American Journal of Intelligent Systems
p-ISSN: 2165-8978 e-ISSN: 2165-8994
2012; 2(4): 35-39
doi: 10.5923/j.ajis.20120204.01
Mehdi Banitalebi Dehkordi
Speech Processing Research Lab Elec. and Comp. Eng. Dept., Yazd University Yazd, Iran
Correspondence to: Mehdi Banitalebi Dehkordi , Speech Processing Research Lab Elec. and Comp. Eng. Dept., Yazd University Yazd, Iran.
| Email: | ![]() |
Copyright © 2012 Scientific & Academic Publishing. All Rights Reserved.
Microphone arrays are today employed to specify the sound source locations in numerous real time applications such as speech processing in large rooms or acoustic echo cancellation. Signal sources may exist in the near field or far field with respect to the microphones. Current Neural Networks (NNs) based source localization approaches assume far field narrowband sources. One of the important limitations of these NN-based approaches is making balance between computational complexity and the development of NNs; an architecture that is too large or too small will affect the performance in terms of generalization and computational cost. In the previous analysis, saliency subject has been employed to determine the most suitable structure, however, it is time-consuming and the performance is not robust. In this paper, a family of new algorithms for compression of NNs is presented based on Compressive Sampling (CS) theory. The proposed framework makes it possible to find a sparse structure for NNs, and then the designed neural network is compressed by using CS. The key difference between our algorithm and the state-of-the-art techniques is that the mapping is continuously done using the most effective features; therefore, the proposed method has a fast convergence. The empirical work demonstrates that the proposed algorithm is an effective alternative to traditional methods in terms of accuracy and computational complexity.
Keywords: Compressive Sampling, Sound Source, Neural Network, Pruning, Multilayer Perceptron, Greedy Algorithms
[2] fig. 1. In this equation
is the minimum wavelength of the source signal, and D is the microphone array length. With this condition, incoming waves are approximately planar. So, the time delay of the received signal between the reference microphone and the
microphone would be[15]:![]() | (1) |
![]() | Figure 1. Estimation of far-field source location |
![]() | Figure 2. Estimation of near-field source location |
microphone would be[15] fig. 2:![]() | (2) |
![]() | Figure 3. Multilayer Perceptron neural network for sound source localization |
is the signal received at the
microphone and
is the reference microphone
. We can write the signal at the
microphone in terms of the signal at the first microphone signal as follow:![]() | (3) |
and sensor
like below:![]() | (4) |
![]() | (5) |
![]() | (6) |
![]() | Figure 4. Relation between number of hidden neurons and error |
and a dictionary
(the columns of D are referred to as the atoms), we seek a vector solution x satisfying:![]() | (7) |
(known as l0 norm), is the number of non-zero coefficient of x.Several iterative algorithms have been proposed to solve this minimization problem (Greedy Algorithms such as Orthogonal Matching Pursuit (OMP) or Matching Pursuit (MP) and Non-convex local optimization like FOCUSS algorithm[16].
was smaller than S, then this matrix called a
matrix.2. If the number of rows that contain nonzero elements in a matrix was smaller than S then this matrix is called a
matrix.We assume that the training input patterns are stored in a matrix I, and the desired output patterns are stored in a matrix O, then the mathematical model for training of the neural network can be extracted in the form of the following expansion:![]() | (8) |
. When we minimize a weight matrix (w1 or w2), the behavior acts like setting, in mathematical viewpoint, the relating elements in w1 or w2 to zero. Deduction from above shows that the goal of finding the smallest number of weights in NNs within a range of accuracy can consider to be equal to finding an
Matrix w1 or w2. So we can write problem as below:![]() | (9) |
![]() | (10) |
![]() | (11) |
![]() | (12) |
is input matrix of the hidden layer for the compressed neural network. Comparing these equations with (7) we can conclude that these minimization problems can be written as CS problems. In these CS equations
,
and
was used as the dictionary matrixes and
and
are playing the role of the signal matrix. The process of compressing NNs can be regarded as finding different sparse solutions for weight matrix
or
.
|
| [1] | R. Reed, "Pruning algorithms-a survey", IEEE Transactions on Neural Networks, vol. 4, pp. 740-747, May, 1993. |
| [2] | P. Lauret, E. Fock, T. A. Mara, "A node pruning algorithm based on a Fourier amplitude sensitivity test method", IEEE Transactions on Neural Networks, vol. 17, pp. 273-293, March, 2006. |
| [3] | B. Hassibi and D. G. Stork, "Second-order derivatives for network pruning: optimal brain surgeon," Advances in Neural Information Processing Systems, vol. 5, pp. 164-171, 1993. |
| [4] | L. Preehcit, I. Proben, A set of neural networks benchmark problems and benchmarking rules. University Karlsruher, Gerrnany, Teeh. Rep, 21/94, 2004. |
| [5] | M. Hagiwara, "Removal of hidden units and weights for back propagation networks," Proceeding IEEE International Joint Conference on Neural Network, vol. 1, pp. 351-354, Aug. 2002. |
| [6] | M. Mozer and P. Smolensky, "Skc lactonization: a technique for trimming the fat from network via relevance assessme network," Advances in Neural Information Processing Systems, vol. 1, pp. 107-115, 1991. |
| [7] | J. Sietsma and R. Dow, "Creating artificial neural networks that generalize," Neural Networks, vol. 4, no. I, pp. 67-79, 1991. |
| [8] | E. J. Candes, J. Rombcrg, T. Tao, "Robust uncertaintyprinciples: exact signal reconstruction from highly incomplete frequency information," IEEE Transactions on Information Theory, vol. 52, pp. 489-509, Jan. 2006. |
| [9] | J. Haupt, R. Nowak, "Signal reconstruction from noisy random projections," IEEE Transaction on Information Theory, vol. 52, pp. 4036-4048, Aug. 2006. |
| [10] | Y. H. Liu, S. W. Luo, A. J. Li, "Information geometry on pruning of neural network," International Conference on Machine Learning and Cybernetics, Shanghai, Aug. 2004. |
| [11] | J. Yang, A. Bouzerdoum, S. L. Phung, "A neural network pruning approach based on compressive sampling," Proceedings of International Joint Conference on Neural Networks, pp. 3428-3435, New Jersey, USA, Jun. 2009. |
| [12] | H. Rauhut, K. Sehnass, P. Vandergheynst, "Compressed sensing and redundant dictionaries," IEEE Transaction on Information Theory, vol. 54, pp. 2210-2219, May. 2008. |
| [13] | T. Xu and W. Wang, "A compressed sensing approach for underdetermined blind audio source separation with sparse representation," 2009. |
| [14] | J. Laurent, P. Yand, and P. Vandergheynst, "Compressed sensing: when sparsity meets sampling," Feb. 2010. |
| [15] | G. Arslan, F. A. Sakarya, B. L. Evans, "Speaker localization for far field and near field wideband sources using neural networks," Proc. IEEE-EURASIP Workshop on Nonlinear Signal and Image Processing, vol. 2, pp. 569-573, Antalya, Turkey, Jun. 1999. |
| [16] | Y. Le. Cun, J. S. Denker, and S. A. Solla, "Optimal brain damage," Advance Neural Information. Process. Systems, vol. 2, pp. 598-605, 1990. |
| [17] | B. Hassibi and D. G. Stork, "Second-order derivatives for network pruning: optimal brain surgeon,” Advances in Neural Information Processing Systems, vol. 5, pp. 164-171, 1993. |
| [18] | M. Hagiwara, "Removal of hidden units and weights for back propagation networks," Proc. IEEE Int. Joint Conf. Neural Network, pp. 351-354, 1993. |
| [19] | SNNS software, available athttp://www.ra.cs.unituebingen.de/SNNS |