Gradient Learning in Structured Parameter Spaces: Adaptive Blind Separation of Signal Sources

The present paper discusses the natural gradient descent learning rules in parameter spaces which have Riemannian geometrical structures. A modi cation is necessary for de ning the steepest descent (gradient) direction in a Riemannian parameter space. Parameter spaces of multilayer perceptrons are good examples of the Riemannian nature. Another example is the space of matrices on which adaptive blind separation of mixtured signals takes place. The ICA (independet component analysis) learning method is discussed from the geometrical point of view. We also show other evidences how the natural gradient learning method is e ective. 1 Gradient Descent Learning Rule Let us consider an observable random variable z and a network speci ed by a parameter vector w 2 R n . Let l(z;w) be a loss function when we use parameter w in processing z. A typical example is the multilayer neural network, where z consists of input vector x and the desired or teacher output vector y, z = (x; y), w is the modi able parameters of the network such that the output of the network is speci ed by

[1]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[2]  Terrence J. Sejnowski,et al.  An Information-Maximization Approach to Blind Separation and Blind Deconvolution , 1995, Neural Computation.

[3]  P. Comon Independent Component Analysis , 1992 .

[4]  Christian Jutten,et al.  Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture , 1991, Signal Process..

[5]  Adel Belouchrani,et al.  A new composite criterion for adaptive and iterative blind source separation , 1994, Proceedings of ICASSP '94. IEEE International Conference on Acoustics, Speech and Signal Processing.

[6]  Shun-ichi Amari,et al.  Differential-geometrical methods in statistics , 1985 .

[7]  Juha Karhunen,et al.  Representation and separation of signals using nonlinear PCA type learning , 1994, Neural Networks.

[8]  Shun-ichi Amari,et al.  Information geometry of Boltzmann machines , 1992, IEEE Trans. Neural Networks.

[9]  Shun-ichi Amari,et al.  Information geometry of the EM and em algorithms for neural networks , 1995, Neural Networks.

[10]  Jean-François Cardoso,et al.  Equivariant adaptive source separation , 1996, IEEE Trans. Signal Process..

[11]  Pierre Comon,et al.  Independent component analysis, A new concept? , 1994, Signal Process..

[12]  S. Amari A Theory ofAdaptive Pattern Classifiers , 1967 .

[13]  Shun-ichi Amari,et al.  Recurrent Neural Networks For Blind Separation of Sources , 1995 .

[14]  Andrzej Cichocki,et al.  A New Learning Algorithm for Blind Signal Separation , 1995, NIPS.