A Context-Sensitive Generalization of ICA

Source separation arises in a surprising number of signal processing applications, from speech recognition to EEG analysis. In the square linear blind source separation problem without time delays, one must find an unmixing matrix which can detangle the result of mixing n unknown independent sources through an unknown n x n mixing matrix. The recently introduced ICA blind source separation algorithm (Baram and Roth 1994; Bell and Sejnowski 1995) is a powerful and surprisingly simple technique for solving this problem. ICA is all the more remarkable for performing so well despite making absolutely no use of the temporal structure of its input! This paper presents a new algorithm, contextual ICA, which derives from a maximum likelihood density estimation formulation of the problem. cICA can incorporate arbitrarily complex adaptive history-sensitive source models, and thereby make use of the temporal structure of its input. This allows it to separate in a number of situations where standard ICA cannot, including sources with low kurtosis, colored gaussian sources, and sources which have gaussian histograms. Since ICA is a special case of cICA, the MLE derivation provides as a corollary a rigorous derivation of classic ICA.

[1]  des lettres et des beaux-arts de Belgique. Nouveaux mémoires de l'Académie royale des sciences et belles-lettres de Bruxelles. , 2022 .

[2]  H. Hotelling Analysis of a complex of statistical variables into principal components. , 1933 .

[3]  William Bialek,et al.  Reading a Neural Code , 1991, NIPS.

[4]  J. Mendel,et al.  Maximum-Likelihood Deconvolution: A Journey into Model-Based Signal Processing , 1990 .

[5]  Geoffrey E. Hinton,et al.  Adaptive Soft Weight Tying using Gaussian Mixtures , 1991, NIPS.

[6]  Barak A. Pearlmutter Temporally continuous vs. clocked networks , 1993 .

[7]  Pierre Comon,et al.  Independent component analysis, A new concept? , 1994, Signal Process..

[8]  Terrence J. Sejnowski,et al.  An Information-Maximization Approach to Blind Separation and Blind Deconvolution , 1995, Neural Computation.

[9]  L. Parra,et al.  Redundancy reduction with information-preserving nonlinear maps , 1995 .

[10]  Yoram Baram,et al.  Forecasting by density shaping using neural networks , 1995, Proceedings of 1995 Conference on Computational Intelligence for Financial Engineering (CIFEr).

[11]  Tzyy-Ping Jung,et al.  Independent Component Analysis of Electroencephalographic Data , 1995, NIPS.

[12]  Andrzej Cichocki,et al.  A New Learning Algorithm for Blind Signal Separation , 1995, NIPS.

[13]  Kari Torkkola,et al.  Blind separation of convolved sources based on information maximization , 1996, Neural Networks for Signal Processing VI. Proceedings of the 1996 IEEE Signal Processing Society Workshop.

[14]  Kari Torkkola,et al.  Blind separation of delayed sources based on information maximization , 1996, 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings.

[15]  T J Sejnowski,et al.  Learning the higher-order structure of a natural sound. , 1996, Network.

[16]  Terrence J. Sejnowski,et al.  The “independent components” of natural scenes are edge filters , 1997, Vision Research.