Statistical Independence and Novelty Detection with Information Preserving Nonlinear Maps
- 15 February 1996
- journal article
- Published by MIT Press in Neural Computation
- Vol. 8 (2) , 260-269
- https://doi.org/10.1162/neco.1996.8.2.260
Abstract
According to Barlow (1989), feature extraction can be understood as finding a statistically independent representation of the probability distribution underlying the measured signals. The search for a statistically independent representation can be formulated by the criterion of minimal mutual information, which reduces to decorrelation in the case of gaussian distributions. If nongaussian distributions are to be considered, minimal mutual information is the appropriate generalization of decorrelation as used in linear Principal Component Analyses (PCA). We also generalize to nonlinear transformations by only demanding perfect transmission of information. This leads to a general class of nonlinear transformations, namely symplectic maps. Conservation of information allows us to consider only the statistics of single coordinates. The resulting factorial representation of the joint probability distribution gives a density estimation. We apply this concept to the real world problem of electrical motor fault detection treated as a novelty detection task.Keywords
This publication has 6 references indexed in Scilit:
- Redundancy reduction with information-preserving nonlinear mapsNetwork: Computation in Neural Systems, 1995
- Independent component analysis, A new concept?Signal Processing, 1994
- Supervised Factorial LearningNeural Computation, 1993
- Learning Factorial Codes by Predictability MinimizationNeural Computation, 1992
- A Mathematical Theory of CommunicationBell System Technical Journal, 1948
- Symplectic GeometryAmerican Journal of Mathematics, 1943