Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
article

Statistical independence and novelty detection with information preserving nonlinear maps

Published: 01 February 1996 Publication History

Abstract

According to Barlow (1989), feature extraction can be understood as finding a statistically independent representation of the probability distribution underlying the measured signals. The search for a statistically independent representation can be formulated by the criterion of minimal mutual information, which reduces to decorrelation in the case of gaussian distributions. If nongaussian distributions are to be considered, minimal mutual information is the appropriate generalization of decorrelation as used in linear Principal Component Analyses (PCA). We also generalize to nonlinear transformations by only demanding perfect transmission of information. This leads to a general class of nonlinear transformations, namely symplectic maps. Conservation of information allows us to consider only the statistics of single coordinates. The resulting factorial representation of the joint probability distribution gives a density estimation. We apply this concept to the real world problem of electrical motor fault detection treated as a novelty detection task.

References

[1]
Abraham, R., and Marsden, J. 1978. Foundations of Mechanics. Benjamin-Cummings, London.
[2]
Barlow, H. 1989. Unsupervised learning. Neural Comp. 1(1), 295-311.
[3]
Bell, A. J., and Sejnowski, T. 1995. An information-maximization approach to blind separation and blind deconvolution. Neural Comp. 7(6), 1129-1159.
[4]
Comon, P. 1994. Independent component analysis, a new concept. Signal Processing 36, 287-314.
[5]
Deco, G., and Brauer, W. 1995. Nonlinear higher order statistical decorrelation by volume conserving neural architectures. Neural Networks (in press).
[6]
Deco, G., and Parra, L. 1996. Nonlinear features extraction by redundancy reduction with stochastic neural networks. Neural Networks (to appear).
[7]
Deco, G., and Schürman, B. 1995. Learning time series evolution by unsupervised extraction of correlations. Phys. Rev. E 51(2), 1780-1790.
[8]
Duda, R. O., and Hart, P. E. 1973. Pattern Classification and Scene Analysis. Wiley, New York.
[9]
Feng Kang, and Qin Meng-zhao. 1985. The symplectic methods for the computation of Hamiltonian equations. In Numerical Methods for Partial Differential Equations, Proceedings of a Conference held in Shanghai, 1987. Lecture Notes in Mathematics. Zhu You-lan and Guo Ben-yu, eds., Vol. 1297, pp. 1- 35. Springer, Berlin.
[10]
Hornik, K., Stinchcombe, M., and White, H. 1989. Multilayer feedforward neural networks are universal approximators. Neural Networks 2, 359-366.
[11]
Miesbach, S., and Pesch, H. J. 1992. Symplectic phase flow approximation for the numerical integration of canonical systems. Numer. Math. 61, 501-521.
[12]
Nadal, J-P., and Parga, N. 1994. Non-linear neurons in the low noise limit: A factorial code maximizes information transfer. Network 5(4), 565-581.
[13]
Papoulis, A. 1991. Probability, Random Variables, and Stochastic Processes, 3rd ed. McGraw-Hill, New York.
[14]
Parra, L., Deco, G., and Miesbach, S. 1995. Redundancy reduction with information preserving nonlinear maps. Network 6, 61-72.
[15]
Redlich, A. N. 1993. Supervised factorial learning. Neural Comp. 5, 750-766.
[16]
Schmidhuber, J. 1992. Learning factorial codes by predictability minimization. Neural Comp. 4(6), 863-879.
[17]
Shannon, C. 1948. A mathematical theory of communication. Bell Syst. Tech. J. 7, 379-423.
[18]
Siegel 1943. Symplectic geometry. Am. J. Math. 65, 1-86.
[19]
Stoer, J., and Bulirsch, R. 1993. Introduction to Numerical Analysis. Springer, New York.

Cited By

View all
  • (2023)Exploiting scatter matrix on one-class support vector machine based on low variance directionIntelligent Data Analysis10.3233/IDA-22703627:6(1663-1679)Online publication date: 1-Jan-2023
  • (2020)Heterogeneous Univariate Outlier Ensembles in Multidimensional DataACM Transactions on Knowledge Discovery from Data10.1145/340393414:6(1-27)Online publication date: 28-Sep-2020
  • (2016)Fast structural ensemble for One-Class ClassificationPattern Recognition Letters10.1016/j.patrec.2016.06.02880:C(179-187)Online publication date: 1-Sep-2016
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Neural Computation
Neural Computation  Volume 8, Issue 2
February 15, 1996
246 pages

Publisher

MIT Press

Cambridge, MA, United States

Publication History

Published: 01 February 1996

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 09 Nov 2024

Other Metrics

Citations

Cited By

View all

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media