Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4682))

Included in the following conference series:

  • 1334 Accesses

Abstract

In this paper, a novel supervised information feature compression algorithm based on divergence criterion is set up. Firstly, according to the information theory, the concept and its properties of the discrete divergence, i.e. average separability information (ASI) is studied, and a concept of symmetry average separability information (SASI) is proposed, and proved that the SASI here is a kind of distance measure, i.e. the SASI satisfies three requests of distance axiomatization, which can be used to measure the difference degree of a two-class problem. Secondly, based on the SASI, a compression theorem is given, and can be used to design information feature compression algorithm. Based on these discussions, we construct a novel supervised information feature compression algorithm based on the average SASI criterion for multi-class. At last, the experimental results demonstrate that the algorithm here is valid and reliable

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Duda, R.O., Hart, P.E. (eds.): Pattern Classification and Scene Analysis. Wiley, New York (1973)

    MATH  Google Scholar 

  2. Devroye, L., Gyorfi, L., Lugosi, G. (eds.): A Probabilistic Theory of Pattern Recognition. Springer, New York (1996)

    MATH  Google Scholar 

  3. Ding, S.F., Shi, Z.Z.: Studies on Incidence Pattern Recognition Based on Information Entropy. Journal of Information Science 31(6), 497–502 (2005)

    Article  Google Scholar 

  4. Fukunaga, K. (ed.): Introduction to Statistical Pattern Recognition, 2nd edn. Academic Press, New York (1990)

    MATH  Google Scholar 

  5. Hand, D.J. (ed.): Discrimination and Classification. Wiley, New York (1981)

    MATH  Google Scholar 

  6. Turk, M., Pentland, A.: Eigenfaces for Recognition. Journal Cognitive Neuroscience 3(1), 71–86 (1991)

    Article  Google Scholar 

  7. Yang, J., Yang, J.Y.: A Generalized K-L Expansion Method That Can Deal With Small Sample Size and High-dimensional Problems. Pattern Analysis Applications 6(6), 47–54 (2003)

    Article  MATH  Google Scholar 

  8. Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley, New York (1991)

    Book  MATH  Google Scholar 

  9. Tang, Q.Y., Feng, M.G.: Practical Statistics and DPS Data Processing System. Science Press, Beijing (2002)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

De-Shuang Huang Laurent Heutte Marco Loog

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ding, S., Ning, W., Jin, F., Xia, S., Shi, Z. (2007). Supervised Information Feature Compression Algorithm Based on Divergence Criterion. In: Huang, DS., Heutte, L., Loog, M. (eds) Advanced Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence. ICIC 2007. Lecture Notes in Computer Science(), vol 4682. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74205-0_95

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-74205-0_95

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-74201-2

  • Online ISBN: 978-3-540-74205-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics