Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

A Non-linear Label Compression Coding Method Based on Five-Layer Auto-Encoder for Multi-label Classification

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2016)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 9949))

Included in the following conference series:

Abstract

In multi-label classification, high-dimensional and sparse binary label vectors usually make existing multi-label classifiers perform unsatisfactorily, which induces a group of label compression coding (LCC) techniques particularly. So far, several linear LCC methods have been introduced via considering linear relations among labels. In this paper, we extend traditional three-layer auto-encoder to construct a five-layer one (i.e., five-layer symmetrical neural network), and then apply the training principle in extreme learning machine to determine all network weights. Therefore, a non-linear LCC approach is proposed to capture non-linear relations of labels, where the first three-layer network is regarded as a encoder and the last two layers act as a decoder. The experimental results on three benchmark data sets show that our proposed method performs better than four existing linear LCC methods according to five performance measures.

This work was supported by the Natural Science Foundation of China (NSFC) under Grant 61273246.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Tsoumakas, G., Katakis, I.: Multi-label classification: an overview. Int. J. Data Warehouse. Min. 3(3), 1–13 (2007)

    Article  Google Scholar 

  2. Zhang, M.L., Zhou, Z.H.: A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 26(8), 1338–1351 (2014)

    Article  Google Scholar 

  3. Gibaji, E., Ventura, S.: A tutorial on multilabel learning. ACM Comput. Surv. 47(3), 52:1–52:38 (2015)

    Google Scholar 

  4. Boutell, M.R., Luo, J., Shen, X., Brown, C.M.: Learning multi-label scene classification. Pattern Recogn. 37(9), 1757–1771 (2004)

    Article  Google Scholar 

  5. Zhang, M.L., Zhou, Z.H.: ML-kNN: a lazy learning approach to multi-label learning. Pattern Recogn. 40(5), 2038–2048 (2007)

    Article  MATH  Google Scholar 

  6. Xu, J.: Multi-label core vector machine with a zero label. Pattern Recogn. 47(7), 2542–2557 (2014)

    Article  MATH  Google Scholar 

  7. Hsu, D., Kakade, S.M., Langfors, J.L., Zhang, T.: Multi-label prediction via compressed sensing. In: Proceedings of the 24th Conference on Neural Information Processing Systems (NIPS2009), pp. 772–780. MIT Press, Cambridge (2010)

    Google Scholar 

  8. Zhou, T., Tao, D., Wu, X.: Compressed labeling on distilled labelsets for multi-label learning. Mach. Learn. 88(1/2), 69–126 (2012)

    Article  MathSciNet  MATH  Google Scholar 

  9. Tai, F., Lin, H.T.: Multi-label classification with principal label space transformation. Neural Comput. 24(9), 2508–2542 (2012)

    Article  MathSciNet  MATH  Google Scholar 

  10. Wicker, J., Pfahringer, B., Kramer, S.: Multi-label classification using Boolean matrix decomposition. In: Proceedings of the 27th Annual ACM Symposium on Applied Computing (SAC2012), pp. 179–186. ACM, New York (2012)

    Google Scholar 

  11. Chen, Y., Lin, H.T.: Feature-aware label ppace dimension reduction for multi-label classification. In: Proceedings of the 27th Conference on Neural Information Processing Systems (NIPS2012), pp. 1538–1546. MIT Press, Cambridge (2012)

    Google Scholar 

  12. Lin, Z., Ding, G., Hu, M., Wang, J.: Multi-label classification via feature-aware implicit label space encoding. In: Proceedings of the 31th International Conference on Machine Learning (ICML2014), pp. 325–333. Microtome Publishing, Brookline (2014)

    Google Scholar 

  13. Cao, L., Xu, J.: A label compression coding approach through maximizing dependence between features and labels for multi-label classification. In: Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN2015), pp. 1–8. IEEE Press, New York (2015)

    Google Scholar 

  14. Zhang, J.J., Fang, M., Wang, H., Li, X.: Dependence maximization based label space dimension reduction for multi-label classification. Eng. Appl. Artif. Intell. 45, 453–463 (2015)

    Article  Google Scholar 

  15. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Leraning representations by back-propagating error. Nature 323, 533–536 (1986)

    Article  Google Scholar 

  16. Bengio, Y.: Learning deep architectures for AI. Found. Trends Mach. Learn. 2(1), 1–127 (2009)

    Article  MathSciNet  MATH  Google Scholar 

  17. Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)

    Article  Google Scholar 

  18. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015)

    Article  Google Scholar 

  19. Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70(1–3), 489–501 (2006)

    Article  Google Scholar 

  20. Huang, G.B., Wang, D.H., Lan, Y.: Extreme learning machine: a survey. Int. J. Mach. Learn. Cybern. 2(2), 107–122 (2011)

    Article  Google Scholar 

  21. Huang, G., Huang, G.B., Song, S., You, K.: Trends in extreme learning machines: a review. Neural Netw. 61, 32–48 (2015)

    Article  MATH  Google Scholar 

  22. Multi-label data sets. http://mulan.sourceforge.net/datasets-mlc.html

  23. Brazdil, P.B., Soares, C.: A comparison of ranking methods for classification algorithm selection. In: Lopez de Mantaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 63–74. Springer, Heidelberg (2000)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianhua Xu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Luo, J., Cao, L., Xu, J. (2016). A Non-linear Label Compression Coding Method Based on Five-Layer Auto-Encoder for Multi-label Classification. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9949. Springer, Cham. https://doi.org/10.1007/978-3-319-46675-0_45

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-46675-0_45

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-46674-3

  • Online ISBN: 978-3-319-46675-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics