Abstract
In multi-label classification, high-dimensional and sparse binary label vectors usually make existing multi-label classifiers perform unsatisfactorily, which induces a group of label compression coding (LCC) techniques particularly. So far, several linear LCC methods have been introduced via considering linear relations among labels. In this paper, we extend traditional three-layer auto-encoder to construct a five-layer one (i.e., five-layer symmetrical neural network), and then apply the training principle in extreme learning machine to determine all network weights. Therefore, a non-linear LCC approach is proposed to capture non-linear relations of labels, where the first three-layer network is regarded as a encoder and the last two layers act as a decoder. The experimental results on three benchmark data sets show that our proposed method performs better than four existing linear LCC methods according to five performance measures.
This work was supported by the Natural Science Foundation of China (NSFC) under Grant 61273246.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Tsoumakas, G., Katakis, I.: Multi-label classification: an overview. Int. J. Data Warehouse. Min. 3(3), 1–13 (2007)
Zhang, M.L., Zhou, Z.H.: A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 26(8), 1338–1351 (2014)
Gibaji, E., Ventura, S.: A tutorial on multilabel learning. ACM Comput. Surv. 47(3), 52:1–52:38 (2015)
Boutell, M.R., Luo, J., Shen, X., Brown, C.M.: Learning multi-label scene classification. Pattern Recogn. 37(9), 1757–1771 (2004)
Zhang, M.L., Zhou, Z.H.: ML-kNN: a lazy learning approach to multi-label learning. Pattern Recogn. 40(5), 2038–2048 (2007)
Xu, J.: Multi-label core vector machine with a zero label. Pattern Recogn. 47(7), 2542–2557 (2014)
Hsu, D., Kakade, S.M., Langfors, J.L., Zhang, T.: Multi-label prediction via compressed sensing. In: Proceedings of the 24th Conference on Neural Information Processing Systems (NIPS2009), pp. 772–780. MIT Press, Cambridge (2010)
Zhou, T., Tao, D., Wu, X.: Compressed labeling on distilled labelsets for multi-label learning. Mach. Learn. 88(1/2), 69–126 (2012)
Tai, F., Lin, H.T.: Multi-label classification with principal label space transformation. Neural Comput. 24(9), 2508–2542 (2012)
Wicker, J., Pfahringer, B., Kramer, S.: Multi-label classification using Boolean matrix decomposition. In: Proceedings of the 27th Annual ACM Symposium on Applied Computing (SAC2012), pp. 179–186. ACM, New York (2012)
Chen, Y., Lin, H.T.: Feature-aware label ppace dimension reduction for multi-label classification. In: Proceedings of the 27th Conference on Neural Information Processing Systems (NIPS2012), pp. 1538–1546. MIT Press, Cambridge (2012)
Lin, Z., Ding, G., Hu, M., Wang, J.: Multi-label classification via feature-aware implicit label space encoding. In: Proceedings of the 31th International Conference on Machine Learning (ICML2014), pp. 325–333. Microtome Publishing, Brookline (2014)
Cao, L., Xu, J.: A label compression coding approach through maximizing dependence between features and labels for multi-label classification. In: Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN2015), pp. 1–8. IEEE Press, New York (2015)
Zhang, J.J., Fang, M., Wang, H., Li, X.: Dependence maximization based label space dimension reduction for multi-label classification. Eng. Appl. Artif. Intell. 45, 453–463 (2015)
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Leraning representations by back-propagating error. Nature 323, 533–536 (1986)
Bengio, Y.: Learning deep architectures for AI. Found. Trends Mach. Learn. 2(1), 1–127 (2009)
Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015)
Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70(1–3), 489–501 (2006)
Huang, G.B., Wang, D.H., Lan, Y.: Extreme learning machine: a survey. Int. J. Mach. Learn. Cybern. 2(2), 107–122 (2011)
Huang, G., Huang, G.B., Song, S., You, K.: Trends in extreme learning machines: a review. Neural Netw. 61, 32–48 (2015)
Multi-label data sets. http://mulan.sourceforge.net/datasets-mlc.html
Brazdil, P.B., Soares, C.: A comparison of ranking methods for classification algorithm selection. In: Lopez de Mantaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 63–74. Springer, Heidelberg (2000)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Luo, J., Cao, L., Xu, J. (2016). A Non-linear Label Compression Coding Method Based on Five-Layer Auto-Encoder for Multi-label Classification. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9949. Springer, Cham. https://doi.org/10.1007/978-3-319-46675-0_45
Download citation
DOI: https://doi.org/10.1007/978-3-319-46675-0_45
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46674-3
Online ISBN: 978-3-319-46675-0
eBook Packages: Computer ScienceComputer Science (R0)