Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Supervised Representation Learning with Double Encoding-Layer Autoencoder for Transfer Learning

Published: 23 October 2017 Publication History

Abstract

Transfer learning has gained a lot of attention and interest in the past decade. One crucial research issue in transfer learning is how to find a good representation for instances of different domains such that the divergence between domains can be reduced with the new representation. Recently, deep learning has been proposed to learn more robust or higher-level features for transfer learning. In this article, we adapt the autoencoder technique to transfer learning and propose a supervised representation learning method based on double encoding-layer autoencoder. The proposed framework consists of two encoding layers: one for embedding and the other one for label encoding. In the embedding layer, the distribution distance of the embedded instances between the source and target domains is minimized in terms of KL-Divergence. In the label encoding layer, label information of the source domain is encoded using a softmax regression model. Moreover, to empirically explore why the proposed framework can work well for transfer learning, we propose a new effective measure based on autoencoder to compute the distribution distance between different domains. Experimental results show that the proposed new measure can better reflect the degree of transfer difficulty and has stronger correlation with the performance from supervised learning algorithms (e.g., Logistic Regression), compared with previous ones, such as KL-Divergence and Maximum Mean Discrepancy. Therefore, in our model, we have incorporated two distribution distance measures to minimize the difference between source and target domains in the embedding representations. Extensive experiments conducted on three real-world image datasets and one text data demonstrate the effectiveness of our proposed method compared with several state-of-the-art baseline methods.

References

[1]
Somnath Banerjee and Martin Scholz. 2008. Leveraging web 2.0 sources for web content classification. In Proceedings of the 2008 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology, Volume 1. IEEE Computer Society, 300--306.
[2]
Yoshua Bengio. 2009. Learning deep architectures for AI. Found. Trends Mach. Learn. 2, 1 (2009), 1--127.
[3]
John Blitzer, Ryan McDonald, and Fernando Pereira. 2006. Domain adaptation with structural correspondence learning. In Proceedings of the 2006 Conference on Empirical Methods on Natural Language Processing. 120--128.
[4]
Minmin Chen, Zhixiang Eddie Xu, Kilian Q. Weinberger, and Fei Sha. 2012. Marginalized denoising autoencoders for domain adaptation. In Proceedings of the 29th International Conference on Machine Learning.
[5]
Koby Crammer, Mark Dredze, and Fernando Pereira. 2012. Confidence-weighted linear classification for text categorization. J. Mach. Learn. Res. 13, 1 (2012), 1891--1926.
[6]
W. Y. Dai, G. R. Xue, Q. Yang, and Y. Yu. 2007a. Co-clustering based classification for out-of-domain documents. In Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.
[7]
W. Y. Dai, Q. Yang, G. R. Xue, and Y. Yu. 2007b. Boosting for transfer learning. In Proceedings of the 24th International Conference on Machine Learning.
[8]
Trevor Hastie, Friedman Jerome, and Rob Tibshirani. 2010. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 33, 1 (2010), 1.
[9]
Yaroslav Ganin and Victor Lempitsky. 2015. Unsupervised domain adaptation by backpropagation. In Proceedings of the 32nd International Conference on Machine Learning (ICML’15). 1180--1189.
[10]
J. Gao, W. Fan, J. Jiang, and J. W. Han. 2008. Knowledge transfer via multiple model local structure mapping. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.
[11]
Mingming Gong, Kun Zhang, Tongliang Liu, Dacheng Tao, Clark Glymour, and Bernhard Schölkopf. 2016. Domain adaptation with conditional transferable components. In Proceedings of the 33rd International Conference on Machine Learning. 2839--2848.
[12]
Jing Jiang and Chengxiang Zhai. 2007. Instance weighting for domain adaptation in NLP. In Proceedings of the 2007 Conference of the Association for Computational Linguistics. 264--271.
[13]
Ivor Tsang, Joey Tianyi Zhou, Sinno Jialin Pan, and Yan Yan. 2014. Hybrid heterogeneous transfer learning through deep learning. In Proceedings of the 28th AAAI Conference on Artificial Intelligence. 2213--2220.
[14]
Solomon Kullback. 1987. Letter to the editor: The Kullback-Leibler distance (1987).
[15]
Andrew R. Liddle, Pia Mukherjee, and David Parkinson. 2010. Model selection and multi-model inference. Bayes. Meth. Cosmol. 1 (2010), 79.
[16]
Tongliang Liu, Dacheng Tao, Mingli Song, and Stephen J. Maybank. 2017. Algorithm-dependent generalization bounds for multi-task learning. IEEE Trans. Pattern Anal. Mach. Intell. 39, 2 (2017), 227--241.
[17]
Mingsheng Long, Yue Cao, Jianmin Wang, and Michael I. Jordan. 2015. Learning transferable features with deep adaptation networks. In Proceedings of the International Machine Learning Society (ICML’15). 97--105.
[18]
Mingsheng Long, Jianmin Wang, Yue Cao, Jianguang Sun, and Philip S. Yu. 2016. Deep learning of transferable representation for scalable domain adaptation. IEEE Trans. Knowl. Data Min. 28, 8 (2016), 2027--2040.
[19]
Yong Luo, Tongliang Liu, Dacheng Tao, and Chao Xu. 2014. Decomposition-based transfer distance metric learning for image classification. IEEE Trans. Image Process. 23, 9 (2014), 3789--3801.
[20]
Cope Mallah and Orwell. 2013. Plant leaf classification using probabilistic integration of shape, texture and margin features. Signal Processing, Pattern Recognition and Applications (2013).
[21]
S. J. Pan, J. T. Kwok, and Q. Yang. 2008. Transfer learning via dimensionality reduction. In Proceedings of the 23rd AAAI Conference on Artificial Intelligence.
[22]
Sinno Jialin Pan, Ivor W. Tsang, James T. Kwok, Qiang Yang, and others. 2011. Domain adaptation via transfer component analysis. IEEE Trans. Neural Networks 22, 2 (2011), 199--210.
[23]
Sinno Jialin Pan and Qiang Yang. 2010. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22, 10 (2010), 1345--1359.
[24]
Christopher Poultney, Sumit Chopra, Yann L. Cun, and others. 2006. Efficient learning of sparse representations with an energy-based model. In Advances in Neural Information Processing Systems. 1137--1144.
[25]
Si Si, Dacheng Tao, and Bo Geng. 2010. Bregman divergence-based regularization for transfer subspace learning. IEEE Trans. Knowl. Data Eng. 22, 7 (2010), 929--942.
[26]
Jan Snyman. 2005. Practical Mathematical Optimization: An Introduction to Basic Optimization Theory and Classical and New Gradient-based Algorithms. Vol. 97. Springer Science 8 Business Media.
[27]
Eric Tzeng, Judy Hoffman, Trevor Darrell, and Kate Saenko. 2015. Simultaneous deep transfer across domains and tasks. In Proceedings of the IEEE International Conference on Computer Vision. 4068--4076.
[28]
Bengio Vincent Larochelle and Manzagol. 2008. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th International Conference on Machine Learning. ACM, 1096--1103.
[29]
Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, and Pierre-Antoine Manzagol. 2010. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 11 (2010), 3371--3408.
[30]
Antoine Xavier and Bengio. 2011. Domain adaptation for large-scale sentiment classification: A deep learning approach. In Proceedings of the 28th International Conference on Machine Learning. 513--520.
[31]
D. K. Xing, W. Y. Dai, G. R. Xue, and Y. Yu. 2007. Bridged refinement for transfer learning. In Proceedings of the 10th Pacific Asia Knowledge Discovery and Data Mining.
[32]
Fuzhen Zhuang, Xiaohu Cheng, Ping Luo, Sinno Jialin Pan, and Qing He. 2015. Supervised representation learning: Transfer learning with deep autoencoders. In Proceedings of 24th International Joint Conference on Artificial Intelligence. 4119--4125.
[33]
Fuzhen Zhuang, Xiaohu Cheng, Sinno Jialin Pan, Wenchao Yu, Qing He, and Zhongzhi Shi. 2014. Transfer learning with multiple sources via consensus regularized autoencoders. In Machine Learning and Knowledge Discovery in Databases. Springer, 417--431.
[34]
Fuzhen Zhuang, Ping Luo, Hui Xiong, Yuhong Xiong, Qing He, and Zhongzhi Shi. 2010. Cross-domain learning from multiple sources: A consensus regularization perspective. IEEE Trans. Knowl. Data Eng. 22, 12 (2010), 1664--1678.

Cited By

View all
  • (2024)Unsupervised Domain Adaptation via Weighted Sequential Discriminative Feature Learning for Sentiment AnalysisApplied Sciences10.3390/app1401040614:1(406)Online publication date: 1-Jan-2024
  • (2024)A separation and alignment framework for black-box domain adaptationProceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence10.1609/aaai.v38i14.29532(16005-16013)Online publication date: 20-Feb-2024
  • (2024)Stable Learning via Triplex LearningIEEE Transactions on Artificial Intelligence10.1109/TAI.2024.34044115:10(5267-5276)Online publication date: Oct-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Intelligent Systems and Technology
ACM Transactions on Intelligent Systems and Technology  Volume 9, Issue 2
Regular Papers
March 2018
191 pages
ISSN:2157-6904
EISSN:2157-6912
DOI:10.1145/3154791
  • Editor:
  • Yu Zheng
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 October 2017
Accepted: 01 June 2017
Revised: 01 June 2017
Received: 01 January 2017
Published in TIST Volume 9, Issue 2

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Double encoding-layer autoencoder
  2. distribution difference measure
  3. representation learning

Qualifiers

  • Research-article
  • Research
  • Refereed

Funding Sources

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)52
  • Downloads (Last 6 weeks)10
Reflects downloads up to 09 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Unsupervised Domain Adaptation via Weighted Sequential Discriminative Feature Learning for Sentiment AnalysisApplied Sciences10.3390/app1401040614:1(406)Online publication date: 1-Jan-2024
  • (2024)A separation and alignment framework for black-box domain adaptationProceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence10.1609/aaai.v38i14.29532(16005-16013)Online publication date: 20-Feb-2024
  • (2024)Stable Learning via Triplex LearningIEEE Transactions on Artificial Intelligence10.1109/TAI.2024.34044115:10(5267-5276)Online publication date: Oct-2024
  • (2023)RePoProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3667531(32445-32467)Online publication date: 10-Dec-2023
  • (2023)A Texture Feature Removal Network for Sonar Image Classification and DetectionRemote Sensing10.3390/rs1503061615:3(616)Online publication date: 20-Jan-2023
  • (2023)Transfer Discriminant Softmax Regression with Weighted MMDIEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences10.1587/transfun.2022EAP1162E106.A:10(1343-1353)Online publication date: 1-Oct-2023
  • (2023)Predicting Aging-Related Bugs Using Network Analysis on Aging-Related Dependency NetworksIEEE Transactions on Emerging Topics in Computing10.1109/TETC.2023.327938811:3(566-579)Online publication date: 1-Jul-2023
  • (2023)Self-Adaptive Imbalanced Domain Adaptation With Deep Sparse AutoencoderIEEE Transactions on Artificial Intelligence10.1109/TAI.2022.31968134:5(1293-1304)Online publication date: Oct-2023
  • (2023)Cross-domain Transfer Learning and State Inference for Soft Robots via a Semi-supervised Sequential Variational Bayes Framework2023 IEEE International Conference on Robotics and Automation (ICRA)10.1109/ICRA48891.2023.10160662(552-559)Online publication date: 29-May-2023
  • (2023)Representation learning via an integrated autoencoder for unsupervised domain adaptationFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-022-1349-517:5Online publication date: 5-Jan-2023
  • Show More Cited By

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media