Abstract
The goal of few-shot learning(FSL) is to learn from a hand of labeled examples and quickly adapt to a new task. The traditional FSL models use the single-scale feature that does not have strong representative ability. Besides, some previous methods construct graph neural network to get better classifications, while they update nodes indiscriminately, which will result in intra-class information passing between inter-class nodes. In this paper, we propose a new method called Multi-scale Feature Self-enhancement Network(MFSN) for few-shot learning, which extracts multi-scale feature through a novel extractor, and then enhance the multiple features by the selective graph neural networks that can filter out the incorrect passings between nodes through a meta-learner. At last, classification is performed by measuring distances between the augmented unlabeled features and the improved prototypes computed from augmented labeled features. Comparing to the traditional method, our method improves 1-shot accuracy by 11.8% and improves 5-shot by 10.3% on MiniImagenet dataset. Experiments on MiniImagenet, Cifar-100, and Caltech-256 datasets show the effectiveness of the proposed model.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Alfassy A, Karlinsky L, Aides A, Shtok J, Harary S, Feris R, ... Bronstein AM (2019). Laso: Label-set operations networks for multi-label few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 6548–6557)
Benaim S, Wolf L (2018). One-shot unsupervised cross domain translation. In advances in neural information processing systems (pp. 2104-2114)
Boney R, Ilin A (2018). Semi-supervised few-shot learning with maml
Chen Z, Fu Y, Zhang Y, Jiang YG, Xue X, Sigal L (2019) Multi-level semantic feature augmentation for one-shot learning. IEEE Trans Image Process 28(9):4594–4605
Chen M, Fang Y, Wang X, Luo H, Geng Y, Zhang X, ... & Wang B (2020). Diversity Transfer Network for Few-Shot Learning. In AAAI (pp. 10559–10566)
Fei-Fei L, Fergus R, Perona P (2006) One-shot learning of object categories. IEEE Trans Pattern Anal Mach Intell 28(4):594–611
Fink M (2005) Object classification from a single example utilizing class relevance metrics. In advances in neural information processing systems (pp. 449-456)
Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. arXiv preprint arXiv:1703.03400
Garcia V, Bruna J (2017). Few-shot learning with graph neural networks. arXiv preprint arXiv:1711.04043
Girshick R (2015) Fast r-cnn. In proceedings of the IEEE international conference on computer vision (pp. 1440-1448)
Girshick R, Donahue J, Darrell T, Malik J (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587)
Hariharan B, Girshick R (2017). Low-shot visual recognition by shrinking and hallucinating features. In proceedings of the IEEE international conference on computer vision (pp. 3018-3027)
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778)
Honari S, Yosinski J, Vincent P, Pal C (2016). Recombinator networks: learning coarse-to-fine feature aggregation. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5743-5752)
Jamal MA, Qi GJ (2019). Task agnostic meta-learning for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 11719-11727)
Keshari R, Vatsa M, Singh R, Noore A (2018) Learning structure and strength of CNN filters for small sample size training. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 9349-9358)
Kim J, Kim T, Kim S, & Yoo CD (2019). Edge-labeling graph neural network for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 11-20)
Koch G, Zemel R, Salakhutdinov R (2015). Siamese neural networks for one-shot image recognition. In ICML deep learning workshop (Vol. 2)
Liu Y, Lee J, Park M, Kim S, Yang E, Hwang SJ, Yang Y (2018) Learning to propagate labels: Transductive propagation network for few-shot learning. arXiv preprint arXiv:1805.10002
Maaten L, Hinton G (2008) Visualizing data using t-SNE[J]. J Mach Learn Res 9(Nov):2579–2605
Mehrotra A, Dukkipati A (2017). Generative adversarial residual pairwise networks for one shot learning. arXiv preprint arXiv:1703.08033
Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J (2013) Distributed representations of words and phrases and their compositionality. In advances in neural information processing systems (pp. 3111-3119)
Mishra N, Rohaninejad M, Chen X, Abbeel P (2017). A simple neural attentive meta-learner. arXiv preprint arXiv:1707.03141
Munkhdalai T, Yu H (2017) Meta networks. Proceedings of machine learning research 70:2554
Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. arXiv preprint arXiv:1802.05365
Qiao L, Shi Y, Li J, Wang Y, Huang T, Tian Y (1910) Transductive episodic-wise adaptive metric for few-shot learning (2019)
Ravi S, Larochelle H (2016). Optimization as a model for few-shot learning
Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: towards real-time object detection with region proposal networks. In advances in neural information processing systems (pp. 91-99)
Snell J, Swersky K, Zemel R (2017). Prototypical networks for few-shot learning. In advances in neural information processing systems (pp. 4077-4087)
Sun Q, Liu Y, Chua TS, Schiele B (2019) Meta-transfer learning for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 403-412)
Sung F, Yang Y, Zhang L, Xiang T, Torr PH, Hospedales TM (2018). Learning to compare: relation network for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1199-1208)
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, ... Rabinovich A (2015). Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1–9)
Tremblay J, Prakash A, Acuna D, Brophy M, Jampani V, Anil C, ... & Birchfield S (2018) Training deep networks with synthetic data: Bridging the reality gap by domain randomization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (pp. 969–977)
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, ... & Polosukhin I (2017). Attention is all you need. In Advances in neural information processing systems (pp. 5998–6008)
Vinyals O, Blundell C, Lillicrap T, Wierstra D (2016). Matching networks for one shot learning. In advances in neural information processing systems (pp. 3630-3638)
Wang P, Liu L, Shen C, Huang Z, van den Hengel A, Tao Shen H (2017) Multi-attention network for one shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2721-2729)
Wang X, Yu F, Wang R, Darrell T, Gonzalez JE (2019) Tafe-net: task-aware feature embeddings for low shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1831-1840)
Xing EP, Jordan MI, Russell SJ, Ng AY (2003). Distance metric learning with application to clustering with side-information. In advances in neural information processing systems (pp. 521-528)
Ye HJ, Hu H, Zhan DC, Sha F (2020) Few-shot learning via embedding adaptation with set-to-set functions. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 8808-8817)
Yu M, Guo X, Yi J, Chang S, Potdar S, Cheng Y, ... & Zhou B. (2018). Diverse few-shot text classification with multiple metrics. arXiv preprint arXiv:1805.07513
Zamir AR, Sax A, Shen W, Guibas LJ, Malik J, Savarese S (2018). Taskonomy: disentangling task transfer learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3712-3722)
Acknowledgments
This work was supported by the National Natural Science Foundation of China (grant number 61672202) and State Key Program of NSFC-Shenzhen Joint Foundation (grant number U1613217).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Dong, B., Wang, R., Yang, J. et al. Multi-scale feature self-enhancement network for few-shot learning. Multimed Tools Appl 80, 33865–33883 (2021). https://doi.org/10.1007/s11042-021-11205-3
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-021-11205-3