Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Multi-scale feature self-enhancement network for few-shot learning

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

The goal of few-shot learning(FSL) is to learn from a hand of labeled examples and quickly adapt to a new task. The traditional FSL models use the single-scale feature that does not have strong representative ability. Besides, some previous methods construct graph neural network to get better classifications, while they update nodes indiscriminately, which will result in intra-class information passing between inter-class nodes. In this paper, we propose a new method called Multi-scale Feature Self-enhancement Network(MFSN) for few-shot learning, which extracts multi-scale feature through a novel extractor, and then enhance the multiple features by the selective graph neural networks that can filter out the incorrect passings between nodes through a meta-learner. At last, classification is performed by measuring distances between the augmented unlabeled features and the improved prototypes computed from augmented labeled features. Comparing to the traditional method, our method improves 1-shot accuracy by 11.8% and improves 5-shot by 10.3% on MiniImagenet dataset. Experiments on MiniImagenet, Cifar-100, and Caltech-256 datasets show the effectiveness of the proposed model.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig.2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Alfassy A, Karlinsky L, Aides A, Shtok J, Harary S, Feris R, ... Bronstein AM (2019). Laso: Label-set operations networks for multi-label few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 6548–6557)

  2. Benaim S, Wolf L (2018). One-shot unsupervised cross domain translation. In advances in neural information processing systems (pp. 2104-2114)

  3. Boney R, Ilin A (2018). Semi-supervised few-shot learning with maml

  4. Chen Z, Fu Y, Zhang Y, Jiang YG, Xue X, Sigal L (2019) Multi-level semantic feature augmentation for one-shot learning. IEEE Trans Image Process 28(9):4594–4605

    Article  MathSciNet  Google Scholar 

  5. Chen M, Fang Y, Wang X, Luo H, Geng Y, Zhang X, ... & Wang B (2020). Diversity Transfer Network for Few-Shot Learning. In AAAI (pp. 10559–10566)

  6. Fei-Fei L, Fergus R, Perona P (2006) One-shot learning of object categories. IEEE Trans Pattern Anal Mach Intell 28(4):594–611

    Article  Google Scholar 

  7. Fink M (2005) Object classification from a single example utilizing class relevance metrics. In advances in neural information processing systems (pp. 449-456)

  8. Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. arXiv preprint arXiv:1703.03400

  9. Garcia V, Bruna J (2017). Few-shot learning with graph neural networks. arXiv preprint arXiv:1711.04043

  10. Girshick R (2015) Fast r-cnn. In proceedings of the IEEE international conference on computer vision (pp. 1440-1448)

  11. Girshick R, Donahue J, Darrell T, Malik J (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587)

  12. Hariharan B, Girshick R (2017). Low-shot visual recognition by shrinking and hallucinating features. In proceedings of the IEEE international conference on computer vision (pp. 3018-3027)

  13. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778)

  14. Honari S, Yosinski J, Vincent P, Pal C (2016). Recombinator networks: learning coarse-to-fine feature aggregation. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5743-5752)

  15. Jamal MA, Qi GJ (2019). Task agnostic meta-learning for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 11719-11727)

  16. Keshari R, Vatsa M, Singh R, Noore A (2018) Learning structure and strength of CNN filters for small sample size training. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 9349-9358)

  17. Kim J, Kim T, Kim S, & Yoo CD (2019). Edge-labeling graph neural network for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 11-20)

  18. Koch G, Zemel R, Salakhutdinov R (2015). Siamese neural networks for one-shot image recognition. In ICML deep learning workshop (Vol. 2)

  19. Liu Y, Lee J, Park M, Kim S, Yang E, Hwang SJ, Yang Y (2018) Learning to propagate labels: Transductive propagation network for few-shot learning. arXiv preprint arXiv:1805.10002

  20. Maaten L, Hinton G (2008) Visualizing data using t-SNE[J]. J Mach Learn Res 9(Nov):2579–2605

    MATH  Google Scholar 

  21. Mehrotra A, Dukkipati A (2017). Generative adversarial residual pairwise networks for one shot learning. arXiv preprint arXiv:1703.08033

  22. Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J (2013) Distributed representations of words and phrases and their compositionality. In advances in neural information processing systems (pp. 3111-3119)

  23. Mishra N, Rohaninejad M, Chen X, Abbeel P (2017). A simple neural attentive meta-learner. arXiv preprint arXiv:1707.03141

  24. Munkhdalai T, Yu H (2017) Meta networks. Proceedings of machine learning research 70:2554

    Google Scholar 

  25. Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. arXiv preprint arXiv:1802.05365

  26. Qiao L, Shi Y, Li J, Wang Y, Huang T, Tian Y (1910) Transductive episodic-wise adaptive metric for few-shot learning (2019)

  27. Ravi S, Larochelle H (2016). Optimization as a model for few-shot learning

  28. Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: towards real-time object detection with region proposal networks. In advances in neural information processing systems (pp. 91-99)

  29. Snell J, Swersky K, Zemel R (2017). Prototypical networks for few-shot learning. In advances in neural information processing systems (pp. 4077-4087)

  30. Sun Q, Liu Y, Chua TS, Schiele B (2019) Meta-transfer learning for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 403-412)

  31. Sung F, Yang Y, Zhang L, Xiang T, Torr PH, Hospedales TM (2018). Learning to compare: relation network for few-shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1199-1208)

  32. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, ... Rabinovich A (2015). Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1–9)

  33. Tremblay J, Prakash A, Acuna D, Brophy M, Jampani V, Anil C, ... & Birchfield S (2018) Training deep networks with synthetic data: Bridging the reality gap by domain randomization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (pp. 969–977)

  34. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, ... & Polosukhin I (2017). Attention is all you need. In Advances in neural information processing systems (pp. 5998–6008)

  35. Vinyals O, Blundell C, Lillicrap T, Wierstra D (2016). Matching networks for one shot learning. In advances in neural information processing systems (pp. 3630-3638)

  36. Wang P, Liu L, Shen C, Huang Z, van den Hengel A, Tao Shen H (2017) Multi-attention network for one shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2721-2729)

  37. Wang X, Yu F, Wang R, Darrell T, Gonzalez JE (2019) Tafe-net: task-aware feature embeddings for low shot learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1831-1840)

  38. Xing EP, Jordan MI, Russell SJ, Ng AY (2003). Distance metric learning with application to clustering with side-information. In advances in neural information processing systems (pp. 521-528)

  39. Ye HJ, Hu H, Zhan DC, Sha F (2020) Few-shot learning via embedding adaptation with set-to-set functions. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 8808-8817)

  40. Yu M, Guo X, Yi J, Chang S, Potdar S, Cheng Y, ... & Zhou B. (2018). Diverse few-shot text classification with multiple metrics. arXiv preprint arXiv:1805.07513

  41. Zamir AR, Sax A, Shen W, Guibas LJ, Malik J, Savarese S (2018). Taskonomy: disentangling task transfer learning. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3712-3722)

Download references

Acknowledgments

This work was supported by the National Natural Science Foundation of China (grant number 61672202) and State Key Program of NSFC-Shenzhen Joint Foundation (grant number U1613217).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Juan Yang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dong, B., Wang, R., Yang, J. et al. Multi-scale feature self-enhancement network for few-shot learning. Multimed Tools Appl 80, 33865–33883 (2021). https://doi.org/10.1007/s11042-021-11205-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-021-11205-3

Keywords