Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3570236.3570237acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiciipConference Proceedingsconference-collections
research-article

Zero-shot Hand-held Objects Recognition Based on Global Feature Relationships: Zero-shot Hand-held Objects Recognition Based on Global Feature Relationships

Published: 27 January 2023 Publication History

Abstract

In recent years, there have been frequent terrorist attacks both at home and abroad. Accurate detection of a wide variety of handheld objects is a problem that needs an urgent solution. To better exploit the effects of handheld movements on handheld object recognition, this paper proposes a zero-shot hand-held object recognition based on global feature relations that comprise two modules: an image processing module and a semantic relations module. The image processing module learns the visual classifier of the input image to obtain the classification weights. The semantic relations module propagates structural knowledge through semantic learning through graph convolution operations on the knowledge graph and obtains the classification weights for all categories. Results indicate that this method significantly improves the recognition rate compared to traditional handheld object recognition methods.

References

[1]
Ren S, He K, Girshick R, Faster r-cnn: Towards real-time object detection with region proposal networks[J]. Advances in neural information processing systems, 2015, 28.
[2]
J. Redmon, S. Divvala, R. Girshick, You only look once: Unified, real­time object detection[C]. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, 779­788.
[3]
Liu W, Anguelov D, Erhan D, Ssd: Single shot multibox detector[C]//European conference on computer vision. Springer, Cham, 2016: 21-37.
[4]
M. Norouzi, T. Mikolov, S. Bengio, Zero­shot learning by convex combination of semantic embeddings[J]. arXiv preprint arXiv:1312.5650, 2013.
[5]
S. Changpinyo, W.­L. Chao, F. Sha. Predicting visual exemplars of unseen classes for zero­shot learning[C]. Proceedings of the IEEE international conference on computer vision, 2017, 3476-3485.
[6]
S. Changpinyo, W.­L. Chao, B. Gong, Synthesized classifiers for zero­shot learning[C]. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, 5327-5336.
[7]
M. Kampffmeyer, Y. Chen, X. Liang, Rethinking knowledge graph propagation for zeroshot learning[C]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, 11487­11496.
[8]
J. Deng, W. Dong, R. Socher, Imagenet: A large­scale hierarchical image database[C]. 2009 IEEE conference on computer vision and pattern recognition, 2009, 248­255.
[9]
X. Wang, Y. Ye, A. Gupta. Zero­shot recognition via semantic embeddings and knowledge graphs[C]. Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, 6857­6866 .
[10]
J. Pennington, R. Socher, C. D. Manning. Glove: Global vectors for word representation[C]. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 2014, 1532­1543 .
[11]
B. Perozzi, R. Al­Rfou, S. Skiena. Deepwalk: Online learning of social representations[C]. Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, 2014, 701­710 .
[12]
T. Mikolov, K. Chen, G. Corrado, Efficient estimation of word representations in vector space[J]. arXiv preprint arXiv:1301.3781, 2013.
[13]
F. Agostinelli, M. Hoffman, P. Sadowski, Learning activation functions to improve deep neural networks[J]. arXiv preprint arXiv:1412.6830, 2014.
[14]
A. Frome, G. Corrado, J. Shlens, Devise: A deep visual­semantic embedding model[J].,2013.
[15]
S. Gupta, J. Malik. Visual semantic role labeling[J]. arXiv preprint arXiv:1505.04474, 2015.
[16]
Y.­W. Chao, Y. Liu, X. Liu, Learning to detect human­object interactions[C]. 2018 IEEE winter conference on applications of computer vision (wacv), 2018, 381­389.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICIIP '22: Proceedings of the 7th International Conference on Intelligent Information Processing
September 2022
367 pages
ISBN:9781450396714
DOI:10.1145/3570236
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 27 January 2023

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Hand-held objects recognition
  2. Zero-shot learning

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICIIP '22

Acceptance Rates

Overall Acceptance Rate 87 of 367 submissions, 24%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 61
    Total Downloads
  • Downloads (Last 12 months)14
  • Downloads (Last 6 weeks)2
Reflects downloads up to 09 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media