Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3379247.3379255acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccdeConference Proceedingsconference-collections
research-article

Decision Making of Mobile Robot based on Multimodal Fusion

Published: 07 March 2020 Publication History

Abstract

To solve the problem of multimodal information fusion, this paper proposes a method based on the filling of scene main components, and evaluates the channel information according to the knowledge base. The modal channel of this paper chooses vision and hearing, which is more suitable for the information transmission in the actual communication. Firstly, the single mode information is identified by the neural network. After processing, the image and audio expression are transformed into text expression, and the component value describing the scene is filled in according to text analysis. According to the knowledge base, the evaluation model is established to calculate the confidence of each channel when the information conflicts. After getting the scene, query the content of the prior knowledge base again and send the corresponding action instructions to the robot. The experimental results show that this paper can correct the modal fusion results under the guidance of prior knowledge base, and achieve effective human-computer cooperation in specific scenes. It reduces the dependence on single channel information in the interaction process, increases fault tolerance mechanism, and improves user experience evaluation.

References

[1]
Zhang, C. 2017. Research on Application of industrial robots based on visual guidance. Shanxi University of science and technology.
[2]
Long H., Zhu D. J., Tian J. 2018. An overview of the application of deep learning in intelligent robots. Computer science, 45 (S2): 53--57 + 62.
[3]
Zhang, W. Z., Shao, B. 2018. Application and research review of intelligent robot technology in Library Information Service. Library science research, no.431 (12): 4--9.
[4]
Chen, B., Guo, D., Y. Shi, K., R. 2003. Research on embedded voice interaction system for humanoid robot. Robot, 25 (5): 452--455.
[5]
Duan, J. H., Zhu, Y. A., Huang, S., J., et al. 2013. Research on multi robot task allocation algorithm based on multimodal fusion. Journal of Northwest University of technology, 2013(6): 974--978.
[6]
Yang, C., Wang, D., Zeng, Y., et al. 2018. Knowledge-Based Multimodal Information Fusion for Role Recognition and Situation Assessment by Using Mobile Robot. Information Fusion.
[7]
Aydemir, A., Pronobis, A., Göbelbecker, M., Jensfelt, P. 2013. Active visual object search in unknown environments using uncertain semantics, IEEE Trans. Rob. 29 (4) (2013)986--1002.
[8]
Xiao, Q. K., Qin, M. Y., Guo, P., Zhao, Y. D. 2019. Multimodal Fusion Based on LSTM and a Couple Conditional Hidden Markov Model for Chinese Sign Language Recognition. IEEE Access 7: 112258--112268 (2019)
[9]
Shi, B. 2003. A review of semantic component analysis. Journal of Hefei University of Technology (SOCIAL SCIENCE EDITION), 2003, 17 (4): 130--133.
[10]
Yan, H. J. 2009. Semantic component analysis from a cognitive perspective. Journal of Hunan University (SOCIAL SCIENCE EDITION), 2009, 23 (3): 96--100.
[11]
Yashwanth, H., Mahendrakar, H., David, S. 2004. Automatic speech recognition using audio visual cues. First India Annual Conference. Kharagpur. 166--169 .

Index Terms

  1. Decision Making of Mobile Robot based on Multimodal Fusion

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICCDE '20: Proceedings of 2020 6th International Conference on Computing and Data Engineering
    January 2020
    279 pages
    ISBN:9781450376730
    DOI:10.1145/3379247
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 March 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Evaluation strategy
    2. Human-computer cooperation
    3. Multimodal fusion

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICCDE 2020

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 81
      Total Downloads
    • Downloads (Last 12 months)6
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 01 Nov 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media