Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article
Open access
Just Accepted

A Survey of Multimodal Perception Methods for Human-Robot Interaction in Social Environments

Online AM: 29 April 2024 Publication History
  • Get Citation Alerts
  • Abstract

    Human-robot interaction (HRI) in human social environments (HSEs) poses unique challenges for robot perception systems, which must combine asynchronous, heterogeneous data streams in real-time. Multimodal perception systems are well-suited for HRI in HSEs, and can provide more rich, robust interaction for robots operating amongst humans. In this article, we provide an overview of multimodal perception systems being used in HSEs, which is intended to be an introduction to the topic and summary of relevant trends, techniques, resources, challenges, and terminology. We surveyed 15 peer-reviewed robotics and HRI publications over the past 10+ years, providing details about the data acquisition, processing, and fusion techniques used in 65 multimodal perception systems across various HRI domains. Our survey provides information about hardware, software, datasets, and methods currently available for HRI perception research, as well as how these perception systems are being applied in HSEs. Based on the survey, we summarize trends, challenges, limitations of multimodal human perception systems for robots, then identify resources for researchers and developers and propose future research areas to advance the field.

    References

    [1]
    Ayodeji Opeyemi Abioye, Stephen D. Prior, Peter Saddington, and Sarvapali D. Ramchurn. 2022. The performance and cognitive workload analysis of a multimodal speech and visual gesture (mSVG) UAV control interface. Robotics and Autonomous Systems 147 (Jan. 2022), 103915. https://doi.org/10.1016/j.robot.2021.103915
    [2]
    Ayodeji O. Abioye, Stephen D. Prior, Glyn T. Thomas, Peter Saddington, and Sarvapali D. Ramchurn. 2018. The Multimodal Speech and Visual Gesture (mSVG) Control Model for a Practical Patrol, Search, and Rescue Aerobot. In Towards Autonomous Robotic Systems, Manuel Giuliani, Tareq Assaf, and Maria Elena Giannaccini (Eds.). Vol. 10965. Springer International Publishing, Cham, 423–437. https://doi.org/10.1007/978-3-319-96728-8_36 Series Title: Lecture Notes in Computer Science.
    [3]
    Samer Al Moubayed, Jonas Beskow, and Gabriel Skantze. 2014. Spontaneous spoken dialogues with the furhat human-like robot head. In Proceedings of the 2014 ACM/IEEE international conference on Human-robot interaction. ACM, Bielefeld Germany, 326–326. https://doi.org/10.1145/2559636.2559781
    [4]
    Samer Al Moubayed, Jonas Beskow, Gabriel Skantze, and Björn Granström. 2012. Furhat: a back-projected human-like robot head for multiparty human-machine interaction. In Cognitive Behavioural Systems: COST 2102 International Training School, Dresden, Germany, February 21-26, 2011, Revised Selected Papers. Springer, 114–130.
    [5]
    Xavier Alameda-Pineda, Jordi Sanchez-Riera, Johannes Wienke, Vojtěch Franc, Jan Čech, Kaustubh Kulkarni, Antoine Deleforge, and Radu Horaud. 2012. RAVEL: An Annotated Corpus for Training Robots with Audiovisual Abilities. Journal on Multimodal User Interfaces 7 (03 2012). https://doi.org/10.1007/s12193-012-0111-y
    [6]
    Sean Andrist and Dan Bohus. 2020. Accelerating the Development of Multimodal, Integrative-AI Systems with Platform for Situated Intelligence. In AAAI Fall Symposium on Artificial Intelligence for Human-Robot Interaction: Trust & Explainability in Artificial Intelligence for Human-Robot Interaction. https://www.microsoft.com/en-us/research/publication/accelerating-the-development-of-multimodal-integrative-ai-systems-with-platform-for-situated-intelligence/
    [7]
    Sean Andrist, Dan Bohus, and Ashley Feniello. 2019. Demonstrating a Framework for Rapid Development of Physically Situated Interactive Systems. In 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 668–668. https://doi.org/10.1109/HRI.2019.8673067
    [8]
    Pablo Azagra, Florian Golemo, Yoan Mollard, Manuel Lopes, Javier Civera, and Ana C. Murillo. 2017. A multimodal dataset for object model learning from natural human-robot interaction. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 6134–6141. https://doi.org/10.1109/IROS.2017.8206514
    [9]
    Tadas Baltrusaitis, Chaitanya Ahuja, and Louis-Philippe Morency. 2019. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 41, 2 (feb 2019), 423–443. https://doi.org/10.1109/TPAMI.2018.2798607
    [10]
    Yutong Ban, Xiaofei Li, Xavier Alameda-Pineda, Laurent Girin, and Radu Horaud. 2018. Accounting for Room Acoustics in Audio-Visual Multi-Speaker Tracking. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Calgary, AB, 6553–6557. https://doi.org/10.1109/ICASSP.2018.8462100
    [11]
    Siddhartha Banerjee, Andrew Silva, and Sonia Chernova. 2018. Robot Classification of Human Interruptibility and a Study of Its Effects. ACM Transactions on Human-Robot Interaction 7, 2 (July 2018), 1–35. https://doi.org/10.1145/3277902
    [12]
    Baris Bayram and Gökhan Ince. 2015. Audio-visual multi-person tracking for active robot perception. In 2015 IEEE/SICE International Symposium on System Integration (SII). IEEE, Nagoya, 575–580. https://doi.org/10.1109/SII.2015.7405043
    [13]
    Michal Bednarek, Piotr Kicki, and Krzysztof Walas. 2020. On Robustness of Multi-Modal Fusion—Robotics Perspective. Electronics 9, 7 (July 2020), 1152. https://doi.org/10.3390/electronics9071152
    [14]
    Giulia Belgiovine, Jonas Gonzlez-Billandon, Alessandra Sciutti, Giulio Sandini, and Francesco Rea. 2022. HRI Framework for Continual Learning in Face Recognition. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Kyoto, Japan, 8226–8233. https://doi.org/10.1109/IROS47612.2022.9981671
    [15]
    Tony Belpaeme, Paul E Baxter, Robin Read, Rachel Wood, Heriberto Cuayáhuitl, Bernd Kiefer, Stefania Racioppa, Ivana Kruijff-Korbayová, Georgios Athanasopoulos, Valentin Enescu, Rosemarijn Looije, Mark Neerincx, Yiannis Demiris, Raquel Ros-Espinoza, Aryel Beck, Lola Cañamero, Antione Hiolle, Matthew Lewis, Ilaria Baroni, Marco Nalin, Piero Cosi, Giulio Paci, Fabio Tesser, Giacomo Sommavilla, and Remi Humbert. 2013. Multimodal Child-Robot Interaction: Building Social Bonds. Journal of Human-Robot Interaction 1, 2 (Jan. 2013), 33–53. https://doi.org/10.5898/JHRI.1.2.Belpaeme
    [16]
    Atef Ben-Youssef, Chloé Clavel, Slim Essid, Miriam Bilac, Marine Chamoux, and Angelica Lim. 2017. UE-HRI: a new dataset for the study of user engagement in spontaneous human-robot interactions. In Proceedings of the 19th ACM International Conference on Multimodal Interaction. ACM, Glasgow UK, 464–472. https://doi.org/10.1145/3136755.3136814
    [17]
    Atef Ben-Youssef, Giovanna Varni, Slim Essid, and Chloé Clavel. 2019. On-the-Fly Detection of User Engagement Decrease in Spontaneous Human–Robot Interaction Using Recurrent and Deep Neural Networks. International Journal of Social Robotics 11, 5 (Dec. 2019), 815–828. https://doi.org/10.1007/s12369-019-00591-2
    [18]
    Wafa Benkaouar and Dominique Vaufreydaz. 2012. Multi-sensors engagement detection with a robot companion in a home environment. In Workshop on Assistance and Service robotics in a human environment at IEEE International Conference on Intelligent Robots and Systems (IROS2012). 45–52.
    [19]
    Chiara Bodei, Linda Brodo, and Roberto Bruni. 2013. Open Multiparty Interaction. In Recent Trends in Algebraic Development Techniques, Narciso Martí-Oliet and Miguel Palomino (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 1–23.
    [20]
    Dan Bohus and Eric Horvitz. 2009. Dialog in the open world: platform and applications. In Proceedings of the 2009 international conference on Multimodal interfaces - ICMI-MLMI ’09. ACM Press, Cambridge, Massachusetts, USA, 31. https://doi.org/10.1145/1647314.1647323
    [21]
    Dan Bohus and Eric Horvitz. 2010. Facilitating multiparty dialog with gaze, gesture, and speech. In International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction on - ICMI-MLMI ’10. ACM Press, Beijing, China, 1. https://doi.org/10.1145/1891903.1891910
    [22]
    Dan Bohus, Ece Kamar, and Eric Horvitz. 2012. Towards Situated Collaboration. In In NAACL Workshop on Future Directions and Challenges in Spoken Dialog Systems: Tools and Data. https://www.microsoft.com/en-us/research/publication/towards-situated-collaboration/
    [23]
    Qin Cai, David Gallup, Cha Zhang, and Zhengyou Zhang. 2010. 3D Deformable Face Tracking with a Commodity Depth Camera. In Computer Vision – ECCV 2010, Kostas Daniilidis, Petros Maragos, and Nikos Paragios (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 229–242.
    [24]
    Z. Cao, G. Hidalgo Martinez, T. Simon, S. Wei, and Y. A. Sheikh. 2019. OpenPose: Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. IEEE Transactions on Pattern Analysis and Machine Intelligence (2019).
    [25]
    Federico Castanedo. 2013. A Review of Data Fusion Techniques. The Scientific World Journal 2013 (2013), 1–19. https://doi.org/10.1155/2013/704504
    [26]
    Oya Celiktutan, Efstratios Skordos, and Hatice Gunes. 2019. Multimodal Human-Human-Robot Interactions (MHHRI) Dataset for Studying Personality and Engagement. IEEE Transactions on Affective Computing 10, 4 (2019), 484–497. https://doi.org/10.1109/TAFFC.2017.2737019
    [27]
    Crystal Chao and Andrea Thomaz. 2013. Controlling Social Dynamics with a Parametrized Model of Floor Regulation. Journal of Human-Robot Interaction 2, 1 (March 2013), 4–29. https://doi.org/10.5898/JHRI.2.1.Chao
    [28]
    Aaron Chau, Kouhei Sekiguchi, Aditya Arie Nugraha, Kazuyoshi Yoshii, and Kotaro Funakoshi. 2019. Audio-Visual SLAM towards Human Tracking and Human-Robot Interaction in Indoor Environments. In 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN). IEEE, New Delhi, India, 1–8. https://doi.org/10.1109/RO-MAN46459.2019.8956321
    [29]
    Anjun Chen, Xiangyu Wang, Kun Shi, Shaohao Zhu, Bin Fang, Yingfeng Chen, Jiming Chen, Yuchi Huo, and Qi Ye. 2023. ImmFusion: Robust mmWave-RGB Fusion for 3D Human Body Reconstruction in All Weather Conditions. In 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, London, United Kingdom, 2752–2758. https://doi.org/10.1109/ICRA48891.2023.10161428
    [30]
    Anjun Chen, Xiangyu Wang, Shaohao Zhu, Yanxu Li, Jiming Chen, and Qi Ye. 2022. mmBody Benchmark: 3D Body Reconstruction Dataset and Analysis for Millimeter Wave Radar. In Proceedings of the 30th ACM International Conference on Multimedia. ACM, Lisboa Portugal, 3501–3510. https://doi.org/10.1145/3503161.3548262
    [31]
    Chen Chen, Roozbeh Jafari, and Nasser Kehtarnavaz. 2015. UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. In 2015 IEEE International Conference on Image Processing (ICIP). 168–172. https://doi.org/10.1109/ICIP.2015.7350781
    [32]
    Yingfeng Chen, Feng Wu, Wei Shuai, Ningyang Wang, Rongya Chen, and Xiaoping Chen. 2015. KeJia Robot–An Attractive Shopping Mall Guider. In Social Robotics, Adriana Tapus, Elisabeth André, Jean-Claude Martin, François Ferland, and Mehdi Ammi (Eds.). Vol. 9388. Springer International Publishing, Cham, 145–154. https://doi.org/10.1007/978-3-319-25554-5_15 Series Title: Lecture Notes in Computer Science.
    [33]
    Wongun Choi, Khuram Shahid, and Silvio Savarese. 2009. What are they doing? : Collective activity classification using spatio-temporal relationship among people. In 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops. 1282–1289. https://doi.org/10.1109/ICCVW.2009.5457461
    [34]
    Vivian Chu, Kalesha Bullard, and Andrea L. Thomaz. 2014. Multimodal real-time contingency detection for HRI. In 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, Chicago, IL, USA, 3327–3332. https://doi.org/10.1109/IROS.2014.6943025
    [35]
    Nikhil Churamani, Paul Anton, Marc Brügger, Erik Fließwasser, Thomas Hummel, Julius Mayer, Waleed Mustafa, Hwei Geok Ng, Thi Linh Chi Nguyen, Quan Nguyen, Marcus Soll, Sebastian Springenberg, Sascha Griffiths, Stefan Heinrich, Nicolás Navarro-Guerrero, Erik Strahl, Johannes Twiefel, Cornelius Weber, and Stefan Wermter. 2017. The Impact of Personalisation on Human-Robot Interaction in Learning Scenarios. In Proceedings of the 5th International Conference on Human Agent Interaction. ACM, Bielefeld Germany, 171–180. https://doi.org/10.1145/3125739.3125756
    [36]
    Eleonora D’Arca, Neil M. Robertson, and James R. Hopgood. 2016. Robust indoor speaker recognition in a network of audio and video sensors. Signal Processing 129 (Dec. 2016), 137–149. https://doi.org/10.1016/j.sigpro.2016.04.014
    [37]
    Nicola Doering, Sandra Poeschl, Horst-Michael Gross, Andreas Bley, Christian Martin, and Hans-Joachim Boehme. 2015. User-Centered Design and Evaluation of a Mobile Shopping Robot. International Journal of Social Robotics 7, 2 (April 2015), 203–225. https://doi.org/10.1007/s12369-014-0257-8
    [38]
    Niki Efthymiou, Panagiotis P. Filntisis, Petros Koutras, Antigoni Tsiami, Jack Hadfield, Gerasimos Potamianos, and Petros Maragos. 2022. ChildBot: Multi-robot perception and interaction with children. Robotics and Autonomous Systems 150 (April 2022), 103975. https://doi.org/10.1016/j.robot.2021.103975
    [39]
    Olov Engwall, Ronald Cumbal, José Lopes, Mikael Ljung, and Linnea Månsson. 2022. Identification of Low-engaged Learners in Robot-led Second Language Conversations with Adults. ACM Transactions on Human-Robot Interaction 11, 2 (June 2022), 1–33. https://doi.org/10.1145/3503799
    [40]
    Olov Engwall, José Lopes, and Anna Åhlund. 2021. Robot Interaction Styles for Conversation Practice in Second Language Learning. International Journal of Social Robotics 13, 2 (April 2021), 251–276. https://doi.org/10.1007/s12369-020-00635-y
    [41]
    Chiara Filippini, Edoardo Spadolini, Daniela Cardone, Domenico Bianchi, Maurizio Preziuso, Christian Sciarretta, Valentina Del Cimmuto, Davide Lisciani, and Arcangelo Merla. 2021. Facilitating the Child–Robot Interaction by Endowing the Robot with the Capability of Understanding the Child Engagement: The Case of Mio Amico Robot. International Journal of Social Robotics 13, 4 (July 2021), 677–689. https://doi.org/10.1007/s12369-020-00661-w
    [42]
    Panagiotis Paraskevas Filntisis, Niki Efthymiou, Petros Koutras, Gerasimos Potamianos, and Petros Maragos. 2019. Fusing Body Posture With Facial Expressions for Joint Recognition of Affect in Child–Robot Interaction. IEEE Robotics and Automation Letters 4, 4 (Oct. 2019), 4011–4018. https://doi.org/10.1109/LRA.2019.2930434
    [43]
    David Fischinger, Peter Einramhof, Konstantinos Papoutsakis, Walter Wohlkinger, Peter Mayer, Paul Panek, Stefan Hofmann, Tobias Koertner, Astrid Weiss, Antonis Argyros, and Markus Vincze. 2016. Hobbit, a care robot supporting independent living at home: First prototype and lessons learned. Robotics and Autonomous Systems 75 (Jan. 2016), 60–78. https://doi.org/10.1016/j.robot.2014.09.029
    [44]
    Mary Ellen Foster. 2014. Validating attention classifiers for multi-party human-robot interaction. In Proceedings of the 2014 ACM/IEEE International Conference on Human-Robot Interaction: Workshop on Attention Models in Robotics. ACM Press Bielefeld.
    [45]
    Mary Ellen Foster, Rachid Alami, Olli Gestranius, Oliver Lemon, Marketta Niemelä, Jean-Marc Odobez, and Amit Kumar Pandey. 2016. The MuMMER Project: Engaging Human-Robot Interaction in Real-World Public Spaces. In Social Robotics, Arvin Agah, John-John Cabibihan, Ayanna M. Howard, Miguel A. Salichs, and Hongsheng He (Eds.). Vol. 9979. Springer International Publishing, Cham, 753–763. https://doi.org/10.1007/978-3-319-47437-3_74 Series Title: Lecture Notes in Computer Science.
    [46]
    Mary Ellen Foster, Bart Craenen, Amol Deshmukh, Oliver Lemon, Emanuele Bastianelli, Christian Dondrup, Ioannis Papaioannou, Andrea Vanzo, Jean-Marc Odobez, Olivier Canévet, et al. 2019. Mummer: Socially intelligent human-robot interaction in public spaces. arXiv preprint arXiv:1909.06749 (2019).
    [47]
    Mary Ellen Foster, Andre Gaschler, and Manuel Giuliani. 2017. Automatically Classifying User Engagement for Dynamic Multi-party Human–Robot Interaction. International Journal of Social Robotics 9, 5 (Nov. 2017), 659–674. https://doi.org/10.1007/s12369-017-0414-y
    [48]
    Angus Fung, Beno Benhabib, and Goldie Nejat. 2023. Robots Autonomously Detecting People: A Multimodal Deep Contrastive Learning Method Robust to Intraclass Variations. IEEE Robotics and Automation Letters 8, 6 (June 2023), 3550–3557. https://doi.org/10.1109/LRA.2023.3269306
    [49]
    Israel D. Gebru, Silèye Ba, Xiaofei Li, and Radu Horaud. [n. d.]. Audio-Visual Speaker Diarization Based on Spatiotemporal Bayesian Fusion. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 5 ([n. d.]), 1086–1099.
    [50]
    Israel D. Gebru, Sileye Ba, Xiaofei Li, and Radu Horaud. 2018. Audio-Visual Speaker Diarization Based on Spatiotemporal Bayesian Fusion. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 5 (May 2018), 1086–1099. https://doi.org/10.1109/TPAMI.2017.2648793
    [51]
    Sarah Gillet, Ronald Cumbal, André Pereira, José Lopes, Olov Engwall, and Iolanda Leite. 2021. Robot Gaze Can Mediate Participation Imbalance in Groups with Different Skill Levels. In Proceedings of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Boulder CO USA, 303–311. https://doi.org/10.1145/3434073.3444670
    [52]
    Dylan F. Glas, Satoru Satake, Florent Ferreri, Takayuki Kanda, Hiroshi Ishiguro, and Norihiro Hagita. 2013. The Network Robot System: Enabling Social Human-Robot Interaction in Public Spaces. Journal of Human-Robot Interaction 1, 2 (Jan. 2013), 5–32. https://doi.org/10.5898/JHRI.1.2.Glas
    [53]
    Dylan F. Glas, Kanae Wada, Masahiro Shiomi, Takayuki Kanda, Hiroshi Ishiguro, and Norihiro Hagita. 2017. Personal Greetings: Personalizing Robot Utterances Based on Novelty of Observed Behavior. International Journal of Social Robotics 9, 2 (April 2017), 181–198. https://doi.org/10.1007/s12369-016-0385-4
    [54]
    Matthew Gombolay, Anna Bair, Cindy Huang, and Julie Shah. 2017. Computational design of mixed-initiative human–robot teaming that considers human factors: situational awareness, workload, and workflow preferences. The International Journal of Robotics Research 36, 5-7 (June 2017), 597–617. https://doi.org/10.1177/0278364916688255
    [55]
    Randy Gomez, Levko Ivanchuk, Keisuke Nakamura, Takeshi Mizumoto, and Kazuhiro Nakadai. 2015. Utilizing visual cues in robot audition for sound source discrimination in speech-based human-robot communication. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Hamburg, Germany, 4216–4222. https://doi.org/10.1109/IROS.2015.7353974
    [56]
    Randy Gomez, Alvaro Paez, Yu Fang, Serge Thill, Luis Merino, Eric Nichols, Keisuke Nakamura, and Heike Brock. 2022. Developing The Bottom-up Attentional System of A Social Robot. In 2022 International Conference on Robotics and Automation (ICRA). IEEE, Philadelphia, PA, USA, 7402–7408. https://doi.org/10.1109/ICRA46639.2022.9811759
    [57]
    Randy Gomez, Deborah Szapiro, Kerl Galindo, and Keisuke Nakamura. 2018. Haru: Hardware Design of an Experimental Tabletop Robot Assistant. In Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Chicago IL USA, 233–240. https://doi.org/10.1145/3171221.3171288
    [58]
    Jonas Gonzalez, Giulia Belgiovine, Alessandra Sciutti, Giulio Sandini, and Rea Francesco. 2021. Towards a Cognitive Framework for Multimodal Person Recognition in Multiparty HRI. In Proceedings of the 9th International Conference on Human-Agent Interaction. ACM, Virtual Event Japan, 412–416. https://doi.org/10.1145/3472307.3484675
    [59]
    Jonas Gonzalez-Billandon, Giulia Belgiovine, Matthew Tata, Alessandra Sciutti, Giulio Sandini, and Francesco Rea. 2021. Self-supervised learning framework for speaker localisation with a humanoid robot. In 2021 IEEE International Conference on Development and Learning (ICDL). IEEE, Beijing, China, 1–7. https://doi.org/10.1109/ICDL49984.2021.9515566
    [60]
    Jonas Gonzalez-Billandon, Alessandra Sciutti, Matthew Tata, Giulio Sandini, and Francesco Rea. 2020. Audiovisual cognitive architecture for autonomous learning of face localisation by a Humanoid Robot. In 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 5979–5985.
    [61]
    Michael A. Goodrich and Alan C. Schultz. 2007. Human-Robot Interaction: A Survey. Foundations and Trends® in Human-Computer Interaction 1, 3 (2007), 203–275. https://doi.org/10.1561/1100000005
    [62]
    François Grondin and James Glass. 2019. Fast and Robust 3-D Sound Source Localization with DSVD-PHAT. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Macau, China, 5352–5357. https://doi.org/10.1109/IROS40897.2019.8967690
    [63]
    François Grondin, Dominic Létourneau, Cédric Godin, Jean-Samuel Lauzon, Jonathan Vincent, Simon Michaud, Samuel Faucher, and François Michaud. 2021. ODAS: Open embeddeD Audition System. arXiv:2103.03954 [cs, eess] (March 2021). http://arxiv.org/abs/2103.03954 arXiv: 2103.03954.
    [64]
    François Grondin and François Michaud. 2016. Noise mask for TDOA sound source localization of speech on mobile robots in noisy environments. In 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Stockholm, 4530–4535. https://doi.org/10.1109/ICRA.2016.7487652
    [65]
    François Grondin and François Michaud. 2018. Lightweight and Optimized Sound Source Localization and Tracking Methods for Open and Closed Microphone array Configurations. arXiv:1812.00115 [cs, eess] (Nov. 2018). http://arxiv.org/abs/1812.00115 arXiv: 1812.00115.
    [66]
    Hatice Gunes, Frank Broz, Chris S. Crawford, Astrid Rosenthal-von Der Pütten, Megan Strait, and Laurel Riek. 2022. Reproducibility in Human-Robot Interaction: Furthering the Science of HRI. Current Robotics Reports 3, 4 (Oct. 2022), 281–292. https://doi.org/10.1007/s43154-022-00094-5
    [67]
    Raoul Harel, Zerrin Yumak, and Frank Dignum. 2018. Towards a generic framework for multi-party dialogue with virtual humans. In Proceedings of the 31st International Conference on Computer Animation and Social Agents - CASA 2018. ACM Press, Beijing, China, 1–6. https://doi.org/10.1145/3205326.3205327
    [68]
    Kotaro Hoshiba, Osamu Sugiyama, Akihide Nagamine, Ryosuke Kojima, Makoto Kumon, Kazuhiro Nakadai, Department of Systems and Control Engineering, School of Engineering, Tokyo Institute of Technology, Kyoto University Hospital, Department of Electrical and Electronic Engineering, School of Engineering, Tokyo Institute of Technology, Graduate School of Information Science and Engineering, Tokyo Institute of Technology, Graduate School of Science and Technology, Kumamoto University, and Honda Research Institute Japan Co., Ltd. 2017. Design and Assessment of Sound Source Localization System with a UAV-Embedded Microphone array. Journal of Robotics and Mechatronics 29, 1 (Feb. 2017), 154–167. https://doi.org/10.20965/jrm.2017.p0154
    [69]
    Jian-Fang Hu, Wei-Shi Zheng, Jianhuang Lai, and Jianguo Zhang. 2015. Jointly Learning Heterogeneous Features for RGB-D Activity Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
    [70]
    Jwu-Sheng Hu, Chen-Yu Chan, Cheng-Kang Wang, Ming-Tang Lee, and Ching-Yi Kuo. 2011. Simultaneous Localization of a Mobile Robot and Multiple Sound Sources Using a Microphone array. Advanced Robotics 25, 1-2 (Jan. 2011), 135–152. https://doi.org/10.1163/016918610X538525
    [71]
    Ruihan Hu, Songbing Zhou, Zhi Ri Tang, Sheng Chang, Qijun Huang, Yisen Liu, Wei Han, and Edmond Q. Wu. 2021. DMMAN: A two-stage audio–visual fusion framework for sound separation and event localization. Neural Networks 133 (Jan. 2021), 229–239. https://doi.org/10.1016/j.neunet.2020.10.003
    [72]
    Bahar Irfan, Natalia Lyubova, Michael Garcia Ortiz, and Tony Belpaeme. 2018. Multi-modal Open-Set Person Identification in HRI. In 2018 ACM/IEEE International Conference on Human-Robot Interaction Social Robots in the Wild workshop (Chicago, IL, USA). ACM. http://socialrobotsinthewild.org/wp-content/uploads/2018/02/HRI-SRW_2018_paper_6.pdf
    [73]
    Bahar Irfan, Michael Garcia Ortiz, Natalia Lyubova, and Tony Belpaeme. 2022. Multi-modal Open World User Identification. ACM Transactions on Human-Robot Interaction 11, 1 (March 2022), 1–50. https://doi.org/10.1145/3477963
    [74]
    Carlos T. Ishi, Jani Even, and Norihiro Hagita. 2015. Speech activity detection and face orientation estimation using multiple microphone arrays and human position information. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Hamburg, Germany, 5574–5579. https://doi.org/10.1109/IROS.2015.7354167
    [75]
    Md Mofijul Islam and Tariq Iqbal. 2020. HAMLET: A Hierarchical Multimodal Attention-based Human Activity Recognition Algorithm. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Las Vegas, NV, USA, 10285–10292. https://doi.org/10.1109/IROS45743.2020.9340987
    [76]
    Mithun G. Jacob, Yu-Ting Li, and Juan P. Wachs. 2013. Surgical instrument handling and retrieval in the operating room with a multimodal robotic assistant. In 2013 IEEE International Conference on Robotics and Automation. IEEE, Karlsruhe, Germany, 2140–2145. https://doi.org/10.1109/ICRA.2013.6630864
    [77]
    Shomik Jain, Balasubramanian Thiagarajan, Zhonghao Shi, Caitlyn Clabaugh, and Maja J. Matarić. 2020. Modeling engagement in long-term, in-home socially assistive robot interventions for children with autism spectrum disorders. Science Robotics 5, 39 (Feb. 2020), eaaz3791. https://doi.org/10.1126/scirobotics.aaz3791
    [78]
    Jinhyeok Jang, Dohyung Kim, Cheonshu Park, Minsu Jang, Jaeyeon Lee, and Jaehong Kim. 2020. ETRI-Activity3D: A Large-Scale RGB-D Dataset for Robots to Recognize Daily Activities of the Elderly. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Las Vegas, NV, USA, 10990–10997. https://doi.org/10.1109/IROS45743.2020.9341160
    [79]
    Shu Jiang and Ronald C. Arkin. 2015. Mixed-Initiative Human-Robot Interaction: Definition, Taxonomy, and Survey. In 2015 IEEE International Conference on Systems, Man, and Cybernetics. IEEE, Kowloon Tong, Hong Kong, 954–961. https://doi.org/10.1109/SMC.2015.174
    [80]
    Martin Johansson, Gabriel Skantze, and Joakim Gustafson. 2013. Head Pose Patterns in Multiparty Human-Robot Team-Building Interactions. In Social Robotics, David Hutchison, Takeo Kanade, Josef Kittler, Jon M. Kleinberg, Friedemann Mattern, John C. Mitchell, Moni Naor, Oscar Nierstrasz, C. Pandu Rangan, Bernhard Steffen, Madhu Sudan, Demetri Terzopoulos, Doug Tygar, Moshe Y. Vardi, Gerhard Weikum, Guido Herrmann, Martin J. Pearson, Alexander Lenz, Paul Bremner, Adam Spiers, and Ute Leonards (Eds.). Vol. 8239. Springer International Publishing, Cham, 351–360. https://doi.org/10.1007/978-3-319-02675-6_35 Series Title: Lecture Notes in Computer Science.
    [81]
    Michiel Joosse and Vanessa Evers. 2017. A Guide Robot at the Airport: First Impressions. In Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Vienna Austria, 149–150. https://doi.org/10.1145/3029798.3038389
    [82]
    Swapna Joshi, Sawyer Collins, Waki Kamino, Randy Gomez, and Selma Šabanović. 2020. Social Robots for Socio-Physical Distancing. In Social Robotics, Alan R. Wagner, David Feil-Seifer, Kerstin S. Haring, Silvia Rossi, Thomas Williams, Hongsheng He, and Shuzhi Sam Ge (Eds.). Vol. 12483. Springer International Publishing, Cham, 440–452. https://doi.org/10.1007/978-3-030-62056-1_37 Series Title: Lecture Notes in Computer Science.
    [83]
    Malte Jung and Pamela Hinds. 2018. Robots in the Wild: A Time for More Robust Theories of Human-Robot Interaction. ACM Transactions on Human-Robot Interaction (THRI) 7 (05 2018). https://doi.org/10.1145/3208975
    [84]
    Nikolaos Kardaris, Isidoros Rodomagoulakis, Vassilis Pitsikalis, Antonis Arvanitakis, and Petros Maragos. 2016. A Platform for Building New Human-Computer Interface Systems that Support Online Automatic Recognition of Audio-Gestural Commands. In Proceedings of the 24th ACM international conference on Multimedia. ACM, Amsterdam The Netherlands, 1169–1173. https://doi.org/10.1145/2964284.2973794
    [85]
    S Hamidreza Kasaei, Jorik Melsen, Floris van Beers, Christiaan Steenkist, and Klemen Voncina. 2021. The State of Lifelong Learning in Service Robots: Current Bottlenecks in Object Perception and Manipulation. Journal of Intelligent & Robotic Systems 103 (2021), 1–31.
    [86]
    Woo-Ri Ko, Minsu Jang, Jaeyeon Lee, and Jaehong Kim. 2021. AIR-Act2Act: Human–human interaction dataset for teaching non-verbal social behaviors to robots. The International Journal of Robotics Research 40, 4-5 (2021), 691–697.
    [87]
    Thomas Kollar, Anu Vedantham, Corey Sobel, Cory Chang, Vittorio Perera, and Manuela Veloso. 2012. A Multi-modal Approach for Natural Human-Robot Interaction. In Social Robotics, David Hutchison, Takeo Kanade, Josef Kittler, Jon M. Kleinberg, Friedemann Mattern, John C. Mitchell, Moni Naor, Oscar Nierstrasz, C. Pandu Rangan, Bernhard Steffen, Madhu Sudan, Demetri Terzopoulos, Doug Tygar, Moshe Y. Vardi, Gerhard Weikum, Shuzhi Sam Ge, Oussama Khatib, John-John Cabibihan, Reid Simmons, and Mary-Anne Williams (Eds.). Vol. 7621. Springer Berlin Heidelberg, Berlin, Heidelberg, 458–467. https://doi.org/10.1007/978-3-642-34103-8_46 Series Title: Lecture Notes in Computer Science.
    [88]
    Tsuyoshi Komatsubara, Masahiro Shiomi, Thomas Kaczmarek, Takayuki Kanda, and Hiroshi Ishiguro. 2019. Estimating Children’s Social Status Through Their Interaction Activities in Classrooms with a Social Robot. International Journal of Social Robotics 11, 1 (Jan. 2019), 35–48. https://doi.org/10.1007/s12369-018-0474-7
    [89]
    David Kortenkamp, R Peter Bonasso, Dan Ryan, and Debbie Schreckenghost. 1997. Traded control with autonomous robots as mixed initiative interaction. In AAAI Symposium on Mixed Initiative Interaction, Vol. 97. 89–94.
    [90]
    Arkadiusz Kwasigroch, Agnieszka Mikolajczyk, and Michal Grochowski. 2017. Deep neural networks approach to skin lesions classification — A comparative analysis. In 2017 22nd International Conference on Methods and Models in Automation and Robotics (MMAR). IEEE, Miedzyzdroje, Poland, 1069–1074. https://doi.org/10.1109/MMAR.2017.8046978
    [91]
    Javier Laplaza, Nicolas Rodriguez, J. E. Dominguez-Vidal, Fernando Herrero, Sergi Hernandez, Alejandro Lopez, Alberto Sanfeliu, and Anais Garrell. 2022. IVO Robot: A New Social Robot for Human-Robot Collaboration. In 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, Sapporo, Japan, 860–864. https://doi.org/10.1109/HRI53351.2022.9889458
    [92]
    Ivan Laptev. 2005. On Space-Time Interest Points. International Journal of Computer Vision 64, 2 (Sept. 2005), 107–123. https://doi.org/10.1007/s11263-005-1838-7
    [93]
    Séverin Lemaignan, Charlotte E. R. Edmunds, Emmanuel Senft, and Tony Belpaeme. 2018. The PInSoRo dataset: Supporting the data-driven study of child-child and child-robot social dynamics. PLOS ONE 13, 10 (10 2018), 1–19. https://doi.org/10.1371/journal.pone.0205999
    [94]
    Séverin Lemaignan, Mathieu Warnier, E. Akin Sisbot, Aurélie Clodic, and Rachid Alami. 2017. Artificial cognition for social human–robot interaction: An implementation. Artificial Intelligence 247 (June 2017), 45–69. https://doi.org/10.1016/j.artint.2016.07.002
    [95]
    Timothée Lesort, Vincenzo Lomonaco, Andrei Stoian, Davide Maltoni, David Filliat, and Natalia Díaz-Rodríguez. 2020. Continual learning for robotics: Definition, framework, learning strategies, opportunities and challenges. Information fusion 58 (2020), 52–68.
    [96]
    Yuanchao Li, Tianyu Zhao, and Xun Shen. 2020. Attention-Based Multimodal Fusion for Estimating Human Emotion in Real-World HRI. In Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Cambridge United Kingdom, 340–342. https://doi.org/10.1145/3371382.3378261
    [97]
    Rainer Lienhart, Alexander Kuranov, and Vadim Pisarevsky. 2003. Empirical Analysis of Detection Cascades of Boosted Classifiers for Rapid Object Detection. In Pattern Recognition, Gerhard Goos, Juris Hartmanis, Jan van Leeuwen, Bernd Michaelis, and Gerald Krell (Eds.). Vol. 2781. Springer Berlin Heidelberg, Berlin, Heidelberg, 297–304. https://doi.org/10.1007/978-3-540-45243-0_39 Series Title: Lecture Notes in Computer Science.
    [98]
    Timm Linder, Stefan Breuers, Bastian Leibe, and Kai O. Arras. 2016. On multi-modal people tracking from mobile platforms in very crowded and dynamic environments. In 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Stockholm, Sweden, 5512–5519. https://doi.org/10.1109/ICRA.2016.7487766
    [99]
    Timm Linder, Kilian Y. Pfeiffer, Narunas Vaskevicius, Robert Schirmer, and Kai O. Arras. 2020. Accurate detection and 3D localization of humans using a novel YOLO-based RGB-D fusion approach and synthetic training data. In 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Paris, France, 1000–1006. https://doi.org/10.1109/ICRA40945.2020.9196899
    [100]
    Jeroen Linssen and Mariët Theune. 2017. R3D3: The Rolling Receptionist Robot with Double Dutch Dialogue. In Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Vienna Austria, 189–190. https://doi.org/10.1145/3029798.3038372
    [101]
    Guiyu Liu, Jiuchao Qian, Fei Wen, Xiaoguang Zhu, Rendong Ying, and Peilin Liu. 2019. Action Recognition Based on 3D Skeleton and RGB Frame Fusion. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Macau, China, 258–264. https://doi.org/10.1109/IROS40897.2019.8967570
    [102]
    Hongyi Liu, Tongtong Fang, Tianyu Zhou, and Lihui Wang. 2018. Towards Robust Human-Robot Collaborative Manufacturing: Multimodal Fusion. IEEE Access 6 (2018), 74762–74771. https://doi.org/10.1109/ACCESS.2018.2884793
    [103]
    Dongcai Lu, Shiqi Zhang, Peter Stone, and Xiaoping Chen. 2017. Leveraging commonsense reasoning and multimodal perception for robot spoken dialog systems. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Vancouver, BC, 6582–6588. https://doi.org/10.1109/IROS.2017.8206570
    [104]
    Lyujian Lu, Hua Wang, Brian Reily, and Hao Zhang. 2021. Robust Real-Time Group Activity Recognition of Robot Teams. IEEE Robotics and Automation Letters 6, 2 (April 2021), 2052–2059. https://doi.org/10.1109/LRA.2021.3060723
    [105]
    Yaxiong Ma, Yixue Hao, Min Chen, Jincai Chen, Ping Lu, and Andrej Košir. 2019. Audio-visual emotion fusion (AVEF): A deep efficient weighted approach. Information Fusion 46 (March 2019), 184–192. https://doi.org/10.1016/j.inffus.2018.06.003
    [106]
    Umberto Maniscalco, Aniello Minutolo, Pietro Storniolo, and Massimo Esposito. 2024. Towards a more anthropomorphic interaction with robots in museum settings: An experimental study. Robotics and Autonomous Systems 171 (Jan. 2024), 104561. https://doi.org/10.1016/j.robot.2023.104561
    [107]
    Umberto Maniscalco, Pietro Storniolo, and Antonio Messina. 2022. Bidirectional Multi-modal Signs of Checking Human-Robot Engagement and Interaction. International Journal of Social Robotics 14, 5 (July 2022), 1295–1309. https://doi.org/10.1007/s12369-021-00855-w
    [108]
    Mirko Marras, Pedro A. Marín-Reyes, José Javier Lorenzo Navarro, Modesto Fernando Castrillón Santana, and Gianni Fenu. 2019. AveroBot: an audio-visual dataset for people re-identification and verification in human-robot interaction. https://doi.org/10.5220/0007690902550265
    [109]
    Eric Martinson, Wallace Lawson, and J. Gregory Trafton. 2013. Identifying people with soft-biometrics at Fleet Week. In 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, Tokyo, Japan, 49–56. https://doi.org/10.1109/HRI.2013.6483501
    [110]
    E. Martinson and V. Yalla. 2016. Augmenting deep convolutional neural networks with depth-based layered detection for human detection. In 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Daejeon, South Korea, 1073–1078. https://doi.org/10.1109/IROS.2016.7759182
    [111]
    Youssef Mohamed and Severin Lemaignan. 2021. ROS for Human-Robot Interaction. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Prague, Czech Republic, 3020–3027. https://doi.org/10.1109/IROS51168.2021.9636816
    [112]
    Jesús Morales, Ricardo Vázquez-Martín, Anthony Mandow, David Morilla-Cabello, and Alfonso García-Cerezo. 2021. The UMA-SAR Dataset: Multimodal data collection from a ground vehicle during outdoor disaster response training exercises. The International Journal of Robotics Research 40, 6-7 (June 2021), 835–847. https://doi.org/10.1177/02783649211004959
    [113]
    Kazuhiro Nakadai, Gökhan Ince, Keisuke Nakamura, and Hirofumi Nakajima. 2012. Robot audition for dynamic environments. In 2012 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2012). IEEE, Hong Kong, China, 125–130. https://doi.org/10.1109/ICSPCC.2012.6335729
    [114]
    Kazuhiro Nakadai, Hiroshi G. Okuno, Hirofumi Nakajima, Yuji Hasegawa, and Hiroshi Tsujino. 2008. An open source software system for robot audition HARK and its evaluation. In Humanoids 2008 - 8th IEEE-RAS International Conference on Humanoid Robots. IEEE, Daejeon, Korea (South), 561–566. https://doi.org/10.1109/ICHR.2008.4756031
    [115]
    Keisuke Nakamura, Kazuhiro Nakadai, Futoshi Asano, and Gökhan Ince. 2011. Intelligent Sound Source Localization and its application to multimodal human tracking. In 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, San Francisco, CA, 143–148. https://doi.org/10.1109/IROS.2011.6094558
    [116]
    Eric Nichols, Sarah Rose Siskind, Waki Kamino, Selma Šabanović, and Randy Gomez. 2021. Iterative Design of an Emotive Voice for the Tabletop Robot Haru. In Social Robotics, Haizhou Li, Shuzhi Sam Ge, Yan Wu, Agnieszka Wykowska, Hongsheng He, Xiaorui Liu, Dongyu Li, and Jairo Perez-Osorio (Eds.). Vol. 13086. Springer International Publishing, Cham, 362–374. https://doi.org/10.1007/978-3-030-90525-5_31 Series Title: Lecture Notes in Computer Science.
    [117]
    Matthias Nieuwenhuisen and Sven Behnke. 2013. Human-Like Interaction Skills for the Mobile Communication Robot Robotinho. International Journal of Social Robotics 5, 4 (Nov. 2013), 549–561. https://doi.org/10.1007/s12369-013-0206-y
    [118]
    Aastha Nigam and Laurel D. Riek. 2015. Social context perception for mobile robots. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Hamburg, Germany, 3621–3627. https://doi.org/10.1109/IROS.2015.7353883
    [119]
    Timo Ojala, Matti Pietikäinen, and David Harwood. 1996. A comparative study of texture measures with classification based on featured distributions. Pattern Recognition 29, 1 (1996), 51–59. https://doi.org/10.1016/0031-3203(95)00067-4
    [120]
    German I Parisi, Ronald Kemker, Jose L Part, Christopher Kanan, and Stefan Wermter. 2019. Continual lifelong learning with neural networks: A review. Neural networks 113 (2019), 54–71.
    [121]
    Maria Pateraki, Markos Sigalas, Georgios Chliveros, and Panos Trahanias. 2013. Visual human-robot communication in social settings. In Proceedings of ICRA Workshop on Semantics, Identification and Control of Robot-Human-Environment Interaction.
    [122]
    Andre Pereira, Catharine Oertel, Leonor Fermoselle, Joe Mendelson, and Joakim Gustafson. 2019. Responsive Joint Attention in Human-Robot Interaction. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Macau, China, 1080–1087. https://doi.org/10.1109/IROS40897.2019.8968130
    [123]
    Roberto Pinillos, Samuel Marcos, Raul Feliz, Eduardo Zalama, and Jaime Gómez-García-Bermejo. 2016. Long-term assessment of a service robot in a hotel environment. Robotics and Autonomous Systems 79 (May 2016), 40–57. https://doi.org/10.1016/j.robot.2016.01.014
    [124]
    David Portugal, Paulo Alvito, Eleni Christodoulou, George Samaras, and Jorge Dias. 2019. A Study on the Deployment of a Service Robot in an Elderly Care Center. International Journal of Social Robotics 11, 2 (April 2019), 317–341. https://doi.org/10.1007/s12369-018-0492-5
    [125]
    Shokoofeh Pourmehr, Jack Thomas, Jake Bruce, Jens Wawerla, and Richard Vaughan. 2017. Robust sensor fusion for finding HRI partners in a crowd. In 2017 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Singapore, Singapore, 3272–3278. https://doi.org/10.1109/ICRA.2017.7989373
    [126]
    José Augusto Prado, Carlos Simplício, Nicolás F. Lori, and Jorge Dias. 2012. Visuo-auditory Multimodal Emotional Structure to Improve Human-Robot-Interaction. International Journal of Social Robotics 4, 1 (Jan. 2012), 29–51. https://doi.org/10.1007/s12369-011-0134-7
    [127]
    John Páez and Enrique González. 2022. Human-Robot Scaffolding: An Architecture to Foster Problem-solving Skills. ACM Transactions on Human-Robot Interaction 11, 3 (Sept. 2022), 1–17. https://doi.org/10.1145/3526109
    [128]
    Ricardo Ragel, Rafael Rey, Álvaro Páez, Javier Ponce, Keisuke Nakamura, Fernando Caballero, Luis Merino, and Randy Gómez. 2022. Multi-modal Data Fusion for People Perception in the Social Robot Haru. In Social Robotics, Filippo Cavallo, John-John Cabibihan, Laura Fiorini, Alessandra Sorrentino, Hongsheng He, Xiaorui Liu, Yoshio Matsumoto, and Shuzhi Sam Ge (Eds.). Vol. 13817. Springer Nature Switzerland, Cham, 174–187. https://doi.org/10.1007/978-3-031-24667-8_16 Series Title: Lecture Notes in Computer Science.
    [129]
    Dhanesh Ramachandram and Graham W. Taylor. 2017. Deep Multimodal Learning: A Survey on Recent Advances and Trends. IEEE Signal Processing Magazine 34, 6 (Nov. 2017), 96–108. https://doi.org/10.1109/MSP.2017.2738401
    [130]
    Arnaud Ramey, Javier F Gorostiza, and Miguel A Salichs. 2012. A social robot as an aloud reader: putting together recognition and synthesis of voice and gestures for HRI experimentation. (2012).
    [131]
    Caleb Rascon and Ivan Meza. 2017. Localization of sound sources in robotics: A review. Robotics and Autonomous Systems 96 (Oct. 2017), 184–210. https://doi.org/10.1016/j.robot.2017.07.011
    [132]
    Brian Reily, Peng Gao, Fei Han, Hua Wang, and Hao Zhang. 2022. Real-time recognition of team behaviors by multisensory graph-embedded robot learning. The International Journal of Robotics Research 41, 8 (July 2022), 798–811. https://doi.org/10.1177/02783649211043155
    [133]
    Laurel D Riek. 2013. The social co-robotics problem space: Six key challenges. Robotics: Science, and Systems (RSS), Robotics Challenges and Visions (2013).
    [134]
    Adam Robaczewski, Julie Bouchard, Kevin Bouchard, and Sébastien Gaboury. 2021. Socially Assistive Robots: The Specific Case of the NAO. International Journal of Social Robotics 13, 4 (July 2021), 795–831. https://doi.org/10.1007/s12369-020-00664-7
    [135]
    Fraser Robinson and Goldie Nejat. 2023. A Deep Learning Human Activity Recognition Framework for Socially Assistive Robots to Support Reablement of Older Adults. In 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, London, United Kingdom, 6160–6167. https://doi.org/10.1109/ICRA48891.2023.10161404
    [136]
    Nicole Robinson, Brendan Tidd, Dylan Campbell, Dana Kulić, and Peter Corke. 2023. Robotic Vision for Human-Robot Interaction and Collaboration: A Survey and Systematic Review. ACM Transactions on Human-Robot Interaction 12, 1 (March 2023), 1–66. https://doi.org/10.1145/3570731
    [137]
    Isidoros Rodomagoulakis, Nikolaos Kardaris, Vassilis Pitsikalis, E Mavroudi, Athanasios Katsamanis, Antigoni Tsiami, and Petros Maragos. 2016. Multimodal human action recognition in assistive human-robot interaction. In 2016 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, 2702–2706.
    [138]
    Alessandra Rossi, Mariacarla Staffa, Antonio Origlia, Maria di Maro, and Silvia Rossi. 2021. BRILLO: A Robotic Architecture for Personalised Long-lasting Interactions in a Bartending Domain. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Boulder CO USA, 426–429. https://doi.org/10.1145/3434074.3447206
    [139]
    Miguel A. Salichs, Álvaro Castro-González, Esther Salichs, Enrique Fernández-Rodicio, Marcos Maroto-Gómez, Juan José Gamboa-Montero, Sara Marques-Villarroya, José Carlos Castillo, Fernando Alonso-Martín, and Maria Malfaz. 2020. Mini: A New Social Robot for the Elderly. International Journal of Social Robotics 12, 6 (Dec. 2020), 1231–1249. https://doi.org/10.1007/s12369-020-00687-0
    [140]
    Jordi Sanchez-Riera, Xavier Alameda-Pineda, and Radu Horaud. 2012. Audio-visual robot command recognition: D-META’12 grand challenge. In Proceedings of the 14th ACM international conference on Multimodal interaction - ICMI ’12. ACM Press, Santa Monica, California, USA, 371. https://doi.org/10.1145/2388676.2388760
    [141]
    Yoko Sasaki, Ryo Tanabe, and Hiroshi Takernura. 2018. Online Spatial Sound Perception Using Microphone array on Mobile Robot *. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Madrid, 2478–2484. https://doi.org/10.1109/IROS.2018.8593777
    [142]
    Khadija Shaheen, Muhammad Abdullah Hanif, Osman Hasan, and Muhammad Shafique. 2022. Continual learning for real-world autonomous systems: Algorithms, challenges and frameworks. Journal of Intelligent & Robotic Systems 105, 1 (2022), 9.
    [143]
    Amir Shahroudy, Jun Liu, Tian-Tsong Ng, and Gang Wang. 2016. NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
    [144]
    Zhihao Shen, Armagan Elibol, and Nak Young Chong. 2021. Multi-modal feature fusion for better understanding of human personality traits in social human–robot interaction. Robotics and Autonomous Systems 146 (Dec. 2021), 103874. https://doi.org/10.1016/j.robot.2021.103874
    [145]
    Shreyas S. Shivakumar, Neil Rodrigues, Alex Zhou, Ian D. Miller, Vijay Kumar, and Camillo J. Taylor. 2020. PST900: RGB-Thermal Calibration, Dataset and Segmentation Network. In 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Paris, France, 9441–9447. https://doi.org/10.1109/ICRA40945.2020.9196831
    [146]
    Nikhita Singh, Jin Joo Lee, Ishaan Grover, and Cynthia Breazeal. 2018. P2PSTORY: Dataset of Children as Storytellers and Listeners in Peer-to-Peer Interactions. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada) (CHI ’18). Association for Computing Machinery, New York, NY, USA, 1–11. https://doi.org/10.1145/3173574.3174008
    [147]
    Andrew Specian, Ross Mead, Simon Kim, Maja Mataric, and Mark Yim. 2022. Quori: A Community-Informed Design of a Socially Interactive Humanoid Robot. IEEE Transactions on Robotics 38, 3 (June 2022), 1755–1772. https://doi.org/10.1109/TRO.2021.3111718
    [148]
    Micol Spitale, Chris Birmingham, R. Michael Swan, and Maja J Mataric. 2021. Composing HARMONI: An Open-source Tool for Human and Robot Modular OpeN Interaction. In 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Xi’an, China, 3322–3329. https://doi.org/10.1109/ICRA48506.2021.9560992
    [149]
    Ryu Takeda, Kazunori Komatani, and The Institute of Scientific and Industrial Research, Osaka University. 2017. Noise-Robust MUSIC-Based Sound Source Localization Using Steering Vector Transformation for Small Humanoids. Journal of Robotics and Mechatronics 29, 1 (Feb. 2017), 26–36. https://doi.org/10.20965/jrm.2017.p0026
    [150]
    Zheng-Hua Tan, Nicolai Bæk Thomsen, Xiaodong Duan, Evgenios Vlachos, Sven Ewan Shepstone, Morten Højfeldt Rasmussen, and Jesper Lisby Højvang. 2018. iSocioBot: A Multimodal Interactive Social Robot. International Journal of Social Robotics 10, 1 (Jan. 2018), 5–19. https://doi.org/10.1007/s12369-017-0426-7
    [151]
    Matteo Terreran, Leonardo Barcellona, and Stefano Ghidoni. 2023. A general skeleton-based action and gesture recognition framework for human–robot collaboration. Robotics and Autonomous Systems 170 (Dec. 2023), 104523. https://doi.org/10.1016/j.robot.2023.104523
    [152]
    Nadia Magnenat Thalmann, Nidhi Mishra, and Gauri Tulsulkar. 2021. Nadine the Social Robot: Three Case Studies in Everyday Life. In Social Robotics, Haizhou Li, Shuzhi Sam Ge, Yan Wu, Agnieszka Wykowska, Hongsheng He, Xiaorui Liu, Dongyu Li, and Jairo Perez-Osorio (Eds.). Vol. 13086. Springer International Publishing, Cham, 107–116. https://doi.org/10.1007/978-3-030-90525-5_10 Series Title: Lecture Notes in Computer Science.
    [153]
    Mariët Theune, Daan Wiltenburg, Max Bode, and Jeroen Linssen. 2017. R3D3 in the Wild: Using A Robot for Turn Management in Multi-Party Interaction with a Virtual Human. In IVA Workshop on Interaction with Agents and Robots: Different Embodiments, Common Challenges.
    [154]
    Susanne Trick, Franziska Herbert, Constantin A. Rothkopf, and Dorothea Koert. 2022. Interactive Reinforcement Learning With Bayesian Fusion of Multimodal Advice. IEEE Robotics and Automation Letters 7, 3 (July 2022), 7558–7565. https://doi.org/10.1109/LRA.2022.3182100
    [155]
    Rudolph Triebel, Kai Arras, Rachid Alami, Lucas Beyer, Stefan Breuers, Raja Chatila, Mohamed Chetouani, Daniel Cremers, Vanessa Evers, Michelangelo Fiore, Hayley Hung, Omar A. Islas Ramírez, Michiel Joosse, Harmish Khambhaita, Tomasz Kucner, Bastian Leibe, Achim J. Lilienthal, Timm Linder, Manja Lohse, Martin Magnusson, Billy Okal, Luigi Palmieri, Umer Rafi, Marieke van Rooij, and Lu Zhang. 2016. SPENCER: A Socially Aware Service Robot for Passenger Guidance and Help in Busy Airports. In Field and Service Robotics, David S. Wettergreen and Timothy D. Barfoot (Eds.). Vol. 113. Springer International Publishing, Cham, 607–622. https://doi.org/10.1007/978-3-319-27702-8_40 Series Title: Springer Tracts in Advanced Robotics.
    [156]
    Mario Trobinger, Christoph Jahne, Zheng Qu, Jean Elsner, Anton Reindl, Sebastian Getz, Thore Goll, Benjamin Loinger, Tamara Loibl, Christoph Kugler, Carles Calafell, Mohamadreza Sabaghian, Tobias Ende, Daniel Wahrmann, Sven Parusel, Simon Haddadin, and Sami Haddadin. 2021. Introducing GARMI - A Service Robotics Platform to Support the Elderly at Home: Design Philosophy, System Overview and First Results. IEEE Robotics and Automation Letters 6, 3 (July 2021), 5857–5864. https://doi.org/10.1109/LRA.2021.3082012
    [157]
    Antigoni Tsiami, Panagiotis Paraskevas Filntisis, Niki Efthymiou, Petros Koutras, Gerasimos Potamianos, and Petros Maragos. 2018. Far-Field Audio-Visual Scene Perception of Multi-Party Human-Robot Interaction for Children and Adults. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Calgary, AB, 6568–6572. https://doi.org/10.1109/ICASSP.2018.8462425
    [158]
    Matthew Turk. 2001. Perceptual User Interfaces. In Frontiers of Human-Centered Computing, Online Communities and Virtual Environments, Rae A. Earnshaw, Richard A. Guedj, Andries van Dam, and John A. Vince (Eds.). Springer London, London, 39–51. https://doi.org/10.1007/978-1-4471-0259-5_4
    [159]
    Nguyen Tan Viet Tuyen, Alexandra L. Georgescu, Irene Di Giulio, and Oya Celiktutan. 2023. A Multimodal Dataset for Robot Learning to Imitate Social Human-Human Interaction. In Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Stockholm Sweden, 238–242. https://doi.org/10.1145/3568294.3580080
    [160]
    Jean-Marc Valin, Shun’ichi Yamamoto, Jean Rouat, François Michaud, Kazuhiro Nakadai, and Hiroshi G. Okuno. 2007. Robust Recognition of Simultaneous Speech by a Mobile Robot. IEEE Transactions on Robotics 23, 4 (Aug. 2007), 742–752. https://doi.org/10.1109/TRO.2007.900612
    [161]
    Michel Valstar, Björn W. Schuller, Jarek Krajewski, Roddy Cowie, and Maja Pantic. 2014. AVEC 2014: the 4th international audio/visual emotion challenge and workshop. In Proceedings of the 22nd ACM international conference on Multimedia. ACM, Orlando Florida USA, 1243–1244. https://doi.org/10.1145/2647868.2647869
    [162]
    Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In Advances in Neural Information Processing Systems, I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Inc. https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf
    [163]
    Yurii Vasylkiv, Heike Brock, Yu Fang, Eric Nichols, Keisuke Nakamura, Serge Thill, and Randy Gomez. 2020. An Exploration of Simple Reactive Responses for Conveying Aliveness Using the Haru Robot. In Social Robotics, Alan R. Wagner, David Feil-Seifer, Kerstin S. Haring, Silvia Rossi, Thomas Williams, Hongsheng He, and Shuzhi Sam Ge (Eds.). Vol. 12483. Springer International Publishing, Cham, 108–119. https://doi.org/10.1007/978-3-030-62056-1_10 Series Title: Lecture Notes in Computer Science.
    [164]
    Yurii Vasylkiv, Ricardo Ragel, Javier Ponce-Chulani, Luis Merino, Eleanor Sandry, Heike Brock, Keisuke Nakamura, Irani Pourang, and Randy Gomez. 2021. Design and Development of a Teleoperation System for Affective Tabletop Robot Haru. In Social Robotics, Haizhou Li, Shuzhi Sam Ge, Yan Wu, Agnieszka Wykowska, Hongsheng He, Xiaorui Liu, Dongyu Li, and Jairo Perez-Osorio (Eds.). Vol. 13086. Springer International Publishing, Cham, 564–573. https://doi.org/10.1007/978-3-030-90525-5_49 Series Title: Lecture Notes in Computer Science.
    [165]
    Dominique Vaufreydaz, Wafa Johal, and Claudine Combe. 2016. Starting engagement detection towards a companion robot using multimodal features. Robotics and Autonomous Systems 75 (Jan. 2016), 4–16. https://doi.org/10.1016/j.robot.2015.01.004
    [166]
    Paul Viola and Michael J. Jones. 2004. Robust Real-Time Face Detection. International Journal of Computer Vision 57, 2 (May 2004), 137–154. https://doi.org/10.1023/B:VISI.0000013087.49260.fb
    [167]
    Xiang-Yang Wang, Jun-Feng Wu, and Hong-Ying Yang. 2010. Robust image retrieval based on color histogram of local feature regions. Multimedia Tools and Applications 49, 2 (Aug. 2010), 323–345. https://doi.org/10.1007/s11042-009-0362-0
    [168]
    David Whitney, Miles Eldon, John Oberlin, and Stefanie Tellex. 2016. Interpreting multimodal referring expressions in real time. In 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Stockholm, Sweden, 3331–3338. https://doi.org/10.1109/ICRA.2016.7487507
    [169]
    Jason R. Wilson, Phyo Thuta Aung, and Isabelle Boucher. 2022. When to Help? A Multimodal Architecture for Recognizing When a User Needs Help from a Social Robot. In Social Robotics, Filippo Cavallo, John-John Cabibihan, Laura Fiorini, Alessandra Sorrentino, Hongsheng He, Xiaorui Liu, Yoshio Matsumoto, and Shuzhi Sam Ge (Eds.). Vol. 13817. Springer Nature Switzerland, Cham, 253–266. https://doi.org/10.1007/978-3-031-24667-8_23 Series Title: Lecture Notes in Computer Science.
    [170]
    Luke J. Wood, Abolfazl Zaraki, Ben Robins, and Kerstin Dautenhahn. 2021. Developing Kaspar: A Humanoid Robot for Children with Autism. International Journal of Social Robotics 13, 3 (June 2021), 491–508. https://doi.org/10.1007/s12369-019-00563-6
    [171]
    Kai Wu, Shu Ting Goh, and Andy W. H. Khong. 2013. Speaker localization and tracking in the presence of sound interference by exploiting speech harmonicity. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing. 365–369. https://doi.org/10.1109/ICASSP.2013.6637670
    [172]
    Jorge Wuth, Pedro Correa, Tomás Núñez, Matías Saavedra, and Néstor Becerra Yoma. 2021. The Role of Speech Technology in User Perception and Context Acquisition in HRI. International Journal of Social Robotics 13, 5 (Aug. 2021), 949–968. https://doi.org/10.1007/s12369-020-00682-5
    [173]
    Lu Xia, Chia-Chih Chen, and J. K. Aggarwal. 2012. View invariant human action recognition using histograms of 3D joints. In 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops. 20–27. https://doi.org/10.1109/CVPRW.2012.6239233
    [174]
    Haibin Yan, Marcelo H. Ang, and Aun Neow Poo. 2014. A Survey on Perception Methods for Human–Robot Interaction in Social Robots. International Journal of Social Robotics 6, 1 (Jan. 2014), 85–119. https://doi.org/10.1007/s12369-013-0199-6
    [175]
    Zhi Yan, Tom Duckett, and Nicola Bellotto. 2017. Online learning for human classification in 3D LiDAR-based tracking. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 864–871. https://doi.org/10.1109/IROS.2017.8202247
    [176]
    Zhi Yan, Tom Duckett, and Nicola Bellotto. 2020. Online learning for 3D LiDAR-based human detection: experimental analysis of point cloud clustering and classification methods. Autonomous Robots 44, 2 (Jan. 2020), 147–164. https://doi.org/10.1007/s10514-019-09883-y
    [177]
    Zhi Yan, Li Sun, Tom Duckctr, and Nicola Bellotto. 2018. Multisensor Online Transfer Learning for 3D LiDAR-Based Human Detection with a Mobile Robot. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, Madrid, 7635–7640. https://doi.org/10.1109/IROS.2018.8593899
    [178]
    Guang-Zhong Yang, Jim Bellingham, Pierre E. Dupont, Peer Fischer, Luciano Floridi, Robert Full, Neil Jacobstein, Vijay Kumar, Marcia McNutt, Robert Merrifield, Bradley J. Nelson, Brian Scassellati, Mariarosaria Taddeo, Russell Taylor, Manuela Veloso, Zhong Lin Wang, and Robert Wood. 2018. The grand challenges of Science Robotics. Sci. Robot. 3, 14 (Jan. 2018), eaar7650. https://doi.org/10.1126/scirobotics.aar7650
    [179]
    Mohammad Samin Yasar, Md Mofijul Islam, and Tariq Iqbal. 2023. IMPRINT: Interactional Dynamics-aware Motion Prediction in Teams using Multimodal Context. ACM Transactions on Human-Robot Interaction (Oct. 2023), 3626954. https://doi.org/10.1145/3626954
    [180]
    Karim Youssef, Katsutoshi Itoyama, Kazuyoshi Yoshii, and Graduate School of Informatics, Kyoto University. 2017. Simultaneous Identification and Localization of Still and Mobile Speakers Based on Binaural Robot Audition. Journal of Robotics and Mechatronics 29, 1 (Feb. 2017), 59–71. https://doi.org/10.20965/jrm.2017.p0059
    [181]
    Zerrin Yumak and Nadia Magnenat-Thalmann. 2016. Multimodal and Multi-party Social Interactions. In Context Aware Human-Robot and Human-Agent Interaction, Nadia Magnenat-Thalmann, Junsong Yuan, Daniel Thalmann, and Bum-Jae You (Eds.). Springer International Publishing, Cham, 275–298. https://doi.org/10.1007/978-3-319-19947-4_13 Series Title: Human–Computer Interaction Series.
    [182]
    Zerrin Yumak, Jianfeng Ren, Nadia Magnenat Thalmann, and Junsong Yuan. 2014. Modelling Multi-Party Interactions among Virtual Characters, Robots, and Humans. Presence: Teleoperators and Virtual Environments 23, 2 (Aug. 2014), 172–190. https://doi.org/10.1162/PRES_a_00179
    [183]
    Zerrin Yumak, Jianfeng Ren, Nadia Magnenat Thalmann, and Junsong Yuan. 2014. Tracking and fusion for multiparty interaction with a virtual character and a social robot. In SIGGRAPH Asia 2014 Autonomous Virtual Humans and Social Robot for Telepresence. ACM, Shenzhen China, 1–7. https://doi.org/10.1145/2668956.2668958
    [184]
    Brian J. Zhang and Naomi T. Fitter. 2023. Nonverbal Sound in Human-Robot Interaction: a Systematic Review. ACM Transactions on Human-Robot Interaction (Feb. 2023), 3583743. https://doi.org/10.1145/3583743
    [185]
    Zhao Zhao and Rhonda McEwen. 2022. “Let’s read a book together”: A Long-term Study on the Usage of Pre-school Children with Their Home Companion Robot. In 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, Sapporo, Japan, 24–32. https://doi.org/10.1109/HRI53351.2022.9889672
    [186]
    Xiao-Hu Zhou, Xiao-Liang Xie, Zhen-Qiu Feng, Zeng-Guang Hou, Gui-Bin Bian, Rui-Qi Li, Zhen-Liang Ni, Shi-Qi Liu, and Yan-Jie Zhou. 2020. A Multilayer-Multimodal Fusion Architecture for Pattern Recognition of Natural Manipulations in Percutaneous Coronary Interventions. In 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, Paris, France, 3039–3045. https://doi.org/10.1109/ICRA40945.2020.9197111
    [187]
    A. Zlatintsi, A.C. Dometios, N. Kardaris, I. Rodomagoulakis, P. Koutras, X. Papageorgiou, P. Maragos, C.S. Tzafestas, P. Vartholomeos, K. Hauer, C. Werner, R. Annicchiarico, M.G. Lombardi, F. Adriano, T. Asfour, A.M. Sabatini, C. Laschi, M. Cianchetti, A. Güler, I. Kokkinos, B. Klein, and R. López. 2020. I-Support: A robotic platform of an assistive bathing robot for the elderly population. Robotics and Autonomous Systems 126 (April 2020), 103451. https://doi.org/10.1016/j.robot.2020.103451
    [188]
    Athanasia Zlatintsi, Isidoros Rodomagoulakis, Vassilis Pitsikalis, Petros Koutras, Nikolaos Kardaris, Xanthi Papageorgiou, Costas Tzafestas, and Petros Maragos. 2017. Social Human-Robot Interaction for the Elderly: Two Real-life Use Cases. In Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Vienna Austria, 335–336. https://doi.org/10.1145/3029798.3038400
    [189]
    Mateusz Żarkowski. 2019. Multi-party Turn-Taking in Repeated Human–Robot Interactions: An Interdisciplinary Evaluation. International Journal of Social Robotics 11, 5 (Dec. 2019), 693–707. https://doi.org/10.1007/s12369-019-00603-1

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Human-Robot Interaction
    ACM Transactions on Human-Robot Interaction Just Accepted
    EISSN:2573-9522
    Table of Contents
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Online AM: 29 April 2024
    Accepted: 29 February 2024
    Revised: 02 November 2023
    Received: 16 May 2022

    Check for updates

    Author Tags

    1. Human-Robot Interaction
    2. Multimodal Perception
    3. Situated Interaction
    4. Social Robotics
    5. Human Social Environments

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 273
      Total Downloads
    • Downloads (Last 12 months)273
    • Downloads (Last 6 weeks)113
    Reflects downloads up to

    Other Metrics

    Citations

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Full Access

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media