Abstract
Socially cognitive robots are supposed to communicate and interact with humans and other robots in the most natural way. Listeners turn their heads to-ward speakers to enhance communicative attention; this is also an act of appreciation to the speaker. In this paper we have designed and implemented a robotic head, “Ava”, which turns toward the speaker in noisy environments. Ava employs a Speech Activity Detection system which differentiates speech segments of non-speech. Then the speech segments are processed to reduce different kinds of noise levels. The speaker localization system then finds the speaker position in the azimuth plane and commands motors to turn horizon-tally toward the speaker in a smooth trajectory. Ava has two built-in micro-phones inside its ears and employs three different algorithms simultaneously for feature extraction and a two-layer perceptron neural network for localization. Ava operates real-time and updates the position even in its moving phase. Experiments show a precision of +/-5 degrees in white noise in SNR of 10 dB.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Meghdari, A, Alemi, M, Ghazisaedy, M., Taheri, A.R., Karimian, A., Zandvakili, M.: Applying robots as teaching assistant in EFL classes at iranian middle-schools, CD. In: Proc. Int. Conf. on Education & Modern Edu. Tech. (EMET 2013), Venice, Italy, September 28-30 (2013)
Alemi, M., Meghdari, A., Ghazisaedy, M.: The Impact of Social Robotics on L2 Learners’ Anxiety and Attitude in English Vocabulary Acquisition. Int. Journal of Social Robotics (2015)
Alemi, M., Ghanbarzadeh, A., Meghdari, A., Moghaddam, L.J.: Clinical Application of a Humanoid Robot in Pediatric Cancer Interventions. Int. Journal of Social Robotics (2015)
Alemi, M., Meghdari, A., Ghazisaedy, M.: Employing Humanoid Robots for Teaching English Language in Iranian Junior High-Schools. Int. Journal of Humanoid Robotics 11(3) (2014)
Taheri, A.R., Alemi, M., Meghdari, M., Pouretemad, H.R., Holderread S.L.: Clinical application of a humanoid robot in playing imitation games for autistic children in Iran. In: Proc. of the 14th Int. Educational Technology Conf. (IETC), Chicago, USA, September (2014)
Mavridis, N.: A review of verbal and non-verbal human–robot interactive communication. Robotics and Autonomous Systems 63, 22–35 (2014)
Badali, A., Valin, J.M., Michaud, F., Aarabi, P.: Evaluating Real-time Audio Localization Algorithms for Artificial Audition in Robotics, pp. 2033–2038 (2008)
Shafiee, S., Almasganj, F., Vazirnezhad, B., Jafari, A.: A two-stage speech activity detection system considering fractal aspects of prosody. Pattern Recognition Letters 31(9), 936–948 (2010). ISSN 0167-8655
Nakadai, K., Hidai, K., Okuno, H.G., Kitano, H.: Real-time speaker localization and speech separation by audio-visual integration. In: Int. Proc. of IEEE-RAS Intern’l Conf. on Robotics and Automation, pp. 1043–1049 (2001)
Nakadai, K., Okuno, H.G., Kitano, H.: Real-time sound source localization and separation for robot audition. In: Proc. IEEE International Conference on Spoken Language Processing, pp. 193–196 (2001)
Trifa, V.M., Koene, A., Mor´en, J., Cheng, G.: Real-time acoustic source localization in noisy environments for human-robot multimodal interaction. presented at the Robot and Human interactive Communication, Jeju, pp. 393–398 (2007)
Alonso-Martin, F., Gorostiza, J.F., Malfaz, M., Salichs, M.A.: User Localization during Human-Robot Interaction. Sensors, 9913–9935 (2012)
Cech, J., Mittal, R., Deleforge, A., Sanchez-Riera, J., Alameda-Pineda, X., Horaud, R.: Active-speaker detection and localization with microphones and cameras embedded into a robotic head. In: International Conference on Humanoid Robots, pp. 203–210 (2013)
Kim, U.H., Nakadai, K., Okuno, H.G.: Improved sound source localization and front-back disambiguation for humanoid robots with two ears. In: Applied Artificial Intelligence, pp. 282–291 (2013)
Park, Y., Hwang, S.: Artificial robot ear design for sound direction estimation. presented at the Robot & Human Interactive Communication, Jeju, pp. 405–409 (2007)
Hwang, S., Park, Y., Park, Y.S.: Sound direction estimation using an artificial ear for robots. Robotics and Autonomous Systems, 208–217 (2011)
Lee, S., Park, Y., Choi, J.S.: Estimation of multiple sound source directions using artificial robot ears. Applied Acoustics 77, 49–58 (2014)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Saffari, E., Meghdari, A., Vazirnezhad, B., Alemi, M. (2015). Ava (A Social Robot): Design and Performance of a Robotic Hearing Apparatus. In: Tapus, A., André, E., Martin, JC., Ferland, F., Ammi, M. (eds) Social Robotics. ICSR 2015. Lecture Notes in Computer Science(), vol 9388. Springer, Cham. https://doi.org/10.1007/978-3-319-25554-5_44
Download citation
DOI: https://doi.org/10.1007/978-3-319-25554-5_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-25553-8
Online ISBN: 978-3-319-25554-5
eBook Packages: Computer ScienceComputer Science (R0)