Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3237383.3238142acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

NADiA - Towards Neural Network Driven Virtual Human Conversation Agents

Published: 09 July 2018 Publication History

Abstract

Advances in artificial intelligence and machine learning - in particular neural networks - have given rise to a new generation of virtual assistants and chatbots. Within this work, we describe the motivation and architecture of NADiA - Neurally Animated Dialog Agent - which leverages both the user's verbal input and facial expressions for multi-modal conversation. NADiA combines a neural language model that generates conversational responses, a convolutional neural network for facial expression analysis, and virtual human technology that is deployed on a mobile phone.

References

[1]
Timothy W Bickmore and Rosalind W Picard. 2005. Establishing and maintaining long-term human-computer relationships. ACM Transactions on Computer-Human Interaction (TOCHI), Vol. 12, 2 (2005), 293--327.
[2]
Justine Cassell. 2000. Embodied conversational agents. MIT press.
[3]
Jonathan Chang and Stefan Scherer. 2017. Learning Representations of Emotional Speech with Deep Convolutional Generative Adversarial Networks. arXiv preprint arXiv:1705.02394 (2017).
[4]
Robert Dale. 2016. The return of the chatbots. Natural Language Engineering Vol. 22, 5 (2016), 811--817.
[5]
Asbjørn Følstad and Petter Bae Brandtzæg. 2017. Chatbots and the new world of HCI. interactions, Vol. 24, 4 (2017), 38--42.
[6]
Sayan Ghosh, Mathieu Chollet, Eugene Laksana, Louis-Philippe Morency, and Stefan Scherer. 2017. Affect-LM: A Neural Language Model for Customizable Affective Text Generation. arXiv preprint arXiv:1704.06851 (2017).
[7]
Sayan Ghosh, Eugene Laksana, Louis-Philippe Morency, and Stefan Scherer. 2016. Representation Learning for Speech Emotion Recognition. INTERSPEECH. 3603--3607.
[8]
Sayan Ghosh, Eugene Laksana, Stefan Scherer, and Louis-Philippe Morency. 2015. A multi-label convolutional neural network approach to cross-domain action unit detection Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on. IEEE, 609--615.
[9]
Mark L Knapp, Judith A Hall, and Terrence G Horgan. 2013. Nonverbal communication in human interaction. Cengage Learning.
[10]
Stefan Kopp, Brigitte Krenn, Stacy Marsella, Andrew N Marshall, Catherine Pelachaud, Hannes Pirker, Kristinn R Thórisson, and Hannes Vilhjálmsson. 2006. Towards a common framework for multimodal generation: The behavior markup language International workshop on intelligent virtual agents. Springer, 205--217.
[11]
Gale M Lucas, Jonathan Gratch, Aisha King, and Louis-Philippe Morency. 2014. It's only a computer: Virtual humans increase willingness to disclose. Computers in Human Behavior Vol. 37 (2014), 94--100.
[12]
Albert Mehrabian. 1972. Nonverbal communication. Transaction Publishers.
[13]
Sharon Mozgai, Gale Lucas, and Jonathan Gratch. 2017. To Tell the Truth: Virtual Agents and Morning Morality International Conference on Intelligent Virtual Agents. Springer, 283--286.
[14]
Najmeh Sadoughi and Carlos Busso. 2017. Joint learning of speech-driven facial motion with bidirectional long-short term memory International Conference on Intelligent Virtual Agents. Springer, 389--402.
[15]
Ari Shapiro. 2011. Building a character animation system. Motion in Games (2011), 98--109.
[16]
Joseph Weizenbaum. 1966. ELIZA - a computer program for the study of natural language communication between man and machine. Commun. ACM Vol. 9, 1 (1966), 36--45.
[17]
Joseph Weizenbaum. 1976. Computer power and human reason: From judgment to calculation. (1976).
[18]
Wojciech Zaremba, Ilya Sutskever, and Oriol Vinyals. 2014. Recurrent neural network regularization. arXiv preprint arXiv:1409.2329 (2014).
[19]
Geoffrey Zweig, Chengzhu Yu, Jasha Droppo, and Andreas Stolcke. 2017. Advances in all-neural speech recognition. In Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 4805--4809.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
AAMAS '18: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems
July 2018
2312 pages

Sponsors

In-Cooperation

Publisher

International Foundation for Autonomous Agents and Multiagent Systems

Richland, SC

Publication History

Published: 09 July 2018

Check for updates

Author Tags

  1. animation
  2. chatbot
  3. convolutional neural network
  4. neural language model
  5. virtual agent

Qualifiers

  • Research-article

Funding Sources

  • U.S. Army Research Laboratory

Conference

AAMAS '18
Sponsor:
AAMAS '18: Autonomous Agents and MultiAgent Systems
July 10 - 15, 2018
Stockholm, Sweden

Acceptance Rates

AAMAS '18 Paper Acceptance Rate 149 of 607 submissions, 25%;
Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)0
Reflects downloads up to 10 Oct 2024

Other Metrics

Citations

Cited By

View all

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media