Abstract
Emotion recognition plays a crucial role in computer science, particularly in enhancing human-computer interactions. The process of emotion labeling remains time-consuming and costly, thereby impeding efficient dataset creation. Recently, large language models (LLMs) have demonstrated adaptability across a variety of tasks without requiring task-specific training. This indicates the potential of LLMs to recognize emotions even with fewer emotion labels. Therefore, we assessed the performance of an LLM in emotion recognition using two established datasets: MELD and IEMOCAP. Our findings reveal that for emotion labels with few training samples, the performance of the LLM approaches or even exceeds that of SPCL, a leading model specializing in text-based emotion recognition. In addition, inspired by the Chain of Thought, we incorporated a multi-step prompting technique into the LLM to further enhance its discriminative capacity between emotion labels. The results underscore the potential of LLMs to reduce the time and costs of emotion data labeling.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ahmed, N., Aghbari, Z.A., Girija, S.: A systematic survey on multimodal emotion recognition using learning algorithms. Intell. Syst. Appl. 17, 200171 (2023). https://doi.org/10.1016/j.iswa.2022.200171
Brown, T.B., et al. (eds.): Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, 6–12 December 2020, Virtual (2020). https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html
Busso, C., et al.: IEMOCAP: interactive emotional dyadic motion capture database. Lang. Resour. Eval. 42, 335–359 (2008). https://doi.org/10.1007/s10579-008-9076-6
Chudasama, V., Kar, P., Gudmalwar, A., Shah, N., Wasnik, P., Onoe, N.: M2FNet: multi-modal fusion network for emotion recognition in conversation. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2022, New Orleans, LA, USA, 19–20 June 2022, pp. 4651–4660. IEEE (2022). https://doi.org/10.1109/CVPRW56347.2022.00511
Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019, Volume 1 (Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/n19-1423
Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., Iwasawa, Y.: Large language models are zero-shot reasoners. In: NeurIPS (2022). http://papers.nips.cc/paper_files/paper/2022/hash/8bb0d291acd4acf06ef112099c16f326-Abstract-Conference.html
Li, Z., Tang, F., Zhao, M., Zhu, Y.: EmoCaps: emotion capsule based model for conversational emotion recognition. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022, pp. 1610–1618. Association for Computational Linguistics (2022). https://doi.org/10.18653/v1/2022.findings-acl.126
Poria, S., Hazarika, D., Majumder, N., Naik, G., Cambria, E., Mihalcea, R.: MELD: a multimodal multi-party dataset for emotion recognition in conversations. In: Korhonen, A., Traum, D.R., Màrquez, L. (eds.) Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019, Volume 1: Long Papers, pp. 527–536. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/p19-1050
Poria, S., Majumder, N., Mihalcea, R., Hovy, E.H.: Emotion recognition in conversation: research challenges, datasets, and recent advances. IEEE Access 7, 100943–100953 (2019). https://doi.org/10.1109/ACCESS.2019.2929050
Qiao, S., et al.: Reasoning with language model prompting: a survey. In: Rogers, A., Boyd-Graber, J.L., Okazaki, N. (eds.) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, 9–14 July 2023, pp. 5368–5393. Association for Computational Linguistics (2023). https://doi.org/10.18653/v1/2023.acl-long.294
Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Lear. Res. 21(1), 5485–5551 (2020). https://dl.acm.org/doi/abs/10.5555/3455716.3455856
Song, X., Huang, L., Xue, H., Hu, S.: Supervised prototypical contrastive learning for emotion recognition in conversation. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022, pp. 5197–5206. Association for Computational Linguistics (2022). https://doi.org/10.18653/v1/2022.emnlp-main.347
Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. In: NeurIPS (2022). http://papers.nips.cc/paper_files/paper/2022/hash/9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference.html
Zahiri, S.M., Choi, J.D.: Emotion detection on TV show transcripts with sequence-based convolutional neural networks. In: The Workshops of the the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, Louisiana, USA, 2–7 February 2018. AAAI Technical Report, vol. WS-18, pp. 44–52. AAAI Press (2018). https://aaai.org/ocs/index.php/WS/AAAIW18/paper/view/16434
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Hama, K., Otsuka, A., Ishii, R. (2024). Emotion Recognition in Conversation with Multi-step Prompting Using Large Language Model. In: Coman, A., Vasilache, S. (eds) Social Computing and Social Media. HCII 2024. Lecture Notes in Computer Science, vol 14703. Springer, Cham. https://doi.org/10.1007/978-3-031-61281-7_24
Download citation
DOI: https://doi.org/10.1007/978-3-031-61281-7_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-61280-0
Online ISBN: 978-3-031-61281-7
eBook Packages: Computer ScienceComputer Science (R0)