Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3242969.3264985acmotherconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
short-paper

An Attention Model for Group-Level Emotion Recognition

Published: 02 October 2018 Publication History

Abstract

In this paper we propose a new approach for classifying the global emotion of images containing groups of people. To achieve this task, we consider two different and complementary sources of information: i) a global representation of the entire image (ii) a local representation where only faces are considered. While the global representation of the image is learned with a convolutional neural network (CNN), the local representation is obtained by merging face features through an attention mechanism. The two representations are first learned independently with two separate CNN branches and then fused through concatenation in order to obtain the final group-emotion classifier. For our submission to the EmotiW 2018 group-level emotion recognition challenge, we combine several variations of the proposed model into an ensemble, obtaining a final accuracy of 64.83% on the test set and ranking 4th among all challenge participants.

References

[1]
Roland Goecke Abhinav Dhall, Amanjot Kaur and Tom Gedeon . 2018. EmotiW 2018: Audio-Video, Student Engagement and Group-Level Affect Prediction, ACM ICMI 2018. ACM ICMI.
[2]
Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio . 2014. Neural Machine Translation by Jointly Learning to Align and Translate. CoRR Vol. abs/1409.0473 (2014).
[3]
Kwang-Ho Choi, Junbeom Kim, Oh Sang Kwon, Min Kim, Yeon Hee Ryu, and Ji-Eun Park . 2017. Is heart rate variability (HRV) an adequate tool for evaluating human emotions? -- A focus on the use of the International Affective Picture System (IAPS). Psychiatry Research Vol. 251 (2017), 192--196.
[4]
J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei . 2009. ImageNet: A Large-Scale Hierarchical Image Database CVPR09.
[5]
A. Dhall, J. Joshi, K. Sikka, R. Goecke, and N. Sebe . 2015. The more the merrier: Analysing the affect of a group of people in images 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Vol. Vol. 1. 1--8.
[6]
Samira Ebrahimi Kahou, Vincent Michalski, Kishore Konda, Roland Memisevic, and Christopher Pal . 2015. Recurrent Neural Networks for Emotion Recognition in Video Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (ICMI '15). ACM, New York, NY, USA, 467--474.
[7]
Rohit Girdhar and Deva Ramanan . 2017. Attentional pooling for action recognition. In Advances in Neural Information Processing Systems. 34--45.
[8]
Xin Guo, Luisa F. Polan'ıa, and Kenneth E. Barner . 2017. Group-level Emotion Recognition Using Deep Models on Image Scene, Faces, and Skeletons. In Proceedings of the 19th ACM International Conference on Multimodal Interaction (ICMI 2017). ACM, New York, NY, USA, 603--608.
[9]
Gao Huang, Zhuang Liu, and Kilian Q. Weinberger . 2016. Densely Connected Convolutional Networks. CoRR Vol. abs/1608.06993 (2016). showeprint{arxiv}1608.06993deftempurl%http://arxiv.org/abs/1608.06993 tempurl
[10]
Sergey Ioffe and Christian Szegedy . 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. CoRR Vol. abs/1502.03167 (2015). showeprint{arxiv}1502.03167deftempurl%http://arxiv.org/abs/1502.03167 tempurl
[11]
N. Jaques, S. Taylor, A. Azaria, A. Ghandeharioun, A. Sano, and R. Picard . 2015. Predicting students' happiness from physiology, phone, mobility, and behavioral data 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), Vol. Vol. 00. 222--228.
[12]
Abubakrelsedik Karali, Ahmad Bassiouny, and Motaz El-Saban . 2016. Facial Expression Recognition in the Wild using Rich Deep Features. CoRR Vol. abs/1601.02487 (2016). showeprint{arxiv}1601.02487deftempurl%http://arxiv.org/abs/1601.02487 tempurl
[13]
Diederik P. Kingma and Jimmy Ba . 2014. Adam: A Method for Stochastic Optimization. CoRR Vol. abs/1412.6980 (2014). showeprint{arxiv}1412.6980deftempurl%http://arxiv.org/abs/1412.6980 tempurl
[14]
Ronak Kosti, Jose M Alvarez, Adria Recasens, and Agata Lapedriza . 2017. Emotion recognition in context. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15]
Weiyang Liu, Yandong Wen, Zhiding Yu, Ming Li, Bhiksha Raj, and Le Song . 2017. SphereFace: Deep Hypersphere Embedding for Face Recognition. CoRR Vol. abs/1704.08063 (2017). showeprint{arxiv}1704.08063deftempurl%http://arxiv.org/abs/1704.08063 tempurl
[16]
Weiyang Liu, Yandong Wen, Zhiding Yu, and Meng Yang . 2016. Large-margin Softmax Loss for Convolutional Neural Networks Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48 (ICML'16). JMLR.org, 507--516. deftempurl%http://dl.acm.org/citation.cfm?id=3045390.3045445 tempurl
[17]
Volodymyr Mnih, Nicolas Heess, Alex Graves, et almbox. . 2014. Recurrent models of visual attention. In Advances in neural information processing systems. 2204--2212.
[18]
Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov . 2014. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research Vol. 15 (2014), 1929--1958. deftempurl%http://jmlr.org/papers/v15/srivastava14a.html tempurl
[19]
Lianzhi Tan, Kaipeng Zhang, Kai Wang, Xiaoxing Zeng, Xiaojiang Peng, and Yu Qiao . 2017. Group Emotion Recognition with Individual Facial Emotion CNNs and Global Image Based CNNs. In Proceedings of the 19th ACM International Conference on Multimodal Interaction (ICMI 2017). ACM, New York, NY, USA, 549--552.
[20]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin . 2017. Attention Is All You Need. CoRR Vol. abs/1706.03762 (2017). showeprint{arxiv}1706.03762deftempurl%http://arxiv.org/abs/1706.03762 tempurl
[21]
Fei Wang, Mengqing Jiang, Chen Qian, Shuo Yang, Cheng Li, Honggang Zhang, Xiaogang Wang, and Xiaoou Tang . 2017. Residual Attention Network for Image Classification Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3156--3164.
[22]
Qinglan Wei, Yijia Zhao, Qihua Xu, Liandong Li, Jun He, Lejun Yu, and Bo Sun . 2017. A New Deep-learning Framework for Group Emotion Recognition Proceedings of the 19th ACM International Conference on Multimodal Interaction (ICMI 2017). ACM, New York, NY, USA, 587--592.
[23]
Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio . 2015. Show, attend and tell: Neural image caption generation with visual attention International conference on machine learning. 2048--2057.
[24]
Dong Yi, Zhen Lei, Shengcai Liao, and Stan Z. Li . 2014. Learning Face Representation from Scratch. CoRR Vol. abs/1411.7923 (2014). showeprint{arxiv}1411.7923deftempurl%http://arxiv.org/abs/1411.7923 tempurl
[25]
Kaipeng Zhang, Zhanpeng Zhang, Zhifeng Li, and Yu Qiao . 2016. Joint Face Detection and Alignment using Multi-task Cascaded Convolutional Networks. CoRR Vol. abs/1604.02878 (2016). showeprint{arxiv}1604.02878deftempurl%http://arxiv.org/abs/1604.02878 tempurl

Cited By

View all
  • (2024)Implementing the Affective Mechanism for Group Emotion Recognition With a New Graph Convolutional Network ArchitectureIEEE Transactions on Affective Computing10.1109/TAFFC.2023.332010115:3(1104-1115)Online publication date: Jul-2024
  • (2024)Group-Level Emotion Recognition Using Hierarchical Dual-Branch Cross Transformer with Semi-Supervised Learning2024 IEEE 4th International Conference on Software Engineering and Artificial Intelligence (SEAI)10.1109/SEAI62072.2024.10674336(252-256)Online publication date: 21-Jun-2024
  • (2023)Multimodal Group Emotion Recognition In-the-wild Using Privacy-Compliant FeaturesProceedings of the 25th International Conference on Multimodal Interaction10.1145/3577190.3616546(750-754)Online publication date: 9-Oct-2023
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICMI '18: Proceedings of the 20th ACM International Conference on Multimodal Interaction
October 2018
687 pages
ISBN:9781450356923
DOI:10.1145/3242969
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

  • SIGCHI: Specialist Interest Group in Computer-Human Interaction of the ACM

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 October 2018

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. attention mechanisms
  2. convolutional neural networks
  3. deep learning
  4. group-level emotion recognition

Qualifiers

  • Short-paper

Funding Sources

Conference

ICMI '18
Sponsor:
  • SIGCHI

Acceptance Rates

ICMI '18 Paper Acceptance Rate 63 of 149 submissions, 42%;
Overall Acceptance Rate 453 of 1,080 submissions, 42%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)18
  • Downloads (Last 6 weeks)2
Reflects downloads up to 02 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Implementing the Affective Mechanism for Group Emotion Recognition With a New Graph Convolutional Network ArchitectureIEEE Transactions on Affective Computing10.1109/TAFFC.2023.332010115:3(1104-1115)Online publication date: Jul-2024
  • (2024)Group-Level Emotion Recognition Using Hierarchical Dual-Branch Cross Transformer with Semi-Supervised Learning2024 IEEE 4th International Conference on Software Engineering and Artificial Intelligence (SEAI)10.1109/SEAI62072.2024.10674336(252-256)Online publication date: 21-Jun-2024
  • (2023)Multimodal Group Emotion Recognition In-the-wild Using Privacy-Compliant FeaturesProceedings of the 25th International Conference on Multimodal Interaction10.1145/3577190.3616546(750-754)Online publication date: 9-Oct-2023
  • (2023)A Self-Fusion Network Based on Contrastive Learning for Group Emotion RecognitionIEEE Transactions on Computational Social Systems10.1109/TCSS.2022.320224910:2(458-469)Online publication date: Apr-2023
  • (2023)Audio-Visual Automatic Group Affect AnalysisIEEE Transactions on Affective Computing10.1109/TAFFC.2021.310417014:2(1056-1069)Online publication date: 1-Apr-2023
  • (2023)Automatic Emotion Recognition for Groups: A ReviewIEEE Transactions on Affective Computing10.1109/TAFFC.2021.306572614:1(89-107)Online publication date: 1-Jan-2023
  • (2023)Cohesive Group Emotion Recognition using Deep Learning2023 26th ACIS International Winter Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD-Winter)10.1109/SNPD-Winter57765.2023.10466291(264-269)Online publication date: 14-Dec-2023
  • (2023)Cohesive Group Emotion Recognition using Deep Learning2023 IEEE/ACIS 8th International Conference on Big Data, Cloud Computing, and Data Science (BCD)10.1109/BCD57833.2023.10466291(264-269)Online publication date: 14-Dec-2023
  • (2023)A recent survey on perceived group sentiment analysisJournal of Visual Communication and Image Representation10.1016/j.jvcir.2023.10398897(103988)Online publication date: Dec-2023
  • (2023)Moving Beyond Benchmarks and Competitions: Towards Addressing Social Media Challenges in an Educational ContextDatenbank-Spektrum10.1007/s13222-023-00436-323:1(27-39)Online publication date: 24-Feb-2023
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media