Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1101149.1101300acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
Article

Multimodal affect recognition in learning environments

Published: 06 November 2005 Publication History

Abstract

We propose a multi-sensor affect recognition system and evaluate it on the challenging task of classifying interest (or disinterest) in children trying to solve an educational puzzle on the computer. The multimodal sensory information from facial expressions and postural shifts of the learner is combined with information about the learner's activity on the computer. We propose a unified approach, based on a mixture of Gaussian Processes, for achieving sensor fusion under the problematic conditions of missing channels and noisy labels. This approach generates separate class labels corresponding to each individual modality. The final classification is based upon a hidden random variable, which probabilistically combines the sensors. The multimodal Gaussian Process approach achieves accuracy of over 86%, significantly outperforming classification using the individual modalities, and several other combination schemes.

References

[1]
T. W. Chan and A. Baskin. Intelligent Tutoring Systems: At the Crossroads of Artificial Intelligence and Education, chapter 1: Learning companion systems. 1990.
[2]
C. Conati. Probabilistic assessment of user's emotions in educational games. Applied Artificial Intelligence, special issue on Merging Cognition and Affect in HCI, 16, 2002.
[3]
T. S. Huang, L. S. Chen, and H. Tao. Bimodal emotion recognition by man and machine. In ATR Workshop on Virtual Communication Environments, 1998.
[4]
R. A. Jacobs, M. I. Jordan, S. J. Nowlan, and G. E. Hinton. Adaptive mixtures of local experts. Neural Computation, 3:79--87, 1991.
[5]
A. Kapoor, H. Ahn, and R. W. Picard. Mixture of gaussian processes to combine multiple modalities. In Workshop on MCS, 2005.
[6]
A. Kapoor, S. Mota, and R. W. Picard. Towards a learning companion that recognizes affect. In AAAI Fall Symposium, Nov 2001.
[7]
A. Kapoor and R. W. Picard. Real-time, fully automatic upper facial feature tracking. In Automatic Face and Gesture Recognition, May 2002.
[8]
A. Kapoor, R. W. Picard, and Y. Ivanov. Probabilistic combination of multiple modalities to detect interest. In ICPR, August 2004.
[9]
J. Kittler, M. Hatef, R. P. W. Duin, and J. Matas. On combining classifiers. PAMI, 20(3):226--239, 1998.
[10]
D. J. Miller and L. Yan. Critic-driven ensemble classification. Signal Processing, 47(10), 1999.
[11]
T. P. Minka. Expectation propagation for approximate bayesian inference. In UAI, 2001.
[12]
S. Mota and R. W. Picard. Automated posture analysis for detecting learner's interest level. In CVPR Workshop on HCI, June 2003.
[13]
N. Oliver, A. Garg, and E. Horvitz. Layered representations for learning and inferring office activity from multiple sensory channels. In ICMI, 2002.
[14]
M. Pantic and L. J. M. Rothkrantz. Towards an affect-sensitive multimodal human-computer interaction. Proceedings of IEEE, 91(9), 2003.
[15]
R. W. Picard, E. Vyzas, and J. Healey. Toward machine emotional intelligence: Analysis of affective physiological state. PAMI, 2001.
[16]
K. Toyama and E. Horvitz. Bayesian modality fusion: Probabilistic integration of multiple vision algorithms for head tracking. In ACCV, 2000.

Cited By

View all
  • (2024)NapTune: Efficient Model Tuning for Mood Classification using Previous Night's Sleep Measures along with Wearable Time-seriesProceedings of the 26th International Conference on Multimodal Interaction10.1145/3678957.3685722(204-213)Online publication date: 4-Nov-2024
  • (2024)Toward Supporting Adaptation: Exploring Affect’s Role in Cognitive Load when Using a Literacy GameProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642150(1-17)Online publication date: 11-May-2024
  • (2024)Automatic Context-Aware Inference of Engagement in HMI: A SurveyIEEE Transactions on Affective Computing10.1109/TAFFC.2023.327870715:2(445-464)Online publication date: Apr-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MULTIMEDIA '05: Proceedings of the 13th annual ACM international conference on Multimedia
November 2005
1110 pages
ISBN:1595930442
DOI:10.1145/1101149
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 06 November 2005

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Conference

MM05

Acceptance Rates

MULTIMEDIA '05 Paper Acceptance Rate 49 of 312 submissions, 16%;
Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)73
  • Downloads (Last 6 weeks)6
Reflects downloads up to 12 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)NapTune: Efficient Model Tuning for Mood Classification using Previous Night's Sleep Measures along with Wearable Time-seriesProceedings of the 26th International Conference on Multimodal Interaction10.1145/3678957.3685722(204-213)Online publication date: 4-Nov-2024
  • (2024)Toward Supporting Adaptation: Exploring Affect’s Role in Cognitive Load when Using a Literacy GameProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642150(1-17)Online publication date: 11-May-2024
  • (2024)Automatic Context-Aware Inference of Engagement in HMI: A SurveyIEEE Transactions on Affective Computing10.1109/TAFFC.2023.327870715:2(445-464)Online publication date: Apr-2024
  • (2024)Deep Representation Learning for Multimodal Emotion Recognition Using Physiological SignalsIEEE Access10.1109/ACCESS.2024.343655612(106605-106617)Online publication date: 2024
  • (2024)Bioindicators of Attention Detection in Online Learning EnvironmentsHCI International 2024 Posters10.1007/978-3-031-61953-3_9(75-85)Online publication date: 1-Jun-2024
  • (2024)Robotics in Medical ScienceMechanical Engineering in Biomedical Applications10.1002/9781394175109.ch15(367-396)Online publication date: 2-Jan-2024
  • (2023)A Survey on Facial Emotion Identification using Deep Learning ModelsAdvances in Computational Intelligence in Materials Science10.53759/acims/978-9914-9946-9-8_3(12-16)Online publication date: 7-Jun-2023
  • (2023)An Experimental Platform for Real-Time Students Engagement Measurements from Video in STEM ClassroomsSensors10.3390/s2303161423:3(1614)Online publication date: 2-Feb-2023
  • (2023)Facial Emotion Recognition System2023 5th International Conference on Advances in Computing, Communication Control and Networking (ICAC3N)10.1109/ICAC3N60023.2023.10541680(1612-1617)Online publication date: 15-Dec-2023
  • (2023)Multimodal Assessment of Interest Levels in Reading: Integrating Eye-Tracking and Physiological SensingIEEE Access10.1109/ACCESS.2023.331126811(93994-94008)Online publication date: 2023
  • Show More Cited By

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media