Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3532213.3532245acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccaiConference Proceedingsconference-collections
research-article

Speech Emotion Recognition by using Philips Fingerprint and Spectral Entropy

Published: 13 July 2022 Publication History

Abstract

Speech Emotion Recognition (SER) has become an indispensable part of human-computer interaction. In order to obtain a higher recognition rate, in this paper, a joint feature based on Philips fingerprint and spectral entropy is used, and it is combined with some underlying features to perform speech emotion recognition on the four emotions of angry, neutral, happy, and sad. Firstly, the Mel Frequency Cepstral Coefficients (MFCC), Linear Prediction Coefficient (LPC), logarithmic amplitude-frequency characteristics, Philips fingerprints and spectral entropy are extracted for speech dataset. Secondly, Principal Component Analysis (PCA) is used to reduce the dimension of the extracted feature. Thirdly, SVM is used to train and classify the features after dimension reduction. The experimental results show that by comparing the recognition results with other feature combinations, the method proposed in this paper has good recognition results.

References

[1]
Sicheng Zhao, Guoli Jia, Jufeng Yang, Guiguang Ding and Kurt Keutzer. Emotion Recognition from Multiple Modalities: Fundamentals and methodologies. IEEE Signal Processing Magazine, 2021, 38(6): 59-73.
[2]
Chi-Chun Lee, Kusha Sridhar, Jeng-Lin Li, Wei-Cheng Lin, Bo-Hao Su, and Carlos Busso. Deep Representation Learning for Affective Speech Signal Analysis and Processing: Preventing unwanted signal disparities. IEEE Signal Processing Magazine, 2021, 38(6): 22-38
[3]
Juan Sebastián Gómez-Cañón, Estefanía Cano, Tuomas Eerola, Perfecto Herrera, Xiao Hu, Yi-Hsuan Yang, and Emilia Gómez. Music Emotion Recognition: Toward new, robust standards in personalized and context-sensitive applications. IEEE Signal Processing Magazine, 2021, 38(6): 106-114.
[4]
Yun Jin, Peng Song, Wenming Zheng, Li Zhao. Novel feature fusion method for speech emotion recognition based on multiple kernel learning. Journal of Southeast University (English Edition), 2013, 29(2): 129-133.
[5]
Naga Padmaja. Jagini, Dr. R. Rajeswar Rao. Exploring Emotion Specific Features for Emotion Recognition System using PCA Approach. International Conference on Intelligent Computing and Control Systems, 2017: 58-62, Madurai, India.
[6]
Xianxin Ke, Bin Cao, Jiaojiao Bai, Qichao Yu, Dezhi Yang. Speech Emotion Recognition Based on PCA and CHMM. IEEE 8th Joint International Information Technology and Artificial Intelligence Conference, 2019: 667-671, Chongqing, China.
[7]
Peng Song, Yun Jin, Cheng Cha, Li Zhao. Speech emotion recognition based on sparse feature transfer. Journal of Data Acquisition and Processing, 2016, 31(2): 325-330.
[8]
Chung-Hsien Wu, Wei-Bin Liang. Emotion Recognition of Affective Speech Based on Multiple Classifiers Using Acoustic-Prosodic Information and Semantic Labels. IEEE Transactions on Affective Computing, 2011, 2(1): 10-21.
[9]
Jiajia Yu, Yun Jin, Yong Ma, Fangqiu Jiang, Yanyan Dai. Original speech emotion recognition based on Sinc-Transformer model. Journal of Signal Processing, 2021, 37(10): 1880-1888.
[10]
Xixin Wu, Yuewen Cao, Hui Lu, Songxiang Liu, Disong Wang, Zhiyong Wu, Xunying Liu and Helen Meng. Speech Emotion Recognition Using Sequential Capsule Networks. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2021, 29: 3280-3291.
[11]
Sheguo Wang, Xuxiong Ling, Fuliang Zhang, Jianing Tong. Speech Emotion Recognition Based on Principal Component Analysis and Back Propagation Neural Network. International Conference on Measuring Technology and Mechatronics Automation, 2010: 437-441, Changsha, China.
[12]
Mrinmoy Bhattacharjee, S.R.M. Prasanna and Prithwijit Guha. Speech/Music Classification Using Features fromSpectral Peaks. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2020, 28: 1549-1559.
[13]
Haitsma J, Kalker T. A highly robust audio fingerprinting system. 3rd International Conference on Music Information Retrieval, 2002:107-115, Paris, France.
[14]
F. Burkhardt, A. Paeschke, M. Rolfes, W. Sendlmeier and B.Weiss. A database of German emotional speech. Interspeech 2005: 1517-1520, Lisbon, Portugal.
[15]
C. Busso IEMOCAP: Interactive emotional dyadic motion capture Database. Lang. Resour. Eval., 2008, 42(4): 335-359.
  1. Speech Emotion Recognition by using Philips Fingerprint and Spectral Entropy

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICCAI '22: Proceedings of the 8th International Conference on Computing and Artificial Intelligence
    March 2022
    809 pages
    ISBN:9781450396110
    DOI:10.1145/3532213
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 July 2022

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    Conference

    ICCAI '22

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 33
      Total Downloads
    • Downloads (Last 12 months)4
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 09 Nov 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media