Multimodal Speech Emotion Recognition Using Modality-Specific Self-Supervised Frameworks

Rutherford Agbeshi Patamia, Paulo E. Santos, Kingsley Nketia Acheampong, Favour Ekong, Kwabena Sarpong, She Kun

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Citations (Scopus)

Abstract

Emotion recognition is a topic of significant interest in assistive robotics due to the need to equip robots with the ability to comprehend human behavior, facilitating their effective interaction in our society. Consequently, efficient and dependable emotion recognition systems supporting optimal human-machine communication are required. Multi-modality (including speech, audio, text, images, and videos) is typically exploited in emotion recognition tasks. Much relevant research is based on merging multiple data modalities and training deep learning models utilizing low-level data representations. However, most existing emotion databases are not large (or complex) enough to allow machine learning approaches to learn detailed representations. This paper explores modality-specific pre-trained transformer frameworks for self-supervised learning of speech and text representations for data-efficient emotion recognition while achieving state-of-the-art performance in recognizing emotions. This model applies feature-level fusion using nonverbal cue data points from motion capture to provide multimodal speech emotion recognition. The model was trained using the publicly available IEMOCAP dataset, achieving an overall accuracy of 77.58% for four emotions, outperforming state-of-the-art approaches
Original languageEnglish
Title of host publication2023 IEEE International Conference on Systems, Man, and Cybernetics
Subtitle of host publicationImproving the Quality of Life, SMC 2023 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers
Pages4134-4141
Number of pages8
ISBN (Electronic)9798350337020
ISBN (Print)979-8-3503-3702-0
DOIs
Publication statusPublished - 29 Jan 2024
Event2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC) - Honolulu, Oahu, HI, USA, Honolulu, United States
Duration: 1 Oct 2023 → …
https://ieeexplore.ieee.org/abstract/document/10394418/authors#authors

Publication series

NameConference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
ISSN (Print)1062-922X

Conference

Conference2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC)
Country/TerritoryUnited States
CityHonolulu
Period1/10/23 → …
Internet address

Keywords

  • Training
  • Emotion recognition
  • Speech recognition
  • Computer architecture
  • Feature extraction
  • Motion capture
  • Data models

Fingerprint

Dive into the research topics of 'Multimodal Speech Emotion Recognition Using Modality-Specific Self-Supervised Frameworks'. Together they form a unique fingerprint.

Cite this