Multimodal Emotion Recognition Based on Speech and Physiological Signals Using Deep Neural Networks

A suitable combination of data in a multimodal emotion recognition model allows conveying and combining each channel’s information to achieve a better recognition of the encoded emotion than would be possible using only a single modality and channel. In this paper, we focus on combining speech and p...

Full description

Saved in:
Bibliographic Details
Published inPattern Recognition. ICPR International Workshops and Challenges Vol. 12666; pp. 289 - 300
Main Authors Bakhshi, Ali, Chalup, Stephan
Format Book Chapter
LanguageEnglish
Published Switzerland Springer International Publishing AG 2021
Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text
ISBN9783030687793
3030687791
ISSN0302-9743
1611-3349
DOI10.1007/978-3-030-68780-9_25

Cover

More Information
Summary:A suitable combination of data in a multimodal emotion recognition model allows conveying and combining each channel’s information to achieve a better recognition of the encoded emotion than would be possible using only a single modality and channel. In this paper, we focus on combining speech and physiological signals to predict the arousal and valence levels of the emotional states of a person. We designed a neural network that can use the information from raw audio signals, electrocardiograms, heart rate variability, electro-dermal activity, and skin conductance levels, to predict emotional states. The proposed deep neural network architecture works as an end-to-end process, which means, neither any pre-processing of the input data nor post-processing of the prediction of the network was applied. Using the data of the modalities available in the publicly accessible part of the RECOLA database, we achieved results comparable to other state-of-the-art approaches.
ISBN:9783030687793
3030687791
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-030-68780-9_25