Speech Emotion Recognition Based on Voice Fundamental Frequency

The human voice is one of the basic means of communication, thanks to which one also can easily convey the emotional state. This paper presents experiments on emotion recognition in human speech based on the fundamental frequency. AGH Emotional Speech Corpus was used. This database consists of audio...

Full description

Saved in:
Bibliographic Details
Published inArchives of acoustics Vol. 44; no. 2; pp. 277 - 286
Main Authors Dimitrova-Grekow, Teodora, Klis, Aneta, Igras-Cybulska, Magdalena
Format Journal Article
LanguageEnglish
Published Warsaw Polish Academy of Sciences 01.01.2019
Institute of Fundamental Technological Research Polish Academy of Sciences
Subjects
Online AccessGet full text
ISSN0137-5075
2300-262X
2300-262X
DOI10.24425/aoa.2019.128491

Cover

More Information
Summary:The human voice is one of the basic means of communication, thanks to which one also can easily convey the emotional state. This paper presents experiments on emotion recognition in human speech based on the fundamental frequency. AGH Emotional Speech Corpus was used. This database consists of audio samples of seven emotions acted by 12 different speakers (6 female and 6 male). We explored phrases of all the emotions – all together and in various combinations. Fast Fourier Transformation and magnitude spectrum analysis were applied to extract the fundamental tone out of the speech audio samples. After extraction of several statistical features of the fundamental frequency, we studied if they carry information on the emotional state of the speaker applying different AI methods. Analysis of the outcome data was conducted with classifiers: K-Nearest Neighbours with local induction, Random Forest, Bagging, JRip, and Random Subspace Method from algorithms collection for data mining WEKA. The results prove that the fundamental frequency is a prospective choice for further experiments.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0137-5075
2300-262X
2300-262X
DOI:10.24425/aoa.2019.128491