Acoustic emotion recognition: A benchmark comparison of performances

230Citations
Citations of this article
138Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In the light of the first challenge on emotion recognition from speech we provide the largest-to-date benchmark comparison under equal conditions on nine standard corpora in the field using the two pre-dominant paradigms: modeling on a frame-level by means of Hidden Markov Models and supra-segmental modeling by systematic feature brute-forcing. Investigated corpora are the ABC, AVIC, DES, EMO-DB, eNTERFACE, SAL, SmartKom, SUSAS, and VAM databases. To provide better comparability among sets, we additionally cluster each database's emotions into binary valence and arousal discrimination tasks. In the result large differences are found among corpora that mostly stem from naturalistic emotions and spontaneous speech vs. more prototypical events. Further, supra-segmental modeling proves significantly beneficial on average when several classes are addressed at a time. © 2009 IEEE.

References Powered by Scopus

The eNTERFACE'05 Audio-Visual emotion database

590Citations
N/AReaders
Get full text

Emotional speech: Towards a new generation of databases

359Citations
N/AReaders
Get full text

The Vera am Mittag German audio-visual emotional speech database

349Citations
N/AReaders
Get full text

Cited by Powered by Scopus

The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing

1265Citations
N/AReaders
Get full text

Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge

589Citations
N/AReaders
Get full text

Evaluating deep learning architectures for Speech Emotion Recognition

464Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Schuller, B., Vlasenko, B., Eyben, F., Rigoll, G., & Wendemuth, A. (2009). Acoustic emotion recognition: A benchmark comparison of performances. In Proceedings of the 2009 IEEE Workshop on Automatic Speech Recognition and Understanding, ASRU 2009 (pp. 552–557). https://doi.org/10.1109/ASRU.2009.5372886

Readers over time

‘10‘11‘12‘13‘14‘15‘16‘17‘18‘19‘20‘21‘22‘23‘2405101520

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 78

80%

Researcher 12

12%

Professor / Associate Prof. 6

6%

Lecturer / Post doc 2

2%

Readers' Discipline

Tooltip

Computer Science 58

62%

Engineering 27

29%

Psychology 6

6%

Medicine and Dentistry 3

3%

Save time finding and organizing research with Mendeley

Sign up for free
0