--- license: cc-by-4.0 task_categories: - audio-classification --- # Acted Emotional Speech Dynamic Database v1.0 ## ABOUT AESDD v1.0 was created on October 2017 in the Laboratory of Electronic Media, School of Journalism and Mass Communications, Aristotle University of Thessaloniki, for the needs of Speech Emotion Recognition research of the Multidisciplinary Media & Mediated Communication Research Group (M3C, http://m3c.web.auth.gr/). It is a collection of utterances of emotional speech acted by professional actors. This version is the initial state of AESDD. The purpose of this project the continuous growth of the database through the collaborative effort of the M3C research group and theatrical teams. ## CREATION OF THE DATABASE For the creation of v.1 of the database, 5 (3 female and 2 male) professional actors were recorded. 19 utterances of ambiguous out of context emotional content were chosen. The actors acted these 19 utterances in every one of the 5 chosen emotions. One extra improvised utterance was added for every actor and emotion. The guidance of the actors and the choice of the final recordings were supervised by a scientific expert in dramatology. For some of the utterances, more that one takes were qualified. Consequently, around 500 utterances occured in the final database. UPDATE: Since the AESDD is dynamic by definition, more actors have been recorded and added, following the same naming scheme as described in the Section "ORGANISING THE DATABASE" ## CHOSEN EMOTIONS Five emotions were chosen: 1. a (anger) 2. d (disgust) 3. f (fear) 4. h (happiness) 5. s (sadness) ## ORGANISING THE DATABASE There are five folders, named after the five emotion classes. Every file name in the databased is in the following form: xAA (B) where - x is the first letter of the emotion (a--> anger, h--> happiness etc.) - AA is the number of the utterance (01,02...20) - B is the number of the speaker (1 --> 1st speaker, 2 --> 2nd speaker etc) e.g. 'a03 (4).wav' is the 3rd utterance spoken by the 4th speaker with anger In the case where two takes were qualified for the same utterance, they are distinguished with a lower case letter. e.g. 'f18 (5).wav' and 'f18 (5)b.wav' are two different versions of the 5th actor saying the 18th utterance with fear. ## References 1. Vryzas, N., Kotsakis, R., Liatsou, A., Dimoulas, C. A., & Kalliris, G. (2018). Speech emotion recognition for performance interaction. Journal of the Audio Engineering Society, 66(6), 457-467. 2. Vryzas, N., Matsiola, M., Kotsakis, R., Dimoulas, C., & Kalliris, G. (2018, September). Subjective Evaluation of a Speech Emotion Recognition Interaction Framework. In Proceedings of the Audio Mostly 2018 on Sound in Immersion and Emotion (p. 34). ACM.