Content area
Full Text
Lang Resources & Evaluation (2008) 42:335359 DOI 10.1007/s10579-008-9076-6
Carlos Busso Murtaza Bulut Chi-Chun Lee Abe Kazemzadeh Emily Mower Samuel Kim Jeannette N. Chang Sungbok Lee Shrikanth S. Narayanan
Published online: 5 November 2008 Springer Science+Business Media B.V. 2008
Abstract Since emotions are expressed through a combination of verbal and non-verbal channels, a joint analysis of speech and gestures is required to understand expressive human communication. To facilitate such investigations, this paper describes a new corpus named the interactive emotional dyadic motion capture database (IEMOCAP), collected by the Speech Analysis and Interpretation Laboratory (SAIL) at the University of Southern California (USC). This database was recorded from ten actors in dyadic sessions with markers on the face, head, and hands, which provide detailed information about their facial expressions and hand movements during scripted and spontaneous spoken communication scenarios. The actors performed selected emotional scripts and also improvised hypothetical scenarios designed to elicit specic types of emotions (happiness, anger, sadness, frustration and neutral state). The corpus contains approximately 12 h of data. The detailed motion capture information, the interactive setting to elicit authentic emotions, and the size of the database make this corpus a valuable addition to the existing databases in the community for the study and modeling of multimodal and expressive human communication.
Keywords Audio-visual database Dyadic interaction Emotion
Emotional assessment Motion capture system
1 Introduction
One of the most interesting paralinguistic messages expressed during human interaction is the emotional state of the subjects, which is conveyed through both
C. Busso (&) M. Bulut C.-C. Lee A. Kazemzadeh E. Mower S. Kim
J. N. Chang S. Lee S. S. Narayanan
Speech Analysis and Interpretation Laboratory (SAIL), University of Southern California, Los Angeles, CA 90089, USAe-mail: [email protected]
IEMOCAP: interactive emotional dyadic motion capture database
123
336 C. Busso et al.
speech and gestures. The tone and energy of the speech, facial expressions, torso posture, head position, hand gestures, and gaze are all combined in a nontrivial manner, as they unfold during natural human communication. These communicative channels need to be jointly studied if robust emotional models are to be developed and implemented.
In this context, one of the major limitations in the study of emotion expression is the lack of databases with genuine interaction that comprise integrated information...