Selezione lingua
ItalianoInglese
Cerca nel sito
Area riservata
Tot. visite contenuti : 342622
Home Bollettino Tecnologie del linguaggio 2013/001T - Recognizing affect from speech prosody using hierarchical graphical models

PostHeaderIcon 2013/001T - Recognizing affect from speech prosody using hierarchical graphical models

Raul Fernandez, Rosalind Picard, Recognizing affect from speech prosody using hierarchical graphical models, Speech Communication, Volume 53, Issues 9–10, November–December 2011, Pages 1088-1103, ISSN 0167-6393

DOI: 10.1016/j.specom.2011.05.003.

Abstract: In this work we develop and apply a class of hierarchical directed graphical models on the task of recognizing affective categories from prosody in both acted and natural speech. A strength of this new approach is the integration and summarization of information using both local (e.g., syllable level) and global prosodic phenomena (e.g., utterance level). In this framework speech is structurally modeled as a dynamically evolving hierarchical model in which levels of the hierarchy are determined by prosodic constituency and contain parameters that evolve according to dynamical systems. The acoustic parameters have been chosen to reflect four main components of speech thought to reflect paralinguistic and affect-specific information: intonation, loudness, rhythm and voice quality. The work is first evaluated on a database of acted emotions and compared to human perceptual recognition of five affective categories where it achieves rates within nearly 10% of human recognition accuracy despite only focusing on prosody. The model is then evaluated on two different corpora of fully spontaneous, affectively-colored, naturally occurring speech between people: Call Home English and BT Call Center. Here the ground truth labels are obtained from examining the agreement of 29 human coders labeling arousal and valence. The best discrimination performance on the natural spontaneous speech, using only the prosody features, obtains a 70% detection rate with 30% false alarms when detecting high arousal negative valence speech in call centers.

Ultimo aggiornamento ( Lunedì 04 Marzo 2013 12:05 )