Emotion Detection from Speech and Text
Fecha
2018-11-21Autor
Antón, Josu
Carrilero, Mikel
Metadatos
Mostrar el registro completo del ítem
IberSPEECH 2018 21-23 November 2018, Barcelona, Spain : 68-71 (2018)
Resumen
The main goal of this work is to carry out automatic emo-tion detection from speech by using both acoustic and textualinformation. For doing that a set of audios were extracted froma TV show were different guests discuss about topics of currentinterest. The selected audios were transcribed and annotatedin terms of emotional status using a crowdsourcing platform.A 3-dimensional model was used to define an specific emo-tional status in order to pick up the nuances in what the speakeris expressing instead of being restricted to a predefined set ofdiscrete categories. Different sets of acoustic parameters wereconsidered to obtain the input vectors for a neural network. Torepresent each sequence of words, a models based on word em-beddings was used. Different deep learning architectures weretested providing promising results, although having a corpus ofa limited size.