Academic Journals Database
Disseminating quality controlled scientific knowledge

Utterance independent bimodal emotion recognition in spontaneous communication

ADD TO MY LIST
 
Author(s): Tao Jianhua | Pan Shifeng | Yang Minghao | Li Ya | Mu Kaihui | Che Jianfeng

Journal: EURASIP Journal on Advances in Signal Processing
ISSN 1687-6172

Volume: 2011;
Issue: 1;
Start page: 4;
Date: 2011;
Original page

Keywords: Bimodal emotion recognition | Utterance Independent | Multistream Hidden Markov Model | Fused Hidden Markov Model Inversion

ABSTRACT
Abstract Emotion expressions sometimes are mixed with the utterance expression in spontaneous face-to-face communication, which makes difficulties for emotion recognition. This article introduces the methods of reducing the utterance influences in visual parameters for the audio-visual-based emotion recognition. The audio and visual channels are first combined under a Multistream Hidden Markov Model (MHMM). Then, the utterance reduction is finished by finding the residual between the real visual parameters and the outputs of the utterance related visual parameters. This article introduces the Fused Hidden Markov Model Inversion method which is trained in the neutral expressed audio-visual corpus to solve the problem. To reduce the computing complexity the inversion model is further simplified to a Gaussian Mixture Model (GMM) mapping. Compared with traditional bimodal emotion recognition methods (e.g., SVM, CART, Boosting), the utterance reduction method can give better results of emotion recognition. The experiments also show the effectiveness of our emotion recognition system when it was used in a live environment.
RPA Switzerland

Robotic Process Automation Switzerland

    

Tango Rapperswil
Tango Rapperswil