4.6 Article

Multimodal information fusion application to human emotion recognition from face and speech

Journal

MULTIMEDIA TOOLS AND APPLICATIONS
Volume 49, Issue 2, Pages 277-297

Publisher

SPRINGER
DOI: 10.1007/s11042-009-0344-2

Keywords

Multimodal feature extraction; Multimodal information fusion; Human computer interaction; Multimodal emotion recognition

Funding

  1. Iran Telecommunication Research Center (ITRC) [T500/20592]

Ask authors/readers for more resources

A multimedia content is composed of several streams that carry information in audio, video or textual channels. Classification and clustering multimedia contents require extraction and combination of information from these streams. The streams constituting a multimedia content are naturally different in terms of scale, dynamics and temporal patterns. These differences make combining the information sources using classic combination techniques difficult. We propose an asynchronous feature level fusion approach that creates a unified hybrid feature space out of the individual signal measurements. The target space can be used for clustering or classification of the multimedia content. As a representative application, we used the proposed approach to recognize basic affective states from speech prosody and facial expressions. Experimental results over two audiovisual emotion databases with 42 and 12 subjects revealed that the performance of the proposed system is significantly higher than the unimodal face based and speech based systems, as well as synchronous feature level and decision level fusion approaches.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available