4.7 Article

Multi-View Facial Expression Recognition Based on Group Sparse Reduced-Rank Regression

Journal

IEEE TRANSACTIONS ON AFFECTIVE COMPUTING
Volume 5, Issue 1, Pages 71-85

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TAFFC.2014.2304712

Keywords

Multi-view facial expression recognition; group sparse reduced-rank regression (GSRRR); sparse reduced-rank regression model (SRRR); reduced-rank regression model (RRR)

Funding

  1. National Basic Research Program of China [2011CB302202]
  2. National Natural Science Foundation of China [61231002, 61073137]
  3. Natural Science Foundation of Jiangsu province [BK20130020]
  4. PhD Program Foundation of Ministry Education of China [20120092110054]
  5. Program for Distinguished Talents of Six Domains in Jiangsu [2010-DZ088]

Ask authors/readers for more resources

In this paper, a novel multi-view facial expression recognition method is presented. Different from most of the facial expression methods that use one view of facial feature vectors in the expression recognition, we synthesize multi-view facial feature vectors and combine them to this goal. In the facial feature extraction, we use the grids with multi-scale sizes to partition each facial image into a set of sub regions and carry out the feature extraction in each sub region. To deal with the prediction of expressions, we propose a novel group sparse reduced-rank regression (GSRRR) model to describe the relationship between the multi-view facial feature vectors and the corresponding expression class label vectors. The group sparsity of GSRRR enables us to automatically select the optimal sub regions of a face that contribute most to the expression recognition. To solve the optimization problem of GSRRR, we propose an efficient algorithm using inexact augmented Lagrangian multiplier (ALM) approach. Finally, we conduct extensive experiments on both BU-3DFE and Multi-PIE facial expression databases to evaluate the recognition performance of the proposed method. The experimental results confirm better recognition performance of the proposed method compared with the state of the art methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available