4.7 Article

AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild

Journal

IEEE TRANSACTIONS ON AFFECTIVE COMPUTING
Volume 10, Issue 1, Pages 18-31

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TAFFC.2017.2740923

Keywords

Affective computing in the wild; facial expressions; continuous dimensional space; valence; arousal

Funding

  1. US National Science Foundation [IIS-1111568, CNS-1427872]
  2. Division Of Computer and Network Systems
  3. Direct For Computer & Info Scie & Enginr [1427872] Funding Source: National Science Foundation
  4. Div Of Information & Intelligent Systems
  5. Direct For Computer & Info Scie & Enginr [1111568] Funding Source: National Science Foundation

Ask authors/readers for more resources

Automated affective computing in the wild setting is a challenging problem in computer vision. Existing annotated databases of facial expressions in the wild are small and mostly cover discrete emotions (aka the categorical model). There are very limited annotated facial databases for affective computing in the continuous dimensional model (e.g., valence and arousal). To meet this need, we collected, annotated, and prepared for public distribution a new database of facial emotions in the wild (called AffectNet). AffectNet contains more than 1,000,000 facial images from the Internet by querying three major search engines using 1,250 emotion related keywords in six different languages. About half of the retrieved images were manually annotated for the presence of seven discrete facial expressions and the intensity of valence and arousal. AffectNet is by far the largest database of facial expression, valence, and arousal in the wild enabling research in automated facial expression recognition in two different emotion models. Two baseline deep neural networks are used to classify images in the categorical model and predict the intensity of valence and arousal. Various evaluation metrics show that our deep neural network baselines can perform better than conventional machine learning methods and off-the-shelf facial expression recognition systems.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available