Comparison of Human and Computer Recognition of Facial Emotion

J.M. Susskind, J. Movellan, M.S. Bartlett, G. Littlewort, A.K. Anderson

Proceedings of the of the Cognitive Neuroscience Society
April 2004, New York, NY.


Abstract

One class of computational approaches to facial expression recognition focuses on mapping face images into six basic emotions: anger, disgust, fear, joy, sadness, and surprise. When tested in parallel with human behavioral data, these models can provide useful information about which attributes of visual input are used by humans to make judgments of emotional expression. Of particular interest are models in which similarity is not built into the model but is learned in the mapping from visual input to expression category. The model we are testing uses support vector machines to classify spatial-frequency filtered (Gabor) face images into the six basic emotions above plus neutral, without a predefined similarity schema or predefined facial features. The model achieves 93% accuracy on multiple datasets, and performs in real-time. Because the model makes a 7-way forced choice, it provides judgments that at times are mixtures between facial expressions, allowing the model to make implicit similarity judgments. Comparing model and human performance on generalizing to expressions such as contempt provides information about whether expressions not present in the training set can be represented as combinations of the visual features that make up the pre-defined categories. Expressing contempt as a mixture of basic emotions is of particular interest in the controversy over whether contempt is itself a basic emotion or a subset of emotions such as anger and disgust.