↓ Skip to main content

The MPI Facial Expression Database — A Validated Database of Emotional and Conversational Facial Expressions

Overview of attention for article published in PLOS ONE, March 2012
Altmetric Badge

About this Attention Score

  • In the top 25% of all research outputs scored by Altmetric
  • High Attention Score compared to outputs of the same age (90th percentile)
  • High Attention Score compared to outputs of the same age and source (85th percentile)

Mentioned by

twitter
17 X users
facebook
1 Facebook page

Citations

dimensions_citation
123 Dimensions

Readers on

mendeley
195 Mendeley
You are seeing a free-to-access but limited selection of the activity Altmetric has collected about this research output. Click here to find out more.
Title
The MPI Facial Expression Database — A Validated Database of Emotional and Conversational Facial Expressions
Published in
PLOS ONE, March 2012
DOI 10.1371/journal.pone.0032321
Pubmed ID
Authors

Kathrin Kaulard, Douglas W. Cunningham, Heinrich H. Bülthoff, Christian Wallraven

Abstract

The ability to communicate is one of the core aspects of human life. For this, we use not only verbal but also nonverbal signals of remarkable complexity. Among the latter, facial expressions belong to the most important information channels. Despite the large variety of facial expressions we use in daily life, research on facial expressions has so far mostly focused on the emotional aspect. Consequently, most databases of facial expressions available to the research community also include only emotional expressions, neglecting the largely unexplored aspect of conversational expressions. To fill this gap, we present the MPI facial expression database, which contains a large variety of natural emotional and conversational expressions. The database contains 55 different facial expressions performed by 19 German participants. Expressions were elicited with the help of a method-acting protocol, which guarantees both well-defined and natural facial expressions. The method-acting protocol was based on every-day scenarios, which are used to define the necessary context information for each expression. All facial expressions are available in three repetitions, in two intensities, as well as from three different camera angles. A detailed frame annotation is provided, from which a dynamic and a static version of the database have been created. In addition to describing the database in detail, we also present the results of an experiment with two conditions that serve to validate the context scenarios as well as the naturalness and recognizability of the video sequences. Our results provide clear evidence that conversational expressions can be recognized surprisingly well from visual information alone. The MPI facial expression database will enable researchers from different research fields (including the perceptual and cognitive sciences, but also affective computing, as well as computer vision) to investigate the processing of a wider range of natural facial expressions.

X Demographics

X Demographics

The data shown below were collected from the profiles of 17 X users who shared this research output. Click here to find out more about how the information was compiled.
Mendeley readers

Mendeley readers

The data shown below were compiled from readership statistics for 195 Mendeley readers of this research output. Click here to see the associated Mendeley record.

Geographical breakdown

Country Count As %
Germany 4 2%
France 2 1%
United Kingdom 2 1%
Hungary 1 <1%
Australia 1 <1%
Spain 1 <1%
United States 1 <1%
Luxembourg 1 <1%
Unknown 182 93%

Demographic breakdown

Readers by professional status Count As %
Student > Ph. D. Student 46 24%
Researcher 23 12%
Student > Master 23 12%
Student > Bachelor 21 11%
Student > Doctoral Student 9 5%
Other 30 15%
Unknown 43 22%
Readers by discipline Count As %
Psychology 61 31%
Computer Science 35 18%
Neuroscience 10 5%
Engineering 10 5%
Medicine and Dentistry 8 4%
Other 23 12%
Unknown 48 25%
Attention Score in Context

Attention Score in Context

This research output has an Altmetric Attention Score of 13. This is our high-level measure of the quality and quantity of online attention that it has received. This Attention Score, as well as the ranking and number of research outputs shown below, was calculated when the research output was last mentioned on 25 September 2012.
All research outputs
#2,820,136
of 25,654,806 outputs
Outputs from PLOS ONE
#34,608
of 223,967 outputs
Outputs of similar age
#16,590
of 169,657 outputs
Outputs of similar age from PLOS ONE
#528
of 3,596 outputs
Altmetric has tracked 25,654,806 research outputs across all sources so far. Compared to these this one has done well and is in the 88th percentile: it's in the top 25% of all research outputs ever tracked by Altmetric.
So far Altmetric has tracked 223,967 research outputs from this source. They typically receive a lot more attention than average, with a mean Attention Score of 15.8. This one has done well, scoring higher than 84% of its peers.
Older research outputs will score higher simply because they've had more time to accumulate mentions. To account for age we can compare this Altmetric Attention Score to the 169,657 tracked outputs that were published within six weeks on either side of this one in any source. This one has done particularly well, scoring higher than 90% of its contemporaries.
We're also able to compare this research output to 3,596 others from the same source and published within six weeks on either side of this one. This one has done well, scoring higher than 85% of its contemporaries.