The Racially Diverse Affective Expression (radiate) Face Stimulus Set
May I. Conley
Danielle V. Dellarco
Estée Rubien-Thomas
Alexandra O. Cohen
Alessandra Cervera
SimpleOriginal

Summary

The Diverse face database (RADIATE) was created to improve understanding of racialized experience and emotion, addressing the lack of diversity in research materials.

2018

The Racially Diverse Affective Expression (radiate) Face Stimulus Set

Keywords emotion; multiracial; facial expressions; diversity; RADIATE; reliability; stimuli; validity

Abstract

Faces are often used in psychological and neuroimaging research to assess perceptual and emotional processes. Most available stimulus sets, however, represent minimal diversity in both race and ethnicity, which may confound understanding of these processes in diverse/racially heterogeneous samples. Having a diverse stimulus set of faces and emotional expressions could mitigate these biases and may also be useful in research that specifically examines the effects of race and ethnicity on perceptual, emotional and social processes. The racially diverse affective expression (RADIATE) face stimulus set is designed to provide an open-access set of 1,721 facial expressions of Black, White, Hispanic and Asian adult models. Moreover, the diversity of this stimulus set reflects census data showing a change in demographics in the United States from a white majority to a nonwhite majority by 2020. Psychometric results are provided describing the initial validity and reliability of the stimuli based on judgments of the emotional expressions.

1. Introduction

Census data (2010, 2014) show a changing racial and ethnic distribution in the United States (U.S.) from a White, non-Hispanic Caucasian majority to a non-White majority (Colby and Ortman, 2014). It is predicted that by 2020, more than half of the children in the U.S. will be part of a non-White race or ethnic group. Since faces play a major role in communicating emotional and social information, facial expressions are often used as stimuli in psychological research. The number of studies published on face processing has more than tripled over the past decade from 5,000 (Tottenham et al., 2009) to over 15,000 based on a simple PubMed (2017) search of the terms: “face perception”, “face processing”, and “face expression”. Yet, psychological experiments presenting emotional faces often use stimuli from predominantly (or all) White face stimulus sets (e.g., Ekman and Friesen, 1976, Ekman and Friesen, 1978, Russell and Bullock, 1985, Lundqvist et al., 1998). A growing literature of empirical studies shows that stimuli that portray individuals from minority and ethnic groups can have profound effects on perceptions and actions (Eberhardt et al., 2004, Funk et al., 2016). Face stimuli that reflect the racial and ethnic demographics of study participants are often needed for studies that use faces to examine these perceptual, emotional, and social processes. Diverse stimuli also afford testing of racial minority-majority effects on psychological processes. This paper introduces the open-access Racially Diverse Affective Expression (RADIATE) Face Stimulus Set (http://fablab.yale.edu/page/assays-tools ) from over 100 racially and ethnically diverse adult models presented with initial validity and reliability scores.

Because facial expressions play a central role in communicating emotional and social information, they have been used broadly in psychological experiments. Several factors must be considered when selecting a stimulus set of emotional faces including model attributes (age, gender, race), total number of stimuli available, overall quality of the stimuli (resolution and image clarity, lighting, perceptual differences across models, etc.), fees or restrictions in use of the stimuli, and the range of emotions portrayed in the stimuli.

Initial stimulus sets of facial expressions were developed to examine processing and production of emotions (Ekman and Friesen, 1976, Ekman and Friesen, 1978, Russell and Bullock, 1985), however many of the earliest sets pose some potential limitations for researchers due to the narrow number of emotions depicted, number of models, or homogeneity of race and ethnicity (Erwin et al., 1992, Phillips et al., 1998, Winston et al., 2002). To address some of these limitations researchers have continued to create and standardize face stimuli (Hart et al., 2000, Kanade et al., 2000, Phelps et al., 2000, Gur et al., 2002, Batty and Taylor, 2003, Phelps et al., 2003, Tanaka et al., 2004, Pantic et al., 2005, Ashraf et al., 2009, Samuelsson et al., 2012, Dalrymple et al., 2013, Giuliani et al., 2017).

The growing use of face stimuli in psychological research has led to the creation of many stimulus sets, each with its own specific attributes (e.g., expanded set of models, ages, photographed angles, emotions, races and ethnicities). One of the largest available is the Karolinska Directed Emotional Faces (KDEF) Database (Lundqvist et al., 1998) which consists of 70 adult models each posing 7 different emotional expressions (Angry, Fearful, Disgusted, Sad, Happy, Surprised, and Neutral), photographed from 5 different angles, yielding a robust number of unique images (n = 4900). However, only a subgroup of these stimuli has been validated (n = 490) (Goeleven et al., 2008) and all models are of European descent. Another large database, the Tarr Lab face database (Righi et al., 2012, Tarr, 2013) provides researchers with stimuli from over 200 Asian, African–American, Caucasian, Hispanic and Multiracial adult models. This large database is available in standard resolution movie stills or movie (mpeg) format and provides 3 different angles and 8 different emotions of mostly Caucasian or Asian models (68%).

More recently, several large stimulus sets have been developed that include child faces. The Radboud Faces Database (Langner et al., 2010) contains 1176 faces of children and adults, the Dartmouth Database of Children's Faces (Dalrymple et al., 2013) contains 3,200 images of children's faces (aged 6–16), and the Child Affective Facial Expression (CAFE) set (LoBue and Thrasher, 2015) includes 1,192 emotional expressions from a highly diverse sample of child models (50% non-White stimuli, aged 2–8). Uniquely, the University of Oregon Emotional Expression Stimulus (DuckEES) set provides dynamic, moving stimuli of children and adolescents ages 8–18 (Giuliani et al., 2017). An advantage of these datasets are the inclusion of child models, however like KDEF, these databases include only or majority Caucasian models. The most recently published Developmental Emotional Faces Stimulus Set (DEFSS) similarly includes child models as well as adult models, and has some diversity, but the majority of the stimuli (87%) feature Caucasian models (Meuwissen et al., 2017).

Diversity in face stimuli is important for experiments that involve research participants from minority and ethnic groups that are growing in the U.S. (Colby and Ortman, 2014) because stimuli that portray individuals from racial in- versus out-groups can have significant effects on psychological processes and actions (Hart et al., 2000, Phelps et al., 2000, Golby et al., 2001, Elfenbein and Ambady, 2002, Lieberman et al., 2005, Herrmann et al., 2007, Rubien-Thomas et al., 2016). Face stimuli with diversity in both race and ethnicity are imperative for teasing apart perceptual, cognitive, and social processes and avoiding confounds due to using predominantly in- or all out-group stimuli.

Stimulus sets that include representation of non-Caucasian adult models have been generated, but many of these sets are still predominantly homogenous with limited diversity (Ekman and Matsumoto, 1993-2004; Mandal, 1987, Wang and Markham, 1999, Beaupre and Hess, 2005, Mandal et al., 2001). The Chicago Face Database (CFD), however, has a large set of standardized racially and ethnically diverse adult stimuli with widespread norming data for neutral expressions (e.g. face size, pupil size, attractiveness), but the number of expressions available for each model varies (Ma et al., 2015). Another database, the NimStim set of adult facial expressions, provides images of 43 models posing 16 expressions that include open- and closed-mouth variants of happy, sad, angry, fear, disgust, surprise, calm, and neutral faces (n = 672). This sample includes approximately 40% non-White faces (10 African-American, 6 Asian-American, and 2 Latino-American models) in addition to validity and reliability scores for all 16 emotional expressions (Tottenham et al., 2009) laying the foundation for the current RADIATE face stimulus set.

The RADIATE face stimulus set consists of over 1,700 unique photographs of over 100 racially and ethnically diverse models (25% non-Hispanic White and 75% minority or ethnic group). Each model posed 16 different facial expressions (Tottenham et al., 2009) providing a wide range of emotions in a racially and ethnically diverse stimulus set. Images from the RADIATE stimulus set were developed so that they could be combined with other face stimulus sets with a standardized scarf-template that is included with the face stimuli (see Supplemental Text). The inclusion of the scarf-template affords researchers more flexibility in the selection and number of racially diverse stimuli by providing a standardized tool for harmonizing stimuli from other databases. Thus, the RADIATE stimulus set overcomes potential limitations within pre-existing stimulus packages, by providing a large, standardized set of 16 emotional expressions by racially and ethnically diverse models to help move the field forward in the consideration of race and ethnicity effects on psychological processes.

2. Methods

2.1. Participants

Participants (n = 693) within the United States were recruited using Amazon's Mechanical Turk (MTurk). All data from 31 participants were removed due to incomplete surveys or having duplicate IP addresses in our attempt to avoid repeat ratings from the same MTurk participants on a given survey. The final sample included 662 participants (260 female, 402 males) with a mean age of 27.6 (Range = 18–35, SD = 3.8) and self-identified as Asian (n = 48), Black/ African American (n = 70), Caucasian (n = 470), Hispanic (n = 63), and Other or Mixed races (n = 11). All participants provided consent approved by the institutional review board at Yale University and were paid for taking part in the experiment.

2.2. Stimuli

A collection of 1721 unique photographs of 109 adults (56 female, 53 male; 18–30 years old) posing 8 expressions (angry, calm, disgust, fear, happy, neutral, sad, and surprise) with open- and closed- mouth variants was obtained from the New York City metropolitan area in 2016 and harmonized and subsequently rated in 2016–2017. Twenty-three expressions were not captured in instances where models were unable to pose less-common expressions (e.g. sad open-mouth). Models were recruited from a community sample and were Asian (n = 22), Black/African American (n = 38), Caucasian (n = 28), Hispanic or Latino (n = 20) and Other (n = 1). All models consented to be photographed and released their photos to be used for research and scientific purposes, and all models were paid for their time.

Models were trained to make eight emotional expressions with open- and closed- mouth variants, given examples of each expression, and time to practice posing in a mirror before being photographed. See Supplemental Table 1 for instructions given to models for posing each facial expression. Models provided open- and closed-mouth variants of all expressions except the emotion of surprise for which only an open-mouthed expression was photographed. In addition, three versions of happy (closed-mouth, open-mouth, and high arousal open-mouth/exuberant) were obtained, based on prior work (Tottenham et al., 2009). Prior to being photographed, models were asked to remove any accessories that would visually separate them from the other models (i.e. glasses, headbands, hats). Models were photographed against a white wall and draped with a white scarf to hide clothing and reduce any potential reflected hues on the models’ faces. Adobe Photoshop was used to correct for differences in luminosity, head size, and head position (See Supplemental Text for additional details on standardization of stimuli). The scarf mask and the stimuli are available in both black and white and color as supplemental materials and at http://fablab.yale.edu/page/assays-tools.

2.3. Stimulus rating procedure

After editing and standardizing the images, the stimuli were uploaded to Qualtrics to be rated by MTurk participants. Surveys were subdivided into MTurk surveys3 that contained subsets (emotional expressions from 8 to 9 models) of all 1721 images to keep rating time to 20–30 min to be consistent with other facial stimuli rating tasks hosted in MTurk (Ma et al., 2015). Each stimulus was rated by an average of 50.92 MTurk participants (S.D .= 1.26, Range = 49–53) and detailed demographic information for each survey can be found in Supplemental Table 5. Each MTurk task contained links to 3 Qualtrics surveys. The first survey contained images of all expressions posed by 8–9 models presented in a randomized order and instructed participants to select the emotion (angry, calm, disgust, fear, happy, neutral, sad, surprise, or “none of the above”) being depicted in the image. To prevent subjects from immediately re-rating stimuli for re-test reliability, the second survey contained a simple cognitive task, which asked participants to list as many words as possible starting with a given letter or within a given object-category (e.g. vegetable, animal). Participants had a minute for each category and a total of 3 letters and 3 objects were given to each participant. Upon completion of the cognitive task, the third survey was administered, presenting the exact same images from the first survey in a new randomized order. The first and third surveys were self-paced. After each survey, participants were given a code that was pasted into a box to confirm completion of all surveys and paid.

2.3.1. Validity ratings

Initial validity ratings were obtained in the first of the two surveys by having participants select one of 8 emotion-labels (angry, calm, disgust, fear, happy, neutral, sad, surprise) for each presented face or given the option to choose “none of the above”. The “none of the above” choice was included to avoid inflation in correct labeling inherent in strict forced choice designs. Stimuli were presented in a randomized order that was counterbalanced across participants.

2.3.2. Reliability

Initial reliability ratings were calculated from all participants who were asked to rate the same subset of the 1,721 face stimuli a second time. Stimuli were presented in a new randomized order for each participant.

2.5. Data analytic procedures

2.5.1. Validity

Two validity measures (proportion correct and Cohen's kappa (Cohen,1960)) were calculated for each of the 1721 stimuli modeled after the analyses of the NimStim Set of Facial Expressions (Tottenham et al., 2009). Proportion correct was calculated by comparing the number of participants who correctly endorsed target expressions to the total number of ratings for each stimulus. Though proportion correct is often reported in examinations of facial expression (Ekman and Friesen, 1976, Mandal, 1987, Biehl et al., 1997, Wang and Markham, 1999, Mandal et al., 2001, Beaupre and Hess, 2005), Tottenham et al., (2009) suggests that Cohen's kappa (Cohen, 1960) may be a better dependent variable for evaluations of stimulus sets since proportion correct does not consider false positives (Erwin et al., 1992). Kappa scores, a measure of agreement between participants’ labels and models’ expressions adjusted for agreement due to chance, were used to estimate agreement between selected labels and intended expressions. These scores were calculated across models within each survey, independently for open- and closed-mouth conditions (see Supplemental Text for detailed information on this calculation).

Endorsements of “none of the above” were counted as incorrect. Because of the nuanced differences between the calm and neutral expressions (see Fig. 1 for expression examples and Supplemental Table 1 for instructions), ratings of “calm” and “neutral” were counted as correct for both calm and neutral expressions (Tottenham et al., 2009).

Fig. 1. Download : Download high-res image (754KB) Examples of the 16 expressions in color and black and white. From top left: Angry (closed), Angry (open), Calm (closed), Calm (open), Disgust (open), Disgust (closed), Fear (closed), Fear (open), Happy (closed), Happy (open), Happy (exuberant), Neutral (closed), Neutral (open), Sad (closed), Sad (open), Surprise.

Additionally, a confusion matrix showing mean proportion of MTurk participants endorsing each expression (targets and non-targets) was used to further examine the breakdown of endorsements across expressions. This matrix reiterates proportion correct scores described previously in conjunction with proportion incorrect scores, revealing potential trends in errors (i.e. some emotions were consistently mistaken for the same expressions).

2.5.2. Reliability

Reliability was measured by calculating proportion agreement within subjects between the first and second ratings of each of the 1,721 face stimuli. Ratings from the second survey were not included in computing validity ratings.

3. Results

3.1. Validity

Validity ratings for each of the 16 emotional expressions are presented in Table 1 and Fig. 2 (See Supplemental Tables 2a and 2b for individual proportion correct and kappa scores and Supplemental Table 3 for each actor). The overall proportion correct was high (Mean = 0.71, S.D .= 0.19, Median = 0.75). Kappa scores (the overall measure of agreement between participants’ labels and models’ intended expressions, adjusting for agreement due to chance), were also substantial (Mean = 0.65, S.D .= 0.19, Median = 0.68) indicating stimuli accurately conveyed their intended expressions (Landis and Koch, 1977).Table 1. Validity of ratings for emotional expression categories.

Expression

Median proportion correct

Mean (S.D.) proportion correct

Range proportion correct

Median kappa

Mean (S.D.) kappa

Range kappa

Angry (closed)

0.66

0.62 (0.24)

0.00–0.98

0.61

0.58 (0.21)

−0.05–0.91

Angry (open)

0.74

0.69 (0.24)

0.00–1.00

0.78

0.71 (0.21)

−0.03–0.96

Calm (closed)

0.90

0.86 (0.14)

0.20–1.00

0.76

0.73 (0.14)

0.18–0.95

Calm (open)

0.85

0.78 (0.20)

0.14–1.00

0.81

0.78 (0.17)

0.22–0.98

Disgust (closed)

0.58

0.56 (0.23)

0.10–1.00

0.56

0.53 (0.21)

0.04–0.91

Disgust (open)

0.88

0.81 (0.19)

0.24–1.00

0.66

0.64 (0.17)

−0.06–0.90

Fear (closed)

0.30

0.33 (0.21)

0.00–0.70

0.41

0.40 (0.23)

−0.03–0.81

Fear (open)

0.51

0.48 (0.22)

0.00–0.94

0.50

0.47 (0.21)

−0.02–0.97

Happy (closed)

0.88

0.80 (0.20)

0.08–1.00

0.83

0.80 (0.17)

0.11–0.98

Happy (exuberant)

0.90

0.84 (0.20)

0.00–1.00

0.82

0.78 (0.20)

−0.01–1.00

Happy (open)

0.99

0.98 (0.06)

0.58–1.00

0.92

0.87 (0.13)

0.30–1.00

Neutral (closed)

0.94

0.89 (0.13)

0.42–1.00

0.78

0.75 (0.13)

0.38–0.95

Neutral (open)

0.85

0.78 (0.20)

0.07–0.99

0.83

0.78 (0.17)

0.11–0.96

Sad (closed)

0.79

0.70 (0.26)

0.00–0.99

0.65

0.61 (0.23)

−0.10–0.93

Sad (open)

0.29

0.34 (0.24)

0.00–0.99

0.38

0.41 (0.25)

−0.05–0.87

Surprise (open)

0.88

0.84 (0.14)

0.20–1.00

0.62

0.62 (0.15)

0.17–0.92

Fig. 2. Download : Download high-res image (245KB) Mean validity ratings for each emotional expression. C = Closed Mouth, O = Open Mouth, E = Exuberant, Sur = Surprise. PC = proportion correct.

Kappa scores per actor ranged from 0.6–0.8 in seventy-five percent of the models (i.e., 82 of the total 109 models) reflecting general agreement between participants’ labels and models’ expressions, adjusting for agreement due to chance (Cohen, 1960, Landis and Koch, 1977) and fifty-four percent (59/109) of the mean proportion correct scores were above 0.70. Nearly seventy percent (11/16) of the mean kappa scores calculated for each expression were substantial (Landis and Koch, 1977) including the following expressions: angry-closed, angry-open, calm-closed, calm-open, disgust-open, happy-closed, happy-open, happy-exuberant, neutral-open, neutral-closed, and sad-closed. Mean proportion correct scores were high across expressions, with more than half (9/16) above 0.7 including: calm-closed, calm-open, disgust-open, happy-closed, happy-open, happy-exuberant, neutral-closed, neutral-open, and surprise.

The confusion matrix (Table 2) depicts average proportion of target and non-target labels endorsed for each expression. This matrix demonstrates that expressions were rarely identified as “none of the above” (endorsement across expressions ranged from 0.00–0.05) and that particular poses were consistently mistaken for other expressions. Table 1 and Fig. 2 show low initial validity (proportion correct and kappa) scores for the fear-closed and fear-open expressions, which were consistently mislabeled as surprise (mean fear-closed labeled as surprise = 0.43, S.D .= 0.27, mean fear-open labeled as surprise = 0.41, S.D .= 0.24). Similarly, the open variant of the sad expression, an expression that many models had difficulty posing, was consistently mislabeled as disgust (mean sad labeled as disgust = 0.25, S.D .= 0.22).Table 2. Confusion matrix depicting mean proportion correct (S.D.) for MTurk participants endorsing each of the target expressions.

Photograph

Angry

Calm/Neutral

Disgust

Fear

Happy

Sad

Surprise

None of the Above

Angry

0.62

0.05

0.17

0.01

0.01

0.10

0.01

0.04

(closed)

(0.24)

(0.07)

(0.15)

(0.02)

(0.06)

(0.15)

(0.02)

(0.05)

Angry

0.69

0.01

0.21

0.03

0.02

0.01

0.01

0.02

(open)

(0.24)

(0.02)

(0.18)

(0.05)

(0.09)

(0.04)

(0.03)

(0.03)

Calm

0.01

0.86

0.01

0.00

0.10

0.01

0.01

0.01

(closed)

(0.02)

(0.14)

(0.01)

(0.01)

(0.14)

(0.03)

(0.02)

(0.01)

Calm

0.01

0.78

0.01

0.01

0.12

0.02

0.03

0.02

(open)

(0.02)

(0.20)

(0.03)

(0.03)

(0.18)

(0.03)

(0.08)

(0.03)

Disgust

0.18

0.02

0.56

0.02

0.01

0.18

0.01

0.02

(closed)

(0.17)

(0.03)

(0.23)

(0.03)

(0.05)

(0.22)

(0.04)

(0.04)

Disgust

0.04

0.02

0.81

0.03

0.03

0.01

0.04

0.03

(open)

(0.07)

(0.08)

(0.19)

(0.07)

(0.07)

(0.02)

(0.06)

(0.06)

Fear

0.02

0.05

0.07

0.33

0.01

0.07

0.43

0.02

(closed)

(0.07)

(0.09)

(0.08)

(0.21)

(0.02)

(0.13)

(0.27)

(0.03)

Fear

0.02

0.01

0.05

0.48

0.02

0.01

0.41

0.01

(open)

(0.07)

(0.01)

(0.08)

(0.22)

(0.08)

(0.02)

(0.24)

(0.02)

Happy

0.00

0.18

0.00

0.00

0.80

0.00

0.00

0.01

(closed)

(0.01)

(0.19)

(0.01)

(0.01)

(0.20)

(0.01)

(0.01)

(0.01)

Happy

0.00

0.01

0.00

0.00

0.98

0.00

0.01

0.00

(open)

(0.01)

(0.03)

(0.01)

(0.01)

(0.06)

(0.01)

(0.04)

(0.01)

Happy

0.00

0.02

0.00

0.00

0.84

0.00

0.13

0.01

(exuberant)

(0.01)

(0.10)

(0.01)

(0.01)

(0.20)

(0.01)

(0.16)

(0.01)

Neutral

0.04

0.89

0.01

0.00

0.01

0.04

0.00

0.01

(closed)

(0.07)

(0.13)

(0.03)

(0.02)

(0.02)

(0.05)

(0.01)

(0.02)

Neutral

0.03

0.78

0.04

0.03

0.01

0.04

0.05

0.03

(open)

(0.05)

(0.20)

(0.08)

(0.04)

(0.04)

(0.04)

(0.08)

(0.05)

Sad

0.06

0.11

0.06

0.03

0.00

0.70

0.01

0.02

(closed)

(0.12)

(0.15)

(0.11)

(0.06)

(0.01)

(0.26)

(0.02)

(0.03)

Sad

0.04

0.08

0.25

0.16

0.02

0.34

0.07

0.05

(open)

(0.07)

(0.14)

(0.22)

(0.15)

(0.08)

(0.24)

(0.09)

(0.06)

Surprise

0.00

0.01

0.01

0.06

0.07

0.00

0.84

0.01

(open)

(0.01)

(0.03)

(0.02)

(0.07)

(0.12)

(0.01)

(0.14)

(0.03)

3.2. Reliability

Reliability scores (i.e. proportion agreement) for each emotional expression are reported in Table 3 and presented in Fig. 3, and reliability values for individual stimuli can be found in Supplemental Table 4. Overall, there was consistency between the first and second ratings with a mean reliability of 0.70 (S.D .= 0.16).Table 3. Reliability for emotional expressions between ratings 1 and 2.

Emotion

Mean proportion correct rating 1 (S.D)

Mean proportion correct rating 2 (S.D)

Agreement between ratings 1 and 2 (S.D)

Angry (closed)

0.62 (0.24)

0.62 (0.23)

0.65 (0.15)

Angry (open)

0.69 (0.24)

0.72 (0.23)

0.75 (0.15)

Calm (closed)

0.86 (0.14)

0.87 (0.13)

0.64 (0.10)

Calm (open)

0.78 (0.20)

0.78 (0.19)

0.58 (0.11)

Disgust (closed)

0.56 (0.23)

0.55 (0.23)

0.65(0.12)

Disgust (open)

0.81 (0.19)

0.82 (0.19)

0.79 (0.15)

Fear (closed)

0.33 (0.33)

0.36 (0.22)

0.60 (0.12)

Fear (open)

0.48 (0.22)

0.52 (0.21)

0.65 (0.10)

Happy (closed)

0.80 (0.20)

0.78 (0.19)

0.80 (0.14)

Happy (exuberant)

0.84 (0.20)

0.86 (0.16)

0.84 (0.12)

Happy (open)

0.98 (0.06)

0.97 (0.05)

0.95 (0.05)

Neutral (closed)

0.89 (0.89)

0.90 (0.12)

0.70 (0.09)

Neutral (open)

0.78 (0.20)

0.81 (0.17)

0.61 (0.11)

Sad (closed)

0.70 (0.26)

0.71 (0.26)

0.71 (0.18)

Sad (open)

0.34 (0.24)

0.37 (0.37)

0.52 (0.13)

Surprise (open)

0.84 (0.14)

0.82 (0.14)

0.79 (0.12)

Fig. 3. Download : Download high-res image (242KB) Mean reliability ratings for each emotional expression. C = Closed Mouth, O = Open Mouth, E = Exuberant, Sur = Surprise.

Approximately half (7/16) of the mean reliability scores for each expression ranged from 0.71–0.96 and the remaining (9/16) expressions ranged from 0.52–0.7. Sad-open (Meanagreement= 0.52, S.D. = 0.13) and calm-open (Meanagreement= 0.58, S.D .= 0.11) faces had the greatest variability between rating sessions.

3.3. Race and ethnicity of model

Exploratory examination of variability in accuracy by race of model was performed. Validity and reliability scores are provided for each individual model for each expression in Supplemental Tables 2–4. Fig. 4 provides mean proportion correct for ratings of emotion by race of the model illustrating overall moderate consistency in ratings for each race, however further studies optimized to explore the relationship between emotion perception and race of target are warranted. The variability observed across emotional categories by race of models is primarily for negative emotions. Regardless, mean accuracy for emotional categories for each race was high (70% or above) for 10 of the 16 emotions and above 50% for 3 emotional categories. The remaining expressions of fear-closed mouth, fear-open mouth and sad-open mouth were consistently low across all 4 races (below 50% accuracy). Most of the participants in the current sample were White (470 of 662). Nonetheless, preliminary accuracy plots are provided for each major racial group of participants for visualization purposes in Supplemental Fig. 1a–d. Further research will be needed to assess any effects of participant and model race on the accuracy of emotional ratings.

Fig. 4. Download : Download high-res image (313KB) Accuracy by emotion and race of stimulus across four major race groups of MTurk participants (n = 651). Points and bars represent mean accuracy with standard error.

4. Discussion

This article presents the RADIATE face stimulus set with initial validity and reliability scores calculated from judgments made by research participants on Amazon's MTurk. The large number of racially and ethnically diverse models posing a variety of facial expressions is provided as an open-access resource for researchers interested in psychological processes involving face processing and social stimuli.

Initial validity ratings were obtained with mean proportion correct of 0.71. Because proportion correct does not reflect false alarm judgments, kappa scores were calculated for each stimulus to measure agreement between participants’ labels and models’ intended expressions, adjusting for agreement due to chance. These calculations are based on the entire stimulus set. While some researchers report validity statistics for the stimuli with the highest scores (Dalrymple et al., 2013, Giuliani et al., 2017), other groups (e.g., Ma et al., 2015) do not report validity data and note the need for diverse stimuli to address demographic homogeneity across available databases. The comprehensive RADIATE stimulus set is provided because of the need for representative stimuli of individuals of color, especially considering the growing literature examining in- and out-group interactions (Byatt and Rhodes, 1998, Hart et al., 2000, Golby et al., 2001, Hugenberg and Bodenhausen, 2003, Herrmann et al., 2007, Lieberman et al., 2005). However, for researchers who are primarily interested in high emotion-identification accuracy, we provide Supplemental Tables 6a and 6b with percentage of models per race, sex, and emotion category above 0.70 to aid researchers in determining which category of RADIATE stimuli may be suitable for their purposes. In addition we provide proportion correct and Kappa scores for each of the 1,721 stimuli in Supplemental Tables 2a and 2b.

The validity scores vary across emotion categories, which is consistent with previous face stimulus sets (Ekman and Friesen, 1976, Tottenham et al., 2009) and is likely due to differences in emotion recognition across expressions (Strauss and Moscovitch, 1981, Calder et al., 2003). As previously shown, happy expressions typically have a high identification rate (Hare et al., 2005) while negative expressions have a lower identification rate (Biehl et al., 1997, Lenti et al., 1999, Calder et al., 2003, Calvo and Lundqvist, 2008, Palermo and Coltheart, 2004, Gur et al., 2002, Elfenbein and Ambady, 2003). For example, fear faces are often recognized less accurately than other expressions, particularly in forced-choice design studies (Russell, 1994) and sad faces are often mislabeled as neutral or disgusted (Palermo and Coltheart, 2004). As Tottenham et al. (2009) report, the open- and closed-mouth variants were produced to control for perceptual differences, however this manipulation may have precluded both production and recognition of certain expressions (e.g. some models had difficultly maintaining an expression while adjusting mouth-variant). Some of the less prototypical expressions (e.g. fear-closed, sad-open) are usually not included in face stimulus sets so the lower scores for these categories are not surprising.

It is important to consider the impact that context and culture may have on the perception of emotion in addressing the variance in emotion categories within the RADIATE face stimulus set. Although emotion perception is often considered static, research has demonstrated that contextual and cultural factors from a variety of sources (e.g. culture of stimulus and perceiver, emotion of a situation, or words) can affect how an emotional expression is perceived (Aviezer et al., 2008, Barrett et al., 2011, Kim et al., 2004). Therefore, we provide all stimuli and ratings to facilitate informed decisions regarding which stimuli are appropriate for the experimental design and objectives of future studies.

The exploratory plots illustrating accuracy of emotional categories for each race were moderately consistent. These findings are promising given that the majority of raters were White participants, while the majority of the stimuli are of non-White models. However, future research will be needed to address associations among participant race, model race, and accuracy of categorizing emotional stimuli to support this claim.

Reliability ratings were calculated demonstrating overall consistency across both rating 1 and rating 2 with mean reliability of 0.70. However, there is variance across emotion categories, particularly for the sad-open and calm-open expressions. Given that reliability is a necessary condition for validity, some individual stimuli may not be appropriate for use in studies for which high reliability of emotional category is required.

An advantage of the RADIATE stimulus set is that images were rated using a semi-forced choice design, allowing participants to choose across 9 options (angry, calm, disgust, fear, happy, neutral, sad, surprised, or “none of the above”) for each expression. Consistent with the NimStim (Tottenham et al., 2009) methods, the “none of the above” choice was included because strict forced choice tasks can inflate correct labeling. However, Russell (1994) notes that the subtle complexities of expressions may not fully be captured with this design and that a combination of forced-choice, freely chosen, or spectrum (i.e. slightly happy, moderately happy, very happy) labels may be more appropriate for rating faces. Considering the number of stimuli in the RADIATE set, a major obstacle was obtaining expression ratings of all stimuli from a relatively large sample. To overcome this challenge we used Amazon's MTurk, which is widely used within the research community (Buhrmester et al., 2011, Mason and Suri, 2012a, Crump et al., 2013). However, using this approach required subdivided groups of models’ images, an approach used in other stimulus rating studies (Ma et al., 2015) and the inclusion of additional expression labels would have required further subdivision of models’ images.

The use of MTurk as the sole subject pool is a potential limitation of the study. While all subjects executed this study online, environmental factors such as background noise, room lighting, or the presence of other people could not be controlled and could account for the variance in total survey completion times (Mean = 18.12 min, S.D .= 12.83 min, Range = 5.34–198.17 min). Differences in timing to complete surveys also may be attributed to differences in Internet speed (e.g. image loading) or other individual differences (e.g. rating speed, leaving a browser window open). However, these factors are not known due to the nature of collection methods. Nonetheless, MTurk participants are often recruited to participate in psychology studies, some of which use face stimuli (Tran et al., 2017). The current participants were however “untrained” in that they were not subjected to using the Facial Action Coding System (FACS) (Ekman and Friesen, 1978) to evaluate expressions (Tottenham et al., 2009).

MTurk is often used to obtain robust community samples that can be more diverse than college samples (Paolacci and Chandler, 2014, Mason and Suri, 2012b). While research comparing MTurk participants to in-person samples is limited, one such study (Casler et al., 2013) showed that data obtained via MTurk, social media, and in-person did not vary across these different collection-methods. Additionally, Hauser and Schwarz (2016) found that MTurk participants were more attentive to task instructions than college students studied in-person. Further, in the current study, data were examined for incomplete surveys and suspected duplicate-raters and removed prior to analysis to help minimize the impact of this data collection procedure on the quality of measures.

Another potential limitation of the current study is that most of the MTurk participants self-identified as Caucasian, while the majority of the stimuli were of minority and ethnic groups (75%), which has been demonstrated to interfere with face recognition (Levin, 1996, Byatt and Rhodes, 1998, Levin, 2000, MacLin and Malpass, 2001). Even so, we show substantial validity with Kappa scores per model ranging from 0.6 to 0.8 in seventy-five percent (82 of 109) of the models, reflecting general agreement between participants’ labels and models’ expressions (Cohen, 1960, Landis and Koch, 1977).

In addition to the limitations present in the rater population, model attributes must also be considered. Models in the RADIATE set were from a community-sample opposed to a sample of trained models. Additionally, while the set contains some non-stereotypical depictions of gender (e.g. women with short hair, men with long hair), researchers interested in issues of gender-representation or gender-bias should consider including a greater number of androgynous models.

The main objective for developing the RADIATE Face Stimulus Set was to create a large, racially and ethnically diverse set of facial expressions that research participants could accurately identify. The stimulus set contains 1721 images available in black and white and color and a standard scarf template offering researchers flexibility to combine RADIATE stimuli with other facial expression packages. This stimulus set is available to the scientific community as supplemental material and at http://fablab.yale.edu/page/assays-tools. Diverse stimulus sets of this nature may prove useful in examining psychological processes, especially considering the shift from majority to minority trends in the U.S., by providing representative stimuli that reflect the race and ethnicity of research participants and for testing questions specific to processing of in- versus out-group effects on psychological processes.

Link to Article

Abstract

Faces are often used in psychological and neuroimaging research to assess perceptual and emotional processes. Most available stimulus sets, however, represent minimal diversity in both race and ethnicity, which may confound understanding of these processes in diverse/racially heterogeneous samples. Having a diverse stimulus set of faces and emotional expressions could mitigate these biases and may also be useful in research that specifically examines the effects of race and ethnicity on perceptual, emotional and social processes. The racially diverse affective expression (RADIATE) face stimulus set is designed to provide an open-access set of 1,721 facial expressions of Black, White, Hispanic and Asian adult models. Moreover, the diversity of this stimulus set reflects census data showing a change in demographics in the United States from a white majority to a nonwhite majority by 2020. Psychometric results are provided describing the initial validity and reliability of the stimuli based on judgments of the emotional expressions.

Racially Diverse Affective Expression (RADIATE) Face Stimulus Set: A Resource for Psychological Research

Introduction

The increasing racial and ethnic diversity in the United States necessitates the use of diverse stimuli in psychological research, particularly in studies involving face processing. The Racially Diverse Affective Expression (RADIATE) Face Stimulus Set addresses this need by providing a comprehensive collection of facial expressions from over 100 racially and ethnically diverse models. This article presents the RADIATE set, along with initial validity and reliability scores.

Methods

Participants

Amazon Mechanical Turk (MTurk) participants (N = 662) rated the stimuli.

Stimuli

The RADIATE set consists of 1,721 photographs of 109 adults posing 16 facial expressions (angry, calm, disgust, fear, happy, neutral, sad, surprise) with open- and closed-mouth variants. Models were recruited from a community sample and represented various racial and ethnic groups.

Stimulus Rating Procedure

MTurk participants rated each stimulus by selecting one of eight emotion labels or "none of the above." A second rating was obtained for reliability assessment.

Results

Validity

Validity was assessed using proportion correct and Cohen's kappa. Overall proportion correct was 0.71, and kappa scores indicated substantial agreement between participants' labels and models' intended expressions (mean kappa = 0.65).

Reliability

Reliability, measured as proportion agreement between the first and second ratings, was 0.70 on average.

Race and Ethnicity of Model

Exploratory analysis showed moderate consistency in accuracy ratings across races of models, with some variability observed for negative emotions.

Discussion

The RADIATE set provides a valuable resource for researchers interested in psychological processes involving face processing and social stimuli. The large number of racially and ethnically diverse models allows for the examination of in- and out-group effects.

Limitations

Limitations include the use of MTurk participants as raters, the majority of whom were Caucasian, and the use of a community sample of models.

Conclusion

The RADIATE Face Stimulus Set offers a diverse and standardized set of facial expressions that can facilitate research on psychological processes in a changing demographic landscape.

Link to Article

Abstract

Faces are often used in psychological and neuroimaging research to assess perceptual and emotional processes. Most available stimulus sets, however, represent minimal diversity in both race and ethnicity, which may confound understanding of these processes in diverse/racially heterogeneous samples. Having a diverse stimulus set of faces and emotional expressions could mitigate these biases and may also be useful in research that specifically examines the effects of race and ethnicity on perceptual, emotional and social processes. The racially diverse affective expression (RADIATE) face stimulus set is designed to provide an open-access set of 1,721 facial expressions of Black, White, Hispanic and Asian adult models. Moreover, the diversity of this stimulus set reflects census data showing a change in demographics in the United States from a white majority to a nonwhite majority by 2020. Psychometric results are provided describing the initial validity and reliability of the stimuli based on judgments of the emotional expressions.

Racially Diverse Faces for Research

What's the Problem?

When scientists study how people think and feel, they often use pictures of faces. But most of the faces in these pictures are white. This is a problem because the United States is diverse, and it's important to study people from all backgrounds.

The RADIATE Face Stimulus Set

To solve this problem, researchers created the RADIATE Face Stimulus Set. This set includes over 1,700 pictures of faces from people of different races and ethnicities. The faces show a variety of emotions, such as happiness, sadness, anger, and fear.

How Was the Set Created?

The researchers recruited over 100 people from different backgrounds to pose for the pictures. The people were trained to make different facial expressions. Then, the researchers took pictures of the people and edited the pictures to make them all look similar.

How Accurate Are the Pictures?

To test how accurate the pictures were, the researchers asked people to look at the pictures and identify the emotions being shown. The results showed that people were able to identify the emotions accurately most of the time.

Why Is This Important?

The RADIATE Face Stimulus Set is important because it allows researchers to study how people from different backgrounds process emotions. This information can be used to develop better treatments for mental health problems and to improve communication between people from different cultures.

Link to Article

Abstract

Faces are often used in psychological and neuroimaging research to assess perceptual and emotional processes. Most available stimulus sets, however, represent minimal diversity in both race and ethnicity, which may confound understanding of these processes in diverse/racially heterogeneous samples. Having a diverse stimulus set of faces and emotional expressions could mitigate these biases and may also be useful in research that specifically examines the effects of race and ethnicity on perceptual, emotional and social processes. The racially diverse affective expression (RADIATE) face stimulus set is designed to provide an open-access set of 1,721 facial expressions of Black, White, Hispanic and Asian adult models. Moreover, the diversity of this stimulus set reflects census data showing a change in demographics in the United States from a white majority to a nonwhite majority by 2020. Psychometric results are provided describing the initial validity and reliability of the stimuli based on judgments of the emotional expressions.

Racially Diverse Faces for Research

What's the Big Deal?

Imagine a person takes a test and all the pictures of people look different from themselves. It might be harder to do well, right? That's why it's important to have pictures of people from all different backgrounds in research studies.

Researchers use pictures of faces to study how we think, feel, and interact with each other. But in the past, most of these pictures have been of white people. That's a problem because the United States is diverse, with more people from different races and ethnicities.

The RADIATE Face Stimulus Set

To fix this problem, researchers have created the RADIATE Face Stimulus Set. It's a huge collection of over 1,700 pictures of faces from people of all different races and ethnicities.

The pictures show people making 16 different facial expressions, like happy, sad, angry, and surprised. Researchers can use these pictures to study how people from different backgrounds process emotions and interact with each other.

How They Made It

To make the RADIATE set, researchers took pictures of over 100 people from different racial and ethnic groups. The people in the pictures were asked to make different facial expressions. Then, the researchers had other people rate the pictures to make sure that the expressions were clear and easy to identify.

How Accurate Is It?

The researchers found that people were pretty good at identifying the emotions in the pictures. The average accuracy was 71%, which is considered good. However, some expressions were harder to identify than others, like fear and sadness.

Why It's Important

The RADIATE Face Stimulus Set is a valuable tool for researchers who want to study how race and ethnicity affect our thoughts, feelings, and behaviors. It can help us understand how people from different backgrounds interact with each other and how we can create a more inclusive society.

Link to Article

Abstract

Faces are often used in psychological and neuroimaging research to assess perceptual and emotional processes. Most available stimulus sets, however, represent minimal diversity in both race and ethnicity, which may confound understanding of these processes in diverse/racially heterogeneous samples. Having a diverse stimulus set of faces and emotional expressions could mitigate these biases and may also be useful in research that specifically examines the effects of race and ethnicity on perceptual, emotional and social processes. The racially diverse affective expression (RADIATE) face stimulus set is designed to provide an open-access set of 1,721 facial expressions of Black, White, Hispanic and Asian adult models. Moreover, the diversity of this stimulus set reflects census data showing a change in demographics in the United States from a white majority to a nonwhite majority by 2020. Psychometric results are provided describing the initial validity and reliability of the stimuli based on judgments of the emotional expressions.

Faces That Show Feelings

What are the RADIATE Faces?

The RADIATE faces are a big group of pictures of people from different races and cultures making different facial expressions. There are over 1,700 pictures of over 100 people. Each person is making 16 different faces, like happy, sad, angry, and scared.

Why Are They Important?

These pictures are important because they help scientists study how people understand and react to different emotions. When we see someone's face, we can often tell how they're feeling. But sometimes it's hard to tell, especially if the person is from a different culture or race than us.

The RADIATE faces can help scientists learn more about how people from different backgrounds understand emotions. This is important because it can help us get along better with people who are different from us.

How Were They Made?

People from the community posed for the pictures. They were asked to make different facial expressions, and then their pictures were taken. The pictures were then checked and edited to make sure they were clear and showed the right emotions.

How Accurate Are They?

To find out how accurate the pictures were, people were asked to look at them and say what emotion they thought each face was showing. Most people were able to guess the right emotion most of the time. This means that the RADIATE faces are a good tool for scientists to use to study emotions.

How Can I Use Them?

The RADIATE faces are available for scientists to use for free. They can be used in experiments to study how people understand emotions, how emotions affect our behavior, and more.

Link to Article

Footnotes and Citation

Cite

Conley, M. I., Dellarco, D. V., Rubien-Thomas, E., Cohen, A. O., Cervera, A., Tottenham, N., & Casey, B. J. (2018). The racially diverse affective expression (RADIATE) face stimulus set. Psychiatry research, 270, 1059-1067. https://doi.org/10.1016/j.psychres.2018.04.066

    Highlights