Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (2024)

As a library, NLM provides access to scientific literature. Inclusion in an NLM database does not imply endorsem*nt of, or agreement with, the contents by NLM or the National Institutes of Health.
Learn more: PMC Disclaimer | PMC Copyright Notice

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (1)

Cognitive Research: Principles and Implications

Cogn Res Princ Implic. 2023 Dec; 8: 67.

Published online 2023 Nov 3. doi:10.1186/s41235-023-00521-x

PMCID: PMC10622392

PMID: 37919608

Author information Article notes Copyright and License information PMC Disclaimer

Associated Data

Data Availability Statement

Abstract

People are able to perceive emotions in the eyes of others and can therefore see emotions when individuals wear face masks. Research has been hampered by the lack of a good test to measure basic emotions in the eyes. In two studiesrespectively with 358 and 200 participants, we developed a test to see anger, disgust, fear, happiness, sadness and surprise in images of eyes. Each emotion is measured with 8 stimuli (4 male actors and 4 female actors), matched in terms of difficulty and item discrimination. Participants reliably differed in their performance on the Seeing Emotions in the Eyes test (SEE-48). The test correlated well not only with Reading the Mind in the Eyes Test (RMET) but also with the Situational Test of Emotion Understanding (STEU), indicating that the SEE-48 not only measures low-level perceptual skills but also broader skills of emotion perception and emotional intelligence. The test is freely available for research and clinical purposes.

Keywords: Emotion perception, Face mask, Social cognition, Emotional intelligence

Introduction

The ability to accurately interpret facial expressions is considered an important aspect of social cognition (Beaudoin & Beauchamp, 2020;Frith, 2009; Henry et al., 2016). It affects well-being in relationships, and along with other important social cognitive skills, enables us to understand, interpret and process social information to produce appropriate behavior in relation to the situation (Adolphs, 2003; Henry et al., 2016).

From a neuroscientific perspective, the ability to recognize emotional facial features is supported by integrated and distributed neural systems involving posterior areas such as the posterior superior temporal sulcus, the inferior occipital and fusiform gyri, as well as more rostral areas such as the inferior parietal lobes, the frontal operculum, the orbitofrontal cortex, and subcortical areas such as the amygdala, insula and striatum (Haxby & Gobbini, 2011; Leppänen & Nelson, 2009).

Given the extent of the neural substrates of emotional processes, it is not surprising that there are several clinical conditions that can affect emotion recognition. Autism, for example, can have a strong impact on the development of the ability to recognize facial emotions (Lozier et al, 2014). In other conditions, emotion recognition may be compromised after appropriate development, for example after brain injury (Milders et al., 2003; Yuvaraj et al., 2013), and in neurological (Marcó García et al., 2019) or psychiatric diseases (Dalili et al., 2015; Kohler et al., 2010).

Psychologist are interested in understanding how individuals perceive and respond to facial expressions under normal and pathological conditions. There are various factors that might influence this ability which warrant further investigation, as for instance effects of gender (Abbruzzese et al., 2019; Lawrence et al., 2015), age (Abbruzzese et al., 2019; Guarnera et al., 2018; Sullivan et al., 2007) and personality (Konrath et al., 2014); the role of these factors on facial emotion perception has yet to be fully delineated.

The eyes region represents a fundamental element in emotion recognition (Baron-Cohen et al., 1997). Although full face vision allows for more accurate identification of emotions (Guarnera et al., 2018), eye-tracking studies have shown that the eyes are the facial part most often observed during emotion processing (Guo, 2012; Vassallo et al., 2009), suggesting a crucial role of the eyes in transmitting important cues that allow encoding others’ emotions and internal states.

In 2020, the world faced a reality where, because of Covid-19, it became mandatory in many countries to wear personal protective equipment, especially a face mask, to prevent infection and limit the spread of the virus. For more than a year, human communication was devoid of essential social cues coming from the lower face. People had to rely entirely on emotions conveyed through the eyes. This led to an increase in research on how well people can see emotions in the eyes of others (Blazenkova et al., 2022; Carbon & Serrano, 2021; Fitousi et al., 2021; Grahlow et al., 2022; Grenville & Dwyer, 2022; Grundmann et al., 2021; Kastendiek et al., 2022; Langbehn et al., 2022; Lau & Huckauf, 2021; Leitner et al., 2022; Parada-Fernández et al., 2022; Pazhoohi et al., 2021; Swain et al., 2022; Tsantani et al., 2022; Verroca et al., 2022; Wong & Estudillo, 2022).

For example, Swain et al (2022) asked participants to identify the six basic emotions outlined by Ekman (1992) from faces with and without masks. The authors observed reduced performance with masks, which was especially true for the emotions of disgust, fear and sadness. Anger was slightly harder to identify with face masks than without; no difference was found for happiness and surprise. In addition, participants were asked to complete the Reading the Mind in the Eyes Test (RMET; Baron-Cohen et al., 2001) and the Tromsø Social Intelligence Scale (Silvera et al., 2001). The RMET is a performance test, in which eyes are shown and participants must indicate which emotions are displayed. Although the test was initially developed as a test for theory of mind (the ability to understand other people's mental states), a meta-analysis showed that the test correlates as much with measures of emotion recognition (r = 0.33) as with measures of theory of mind (r = 0.29; Kittel et al., 2022). The Tromsø Social Intelligence Scale is a subjective estimate of social intelligence in which participants indicate on Likert scales how well they think they perform on various social tasks. Swain et al. (2022) reported that emotion perception of faces both with and without masks correlated significantly with the RMET but not with the Tromsø Social Intelligence Scale, in line with many studies showing good correlations between objective tests of emotion recognition but not with tests based on self-report (Murphy & Lilienfeld, 2019).

As with all emotion perception research, having access to good stimulus material is a challenge. This was a weakness in Swain et al (2022), who used only four faces from an existing database: two from male actors and two from female actors, showing different emotions. A low number of stimuli carries a significant risk that the findings are limited to the stimulus set used and do not generalize to other stimuli (Lewis, 2023; Westfall et al., 2014). Looking at the publications listed above, we see that Swain et al. (2022) were not an exception: Ten out of the 16 papers used 10 or fewer face stimuli. Only six studies used more faces (see Table Table1).1). Twelve studies added masks digitally to pictures from existing databases. Two studies (Fitousi et al., 2021; Leitner et al., 2022) developed face stimuli ex novo using actors with real masks. Grenville and Dwyer (2022) used a mixed approach in which they made new face stimuli of six people with and without real masks and added a third condition in which masks were added digitally to the photos without masks.

Table 1

Number of stimuli used in studies investigating the effects of face masks on emotion recognition

SourceStimuli used
Blazhenkova et al. (2022)4 Faces from database
Carbon and Serrano (2021)4 Faces from database
Fitousi et al. (2021)16 New faces
Grahlow et al. (2022)36 Faces from database
Grenville and Dwyer (2022)6 New faces
Grundmann et al. (2021)36 Faces from database
Kastendieck et al. (2022)6 Videos from database
Langbehn et al. (2022)14 Videos from database
Lau and Huckauf (2021)8 Faces from database
Leitner et al. (2022)2 New videos
Parada-Fernández et al. (2022)48 Faces from database
Pazhoohi et al. (2021)16 Faces from database
Swain et al. (2022)4 Faces from database
Tsantani et al. (2022)8 Faces from database
Verroca et al. (2022)10 Faces from database
Wong and Estudillo (2022)6 Faces from database

Other researchers used the RMET (Baron-Cohen et al., 2001) to study the impact of wearing a face mask. Because face masks mainly leave the eyes visible, performance on the RMET can be seen as a good proxy of emotion perception in masked faces. Kulke et al (2022), Ong and Liu (2022), and Trainin and Yeshurun (2021) reported that performance on the RMET was better several weeks after the introduction of mandatory mask wearing than before. Trainin and Yeshurun (2021) further reported a correlation between an individual’s tendency to look at the interlocutor’s eyes and the change in RMET performance after one month of mask wearing. They concluded that ongoing everyday experiences can lead to an enhanced capacity for reading mental and emotional states by looking into the eyes of individuals, especially for people motivated to understand the mental states of others.

The RMET has an advantage over self-selected stimuli because the test is known to have acceptable reliability. The reliability of a test indicates how dependable the test result is. A first way to measure reliability is through test–retest reliability. If a test measures a skill consistently, there will be a high correlation between the scores obtained in the first and second administration. For example, participants who performed poorly in the first session should also perform poorly in the second session.

A second way to measure reliability is by looking at the internal consistency of the test. If a test measures a single skill, all items will have positive intercorrelations. Measures commonly used to assess internal consistency are Cronbach's alpha and McDonald’s omega. For example, Jankowiak-Siuda et al. (2016) reported a Cronbach’s alpha equal to 0.67 for the Polish version of the RMET and a test–retest reliability of 0.89. Koo et al. (2021) found values of, respectively, 0.54 and 0.78 for the Korean version, and Vellante et al. (2013) reported values of 0.60 and 0.83 for the Italian version. Fernández-Abascal et al. (2013) reported a 1-year test–retest reliability of 0.63 for the Spanish version.

Reliability is an important element in research on individual differences because a variable cannot correlate with another variable any more than with itself (Olderbak et al., 2021). Correlational research with variables for which no reliability information is available is, therefore, a risk, because it is not possible to interpret low correlations. Indeed, these may be due to the low reliability of the variables or to the lack of correlation between them. Stimulus sets with low reliability are also suboptimal for experimental research because low reliability often comes from having too many easy and/or difficult items. Performance on such items is unlikely to change much from one condition to another. Ideally, stimulus materials should include the whole range of difficult levels, going from easy to difficult. This is hard to achieve if researchers have to assemble the stimulus set themselves (as happened in the studies of Table Table11).

At the same time, RMET is not an ideal test to study emotion recognition. Although it measures emotion recognition (Kittel et al., 2022; Oakley et al., 2016), it was originally developed to measure theory of mind. It measures complex emotions, such as jealousy, arrogance, and irritation, which differ from the basic emotions (anger, disgust, fear, happiness, sadness, surprise) examined in much emotion recognition research. The wide range of emotions means that the RMET measures not a single factor (skill) but a group of factors (Higgins et al., 2022), as seen in the fact that the test’s internal consistency is lower than its test–retest reliability. There are also concerns that recognizing complex emotions requires a good vocabulary (Kittel et al., 2022; Pavlova & Sokolov, 2022), which limits research with less verbally proficient groups, and the low quality of the old black-and-white pictures becomes a concern.

Because of the aforementioned theoretical context and methodological problems with available tests, the aim of the current study was to develop a new test that measures how well a person can recognize basic emotions in the eyes of others. Such a test should avoid ceiling effects in healthy participants and be useful in clinical populations to detect even subtle deficits in emotion processing.

Study 1

The key requirement for an accuracy test is to have stimuli that differ in difficulty. This is more difficult than is sometimes thought, because it is not enough to have easy and difficult items. If the easy items are achievable for everyone and the difficult items for no one, then there will be no individual differences in scores and placing participants in an easier or more difficult condition will make little difference. What is important is to find the sweet spot of items that are not too easy and not too difficult, so that they are achievable for some participants but not for others, and so that they are achievable under some conditions but not under others. Finding critical items works best if you can start from a rich database, as it is then possible to try out many stimuli in search of good ones. Such databases have recently become available through digitization and database sharing. Many databases for emotional stimuli are reviewed on KAPODI (Diconne et al., 2022). Ultimately, we decided to use the Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) dataset (Livingstone & Russo, 2018).

Method

Participants

A total of 358 healthy adults completed the test. They included 159 Italian participants, consisting of master students of the University of Milano Bicocca (105 females; 53 males; 1 non-binary), together with 199 first-year psychology students from Ghent University (174 females; 23 males; 2 non-binary). The study was conducted following the ethical standards of the Declaration of Helsinki, and the local Ethical Committees. All participants gave their written informed consent, before taking part in the study, and received European University Credits (ECTs) for their participation.

Stimulus materials

The RAVDESS dataset contains clips of 24 professional actors (gender-balanced) expressing Ekman’s six basic emotions, along with a calm (neutral) version. Each expression is performed four times: with weak and strong intensity and while participants are speaking or singing, resulting in 7356 recordings. Other advantages of the dataset are that the stimuli are freely available for research, can be shared with others, and have been evaluated by independent raters. The ratings provide information on how well the emotion can be recognized. Half of the assessments were done without sound, which were particularly interesting for us. The ratings showed that not all emotions were equally easy to recognize (in descending order: disgust, anger, surprise, happiness, fear, calmness, sadness).

For each actor, we selected two expressions for each emotion: one that was recognized by about 90% of the raters and one that was recognized by about 40–60% of the raters. This ensured that the expressed emotions were real and recognizable, at least when the full video clip was watched without sound. At the same time, we hoped that the stimuli would not be too easy and more or less equally difficult for the different emotions. Due to an oversight, we forgot to include the videos of actor 23. This gave a total sample of 315 clips, because seven out of 322 combinations (23 actors × 7 emotions × 2 recognition rates) were not present in the dataset.

We selected what seemed to us to be the best frame from the video and cut a rectangle around the eyes. The size of the rectangle varied from item to item in a further attempt to increase variability in the stimuli (see Fig.1).

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (4)

Examples of stimuli used in Study 1 showing various basic emotions

Procedure

Testing happened online, based on Qualtrics software. Participants were first informed about the task and gave informed consent. Then, they were shown very clear examples of each of the 7 emotions (including full face and part of the body) together with the labels. The Italian labels were in alphabetical order: disgusto [disgust], felicità [happiness], neutrale [neutral], paura [fear], rabbia [anger], sorpresa [surprise], and tristezza [sadness]. The Dutch labels were: angst [fear], neutraal [neutral], verdriet [sadness], verrassing [surprise], vreugde [happiness], walging [disgust], and woede [anger].

After the examples, participants were told that we were investigating whether people could see emotions in eyes. They were told they would see 315 stimuli with one pair of eyes and seven labels in alphabetical order (six emotions and neutral). They had to choose the label they felt best represented the face expression. We indicated that we did not know if the task was doable, so we tried it out. Most participants found the task interesting and life-relevant.

Statistical analysis

We performed a psychometric analysis of the test. First, we wanted to know whether the test would reliably measure differences in performance between participants. Because we did not have two test measurements with some time in-between, we could not look at test–retest reliability, but we could measure the internal consistency of the test. We used the R package psych (Revelle, 2023) to calculate Cronbach’s Alpha, Omega total, and Omega hierarchical. Cronbach’s Alpha is the traditional index but makes a few assumptions that are not present in Omega total. Omega hierarchical is assumed to measure how much of the total variance is due to a single, general construct affecting all items, despite the multidimensional nature of the item set (see Flora, 2020, for a tutorial on reliability coefficients).

Second, we wanted to know whether the test measured a single factor (emotion recognition) or multiple factors. A good test is one that measures the intended trait as purely as possible. If seeing emotions in the eyes is a unitary skill, we should see that the test mainly measures a single factor. This was assessed with a scree plot analysis using the R package psych (Revelle, 2023). A scree plot shows the weights (eigenvalues) given to factors extracted in a factor analysis (FA) and/or a principal component analysis (PC). A test that measures a single factor is characterized by a high weight of the first factor, ideally followed by weights below 1 for the second and subsequent factors. Because the latter is not always the case, modern techniques compare the weights obtained with weights expected from a random data set with the same structure (see Sakaluk & Short, 2017, for a tutorial of exploratory factor analysis). All the figures of the scree plot analysis were taken from the R package psych (Revelle, 2023) and based on tetrachoric correlations (to take into account the binary dependent variable). A full-page version of the figures is available in the osf archive; they can also be generated from the data and the R code in the repository.

Finally, we wanted to reduce the item pool and select only items that were interesting for our test. These were items on which (1) participants with low skill scored lower than participants with high skill, (2) which came from different actors, (3) which were balanced for gender and emotion, and (4) which had varying difficult levels.

Results

Data from Milan and Ghentlabs were analyzed together, as we wanted to have a large and diverse sample to start from. No data had to be deleted because of careless responding (either random responding or repetition of the same response). Participants selected the correct response most often for happy and calm expressions, and least often for surprise and disgust (see Table Table2),2), even though we started from equally difficult video clips (see stimulus selection). It is important to realize that the percentage correct per emotion is a combination of sensitivity and response bias. Participants who decided to press calm whenever they were in doubt, would be correct more often with calm stimuli than participants who chose to press happy whenever they were in doubt. Table Table22 also shows a good standard deviation in the percentage correct per stimulus. This opens perspective for stimulus selection. Overall, participants were correct on 50% of the trials, whereas 14% was the guessing rate (100/7). All emotions were recognized above chance level.

Table 2

Percentage of emotions correctly chosen (guessing level = 0.14)

EmotionMeanSD
Anger0.400.32
Calm0.700.20
Disgust0.300.22
Fear0.390.25
Happiness0.730.26
Sadness0.650.22
Surprise0.340.29

Mean represents the accuracy data calculated on the stimulus average. SD Standard Deviation

Reliability of the test (i.e., internal consistency) was very good: Cronbach’s alpha was 0.89, Omega total 0.90. And Omega hierarchical was 0.84. The high value for the Omega hierarchical, can be interpreted as suggesting that the test primarily measured a single construct (which may be subdivided into more than one part, hence the name hierarchical).

A less favorable result was found in the scree plot. As shown in Fig.2, there were several factors with weights above 5, and more importantly, with weights considerably above those that would be expected for noise data (simulated or resampled). This indicated that the test measured multiple factors and this was true for both the outcome of a factor analysis (FA) and principal component analysis (PC). The scree plot analysis indicated that the high values of alpha and omega were not due to high intercorrelations between the items but to the large number of items in the test. The mean test result is then stable (reliable), but the test itself is influenced by many factors, even if we assume that the total test score measures a single factor (skill in emotion recognition).

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (5)

Result of a scree plot analysis of all items from Study 1. The plot shows that there are many factors with high weights (eigenvalues), which is not ideal for a test supposed to measure a specific skill

Based on the outcome of the analyses, we decided to select the best stimuli per basic emotion, except for the calm (neutral) expression. The latter was dropped because it became clear that participants differed considerably in the selection of this category, introducing noise in the dataset.

The following criteria were used for item selection. First, we only selected items with an item-rest correlation higher than 0.2. Item-rest correlation refers to the correlation between performance on an item and the average performance on the other items. A good item is an item with low scores for participants who perform poorly on the other items, and high scores for participants who perform well on the other items. High item-rest correlations increase the likelihood that a single factor accounts for most of the variance. The modest item-rest correlations obtained in Study 1 indicate that there is quite a bit of measurement noise and thus that the test needs a fairly large number of items to obtain good reliability. We additionally used various IRT analyses to select items that differed in difficulty and had good discrimination power. These are not reported here but can be replicated with the R program in the osf archive.

Second, we aimed to decrease the performance differences between emotions and to increase the overall level of accuracy, as low performance is demotivating for participants.

Finally, we wanted to include as many different actors as possible and have gender-balanced lists for the various emotions.

Table Table33 shows the outcome of the selection of the 60 best items. Unfortunately, it was not possible to select an equal number in each category. In particular, the number of surprise expressions was low (N = 6).

Table 3

Statistics of the 60 items chosen for validation

EmotionMeanSDN
Anger0.680.2112
Disgust0.590.179
Fear0.630.1411
Happiness0.690.2810
Sadness0.730.1512
Surprise0.590.266

SD standard deviation, N number of stimuli

Reliability of the selected items was Cronbach’s alpha = 0.85, Omega total = 0.86, and Omega hierarchical = 0.46 (we will return to the low value of Omega hierarchical later, after Study 2). At the same time, the scree plot analysis gave stronger evidence for the importance of the first factor, which now accounted for 19% of the variance (see Fig.3). It was still the case that the test contained several factors with eigenvalues much higher than those expected for randomly simulated or resampled data, suggesting that test performance was affected by several factors other than the one we are interested in. As reported in the introduction, the same observation has been made for the RMET, where internal consistency is lower than test–retest reliability.

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (6)

Outcome of a scree plot analysis of the 60 selected items from Study 1. This shows that the first factor had a larger impact than in the unselected sample of items

Discussion

Study 1 was a good illustration of the fact that items from a standardized database are not sufficient for a good test. Although the sample of items we took had good reliability because of its size, it proved impossible to find 12 good items for each emotion. In particular, participants had difficulties distinguishing between anger, disgust, fear, and surprise on the basis of the eyes alone (see also Swain et al., 2022). They were better at seeing happiness and sadness. They also often selected the calm (neutral) option, but we have reasons to believe this was mainly due to response bias, because the neutral option was chosen quite often for the other emotions as well. To some extent, this was an inevitable consequence of using difficult stimuli that caused accuracy differences between participants: it made participants hesitate and increased the impact of response biases. On the positive side, the data suggested that it was possible to design a reliable test.

An important caveat to the findings so far was that we could not be sure that the item selection resulted in good stimuli until we conducted a validation study on a new group of participants. At worst, it could be that the performance on the selected items was entirely due to overfitting (i.e., selecting items that happened to look good in the sample because of chance fluctuations in human performance). In that case, we would observe that the selected items in a new study would behave very similar to the unselected items of Study 1 (see Fig.1) because of regression to the mean. We tried to avoid such a scenario by testing 358 participants from two different universities, but we could not be sure until we replicated the findings in a new sample.

Another possible danger was that our selection of items would largely measure a single factor, but that this factor would no longer correlate with emotion recognition (for example, because it reflected the perception of a visual feature, present in some faces but absent in others). The only way to find out if the selected items measured the intended skill was to see if performance on them correlated with other, established emotion recognition tests.

Study 2

A test is only useful if the items give the same data in a new, independent sample, so that one can trust the findings. This was the first reason for running the second study.

The second reason was that we needed information on validity. If our SEE test (Seeing Emotions in the Eyes) measured emotion recognition, it should correlate well with other tests for emotion recognition (e.g., Schlegel et al., 2019). The obvious test to compare the SEE test to was the RMET (Baron-Cohen et al., 2001), given that both tests have the same format, differing mainly in the number of emotions tested (36 complex emotions vs. 6 basic emotions) and in the number of response alternatives (4 instead of 6). Another accuracy test often used to measure individual differences in emotion recognition is the Situational Test of Emotional Understanding (STEU; MacCann & Roberts, 2008). In this test, 42 social situations are verbally described and participants are asked to choose from five options the feeling that the protagonists are likely to experience.

In addition to finding convergent evidence with tests that were supposed to measure the same construct, it was informative to check to what extent the test measured related constructs. For example, it is interesting to know whether the SEE test is influenced by anxiety in participants (Demenescu et al., 2010). A widely used test to measure anxiety is the State-Trait Anxiety Inventory (STAI; Spielberger et al., 1983). This test measures anxiety both as a stable trait and in the specific test situation (anxiety as a state). A second related construct we wanted to study was apathy, a reduced motivation for purposeful activity in terms of behavior, cognition, emotion, and social functioning (Robert et al., 2018). Given the evidence that emotion perception can improve after practice (Kulke et al., 2022; Ong & Liu, 2022; Trainin & Yeshurun, 2021), it could be that performance on the SEE test is lower in people with low motivation to train emotion recognition. Apathy may also interfere with emotion recognition because it affects emotional/social dimensions, which can lead to less interest in the other person's feelings and social withdrawal. A test designed to measure apathy is the dimensional apathy scale (DAS; Radakovic & Abrahams, 2014).

All in all, we added five tests (RMET, STEU, STAI-trait, STAI-state, DAS) to the SEE, in order to better assess the value of the test.

A last reason for conducting a second study was that we were not completely satisfied with the item selection we could make based on Study 1. We had hoped to select 12 items per emotion, so that each emotion could be investigated on its own, but this proved impossible (see Table Table3).3). Given the information we gathered in Study 1, it seemed a missed opportunity not to use the information we gained to select additional stimuli that would hopefully be clearer for surprise, fear, and disgust, and more difficult for anger, happiness, and sadness. In addition, it made sense to try to spread the number of different actors in the stimulus set as much as possible, to improve generalization to other faces. So, we added new stimuli from the faces that were not selected.

Method

Participants

The participants were 200 new master students from Milano Bicocca (115 women; 85 men), who completed all tests. The study obtained ethical approval; all participants gave their written informed consent and received European University Credits (ECTs) for their participation.

Stimulus materials

The first set of stimuli consisted of the 60 SEE faces we selected from Study 1 (Table (Table3).3). We call this test SEE1. In addition, 72 new SEE stimuli were created to try to fill gaps. The items were distributed as follows: 10 new ones for anger, 15 for disgust, 10 for fear, 13 for happiness, 11 for sadness, and 13 for surprise. These stimuli form test SEE2. Unlike in Study 1, there were no longer neutral (calm) items, so participants had to choose between six emotionally charged options.

For RMET, we used the Italian version developed by Serafin and Surian (2004) and also used by Vellante et al. (2013) and Maddaluno et al. (2022). It contained, as the English version, one practice item and 36test items, but the labels were translated into Italian. Each item had four response alternatives, one of which was correct.

We translated the STEU ourselves and tried to stay as close as possible to the English version. All 42 items were translated, even though it is known that not all items work well in English (Allen et al., 2014). The reasoning was that it was better to start with the largest possible pool, which could then be pruned, than to start with a reduced item set in the hope that the Italian pattern of good stimuli would follow the English one. Each item had five response alternatives, only one of which was correct.

The Italian version of the STAI was developed and tested by Pedrabissi and Santinello (1989). It contains 40 items, 20 related to anxiety in general (trait), and 20 to anxiety at the moment (state). Participants filled in a Likert scale from 1 to 4 to indicate how strongly the item applied to them.

Finally, we used the Italian version of the DAS (Santangelo et al., 2017). It contained 24 self-ratings on a 4-point Likert scale.

Procedure

All tests were programmed in Qualtrics and administered online. The order of the tests was the same for all participants. After explaining the purpose of the study and asking for informed consent, we first presented the 60 items of SEE1 and immediately afterward (without pause), participants saw the 72 items of SEE2. Subsequently, they were given the RMET, STEU, STAI and DAS, with the usual instructions.

Statistical analysis

In the second study, we focused on the validity of the test developed in Study 1 (SEE1) to examine whether it measured emotion recognition as a skill. Therefore, we investigated the convergent validity between the four tests hypothesized to measure emotion recognition (SEE1, SEE2, RMET, STEU) and to what extent these correlations differed from tests measuring other aspects of human functioning (STAI and DAS).

Next, we looked at the quality of each test, performing the same statistical procedures as in Study 1 (reliability, item analysis, factor analysis, IRT). As in Study 1, all figures from Study 2 can be reproduced using the R code and are also available as full-page figures on the osf repository.

Finally, we selected items from SEE1 and SEE2 to come to the best possible test.

Results

As in Study 1, there was no evidence to distrust the data from participants (probably because they found the test interesting and wanted to help us make a good test).

Before going into the details of the test analysis, we had a look at the correlations between the tests. These correlations contained important validity information, because there was little to expect from item analysis if there were no significant correlations between our tests and the other tests. A comparison of the Pearson and Spearman correlations revealed that our sample included a few low and high scorers, who did not really distort the correlation pattern but still had an upward effect on the Pearson correlation. Because the elevated Pearson correlations were unlikely to be replicated in a new study, we used robust correlation calculation (pbcor; Mair & Wilcox, 2020).

We found that all four emotion recognition tasks (SEE1, SEE2, RMET, STEU) correlated strongly with each other (see Table Table4).4). This confirms that they all measured to a large extent the same skill. The good performance of the new SEE2 items was particularly impressive, as this test was almost as good as the SEE1, even though performance on the new items was significantly lower (48% correct vs. 69% correct). Furthermore, it was interesting to see that the STAI and the DAS had positive intercorrelations, indicating that participants with high anxiety had a higher lack of motivation for goal-directed behavior. The correlations between the emotion recognition tests and DAS were negative, as expected, but not high enough to be significant. Test performance was not affected by the anxiety or apathy level of the participants (who it should be remembered were all university students).

Table 4

Performance on the tests (SEE1, SEE2, RMET, STEU, STAI-T, STAI-S, DAS) and correlations between the tests

MSDSEE2RMETSTEUSTAI-TSTAI-SDAS
SEE10.690.140.580.430.49− 0.030.01− 0.12
SEE20.480.100.390.44− 0.03-0.04− 0.13
RMET0.710.130.47− 0.06− 0.09− 0.09
STEU0.580.130.080.04− 0.01
STAI-T2.310.560.780.46
STAI-S2.160.570.31
DAS2.060.35

For the SEE1, SEE2, RMET, and STEU, the values of M and SD refer to percentage correct; for the STAI-T, STAI-S, and DAS the values refer to a Likert scale from 1 to 4. Correlations in bold are significant at p < 0.01. The other correlations are not significant at p  < 0.05

M mean, SD standard deviation, RMET reading the mind in the eyes test, STEU situation test of emotional understanding, STAI-T stait trait anxiety inventory-Trait, STAI-S stait trait anxiety inventory-state, DAS dimension apathy scale

As can be expected based on the intercorrelations, reliability was good for all tests, slightly better for the SEE tests than for RMET and STEU, in line with the wide selection of stimuli we could start from (see Table Table5).5). A less interesting finding was that Omega hierarchical seems to be a rather unstable reliability estimate. While it was low for the SEE1 items (0.43) in Study 1, it was high (0.83) in Study 2, and at the same time it was low for SEE2 (0.21) even though that test measured the same construct in a very similar way, given the high correlation between SEE1 and SEE2. This indicates that Omega hierarchical may not be as good an indicator of test quality as has been hoped (see also Cho, 2022). The reliability of STEU was lower than 0.8 but can be improved by omitting items with a poor profile, in line with what Allen et al. (2014) did with the English test (see the osf repository for detailed item analysis and selection).

Table 5

Reliability of the tests

AlphaOmegaTOmegaH
SEE10.890.900.83
SEE20.840.860.21
RMET0.770.790.72
STEU0.770.790.15
STAI-T0.920.920.70
STAI-S0.930.950.63
DAS0.770.850.81

OmegaT omega total, OmegaH omega hierarchical

Finally, Fig.4 shows the scree plots of the seven tests. It illustrates that SEE1 did not lose much of its quality compared to Study 1 (see Fig.3). The first factor explained much more variance than the subsequent factors. SEE2 had more features of the unselected items from Study 1 (see Fig.1), with more evidence for the contribution of more than one factor. The scree plot further showed that the STAI was better characterized as a combination of two factors and the DAS tests as a combination of three factors, in line with previous observations (STAI: Shek, 1991; DAS: Lafond-Brina & Bonnefond, 2022; Santangelo et al., 2017). This opens up possibilities for more detailed analyses in the future, when more data are available.

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (7)

Scree plots of all tests used in Study 2. Left part: Scree plots of the four emotion recognition tests show that these tests largely measure a single factor. Right part: Scree plots show that STAI-T and STAI-S probably measure two different factors with weights considerably above those based on simulated or randomly resampled data, and that DAS may even measure three factors

Discussion

The pattern of results in Study 2 was clearer than anticipated. The study confirmed that the four emotion recognition tests (SEE1, SEE2, RMET, STEU) largely measured the same underlying skill, even though they differed methodologically from each other and were not perfect in their details. Indeed, the RMET has been criticized several times for not being a pure measurement of a single factor (Higgins et al., 2022; Kittel et al., 2022; Olderbak et al., 2015). At the same time, it showed high test–retest reliability (Fernández-Abascal et al., 2013; Jankowiak-Siuda et al., 2016; Koo et al., 2021; Vellante et al., 2013) and good correlations with convergent tests (Kim et al., 2022; Schlegel et al., 2019; the present study). Similarly, the STEU has been criticized for containing many suboptimal items (Allen et al., 2014). Nevertheless, looking at the correlations with related tests, the STEU seems to perform well at an average level. The high intercorrelations between the tests suggest that statistical fit criteria of individual tests may be less important than sometimes thought in the context of confirmatory factor analysis (Hopwood & Donnellan, 2010) and may even lead to construct underrepresentation (Messick, 1989) if the underlying skill is hierarchically organized and the test does not systematically map the subcomponents (Neubauer & Hofer, 2022; Steger et al., 2019). Also, Omega hierarchical seems to differ too much between very similar tests or even two runs of the same test to be used as a strong cue for test quality (Cho, 2022).

A further interesting finding of Study 2 was that SEE2 seemed to measure the construct of interest (the ability to recognize emotions) nearly as well as SEE1 (high correlation between both tests and high correlations with the other emotion-recognition tests), even though several items were far from being optimal. The latter can be seen an in item response theory (IRT) analysis. IRT analysis provides the same information as item-rest correlation, but in more detail, because it shows how performance differs as a function of participant skill. An items is good if low-proficiency participants have a low chance of knowing the answer and high-proficiency items have a high chance. In addition, the transition from low to high is rather steep (item discrimination) and differs between items (item difficulty). The IRT analysis was conducted using the R package mirt (Chalmers, 2012). Figure5 shows the outcome of the analysis for SEE1; Fig.6 shows the outcome for SEE2.

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (8)

Item Response Theory analysis of SEE1 in Study 2: Most items show the expected pattern of performance (low at the left of the curve and high at the right) except for a few (e.g., A21-D2, A13-SU1, A9-A2 and A18-Sa2)

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (9)

Item Response Theory analysis of SEE2, showing several bad items with higher scores for low-proficiency participants than for high-proficiency participants (e.g., A20-Su3, A9-Sa3, A4-D3) among good ones (A6-Su3, A20-Sa3, A2-H3, A20-F3, A19-D3, A22-A3)

Combining the items from SEE1 and SEE2, we were able to compose a good test with 48 items (SEE-48). The test contains 8 matched items for each of the emotions, coming from 22 different actors (four male and four female per emotion). Performance on the items varied from 40 to 90% with an average of 63–70% per emotion, as shown in Table Table66.

Table 6

Accuracy rates for the SEE-48 stimuli in Study 2

StimulusAngerDisgustFearHappySadSurprise
S10.910.880.780.910.830.88
S20.870.740.770.800.830.87
S30.690.700.750.800.800.82
S40.640.700.720.770.780.72
S50.610.640.690.690.730.62
S60.560.600.560.680.590.54
S70.550.410.560.540.550.54
S80.410.350.490.420.440.49
Mean0.660.630.670.700.690.69
SD0.170.170.110.160.150.16

S1 is the easiest item for each emotion; S8 is the most difficult item for each emotion. See the osf site to find the pictures corresponding to these stimuli

Figure7 gives the outcome of an IRT analysis of the selected items. These provide a test with reliability of Cronbach’s alpha = 0.87 and Omega total = 0.90 (reliability is likely overestimated due to overfitting but should still be above 0.8 in a new sample, given the similarity of the data in Study 1 and Study 2). Robust correlation was 0.49 between SEE-48 and RMET, and 0.46 between SEE-48 and STEU. The negative correlations with STAI-T, STAI-S and DAS were not significant.

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (10)

Item Response Theory analysis of the SEE-48 items

General discussion

Humans are able to see emotions in the eyes of others (Allen-Walker et al., 2015; Baron-Cohen et al., 1997; Biotti & Cook, 2016). This ability underlies social interactions and well-being and has recently received attention due to the mandatory wearing of a face mask during Covid-19. Despite a considerable number of publications, we could not find a good test when we wanted to include the ability in our research. In particular, we could not find a test for basic eyes emotion recognition that contained information on reliability and validity. Therefore, we decided to create such a test, which we call SEE-48 (Seeing Emotions in Eyes, consisting of 48 items). Making the test was easier than expected. The main problem was to match performance across emotions, because not all basic emotions were equally easy to perceive, even though we started from equivalent video stimuli (see also Swain et al., 2022). Another challenge was finding good images of different actors that were gender-balanced (four different male actors and four different female actors per emotion).

In the end, we had to settle for eight pictures per emotion, bringing the total to 48 instead of the 72 we were hoping for,1 because there were not enough easy items for some emotions and not enough difficult items for others. The number of items in SEE-48 is sufficient for a reliable measure of emotion recognition, but perhaps not for researchers who want to study individual emotions. For anger, happiness and surprise, additional stimuli from SEE1 and SEE2 can be chosen if needed. For the other emotions, this will be difficult. Another option could be to abandon the gender balance criterion.

Test construction was helped by the observation that emotion recognition seems to be a robust ability, affecting performance in SEE as much as in RMET and STEU (see Table Table4).4). The latter is interesting because there is a big difference between seeing basic emotions in the eyes and knowing what emotion is felt in verbally depicted situations. So, the SEE test not only measures a low-level, perceptual skill (recognizing emotions from eyes) but also a broader ability related to emotional intelligence (Simonet et al., 2021). The correlations between SEE, RMET and STEU are in line with other studies showing good correlations between performance-based emotion recognition tests (Bryan & Mayer, 2021; Schlegel et al., 2019). They confirm that all these tests measure a common skill, while at the same time, having unique components (otherwise correlations between tests would be as high as test reliability). Campbell and Fiske (1959) already argued that each test or task used for measurement purposes is a combination of traits and measurement procedures, and that true understanding of a specific trait arises only when researchers are able to examine convergent performance from different tests expected to measure the same trait in different ways.

There are several reasons why we see the SEE-48 test as a good addition to the existing tests. First, the test scores well on reliability (Olderbak et al., 2021). Second, the test makes minimal use of language (only the names of the six basic emotions need to be known) and this makes the test useful for people with a small vocabulary or who have little experience with the test language (e.g., second-language speakers). Third, results show that test scores are not significantly correlated with anxious and apathetic traits, making the test robust to the influence of such factors (at least in university students). Fourth, the SEE-48 test provides high-graphic-quality matched stimuli for six basic emotions and actor’s gender, which opens up possibilities for testing differences in emotion perception across groups. Fifth, it allows to study the ability to recognize emotion in healthy participants avoiding ceiling effects. Finally, if a study willinclude enough participants, it will even become possible to analyze complete confusion matrices between the six emotions (see Barhoom et al., 2021, for methods to use).

Availability

A test for research and assessment in clinical settings is only useful if it is easily and freely available. This allowed us to use the RAVDESS database (Livingstone & Russo, 2018) to create our own test.

To encourage further cumulative research, we are making the SEE-48 test available at https://osf.io/ed7f6/. The repository contains the selected eye stimuli, the full faces from which they were cut (for those who want to change aspects of the stimuli or compare the eye stimuli with masked faces), performance indices, and a file to present the tests in Qualtrics. Interested readers can also try out the test at https://ugent.qualtrics.com/jfe/form/SV_emuHopiA6Cn7IUe. In addition, we provide all the information from Studies 1 and 2 so that colleagues can select more stimuli if needed. The number of stimuli in the SEE-48 test could be further reduced if individual emotions are not important, if a reliability lower than 0.8 is sufficient, or if the test turns out to be too difficult for a particular population (in which case the most difficult items can be dropped). However, for the time being we believe that the SEE-48 test is short enough (usually about five minutes) to be administered in total. Some redundancy is good to make a test robust.

Acknowledgements

Theresearch was partially funded by the Italian Ministry of Health (to N.B.).

Significance statement

Emotion perception is important for social life. Most people are able to infer six basic emotions from others’ faces: anger, disgust, fear, happiness, sadness and surprise. They do not need to see the full face, as emotions can be perceived reasonably well in just the eyes, allowing us to see the emotions in people with a face mask. Emotion perception differs between healthy individuals and some clinical conditions may impair this ability, impacting social cognition at different levels. To properly investigate such differences, it is important to assess them in a reliable and valid way. We present a test that we believe significantly improves the available stimulus material and will enable researchers to better understand the importance of emotion understanding in human functioning.

Abbreviations

DASDimensional apathy scale
RMETReading the mind in the eyes test
SEESeeing emotions in the eyes
SEE-48Seeing emotions in the eyes test with 48 questions
STAIState-Trait Anxiety Inventory
STEUThe situational test of emotion understanding

Author contributions

MF designing the studies, running the studies, analysis of the data, and writing the ms; NB designing the studies and writing the ms; MB designing the studies, stimulus selection, analysis of the data, and writing the ms.

Funding

No funding from outside sources was involved.

Availability of data and materials

All data, figures, and stimulus materials are available at https://osf.io/ed7f6/.

Declarations

Ethics approval and consent to participate

The study was approved by the two universities involved. For the University of Milan, ethical approval was given by the ethics committee; for Ghent University, no explicit approval was needed, as the study followed the General Ethical Protocol of the Faculty of Psychology and Educational Sciences, which covers all research that does not involve specific ethical considerations. All participants gave informed consent for the use of their data in anonymized form.

Consent for publication

All authors and participants provided consent for publication.

Competing interests

The authors of this article have no conflicts of interest to declare. All authors have seen and agree with the contents of the manuscript, and there is no financial interest to report. We certify that the submission is original work and is not under review at any other publication.

Footnotes

1Having 72 items would have been better, because then we had 12 items per emotion, giving more power to tests of a specific emotion, or making it possible to create two parallel versions of the test.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Nadia Bolognini, Email: ti.ocigoloxua@iningolob.n.

Marc Brysbaert, Email: eb.tnegu@treabsyrb.cram.

References

  • Abbruzzese L, Magnani N, Robertson IH, Mancuso M. Age and gender differences in emotion recognition. Frontiers in Psychology. 2019;10:2371. doi:10.3389/fpsyg.2019.02371. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Adolphs R. Cognitive neuroscience of human social behaviour. Nature Reviews Neuroscience. 2003;4(3):165–178. doi:10.1038/nrn1056. [PubMed] [CrossRef] [Google Scholar]
  • Allen VD, Weissman A, Hellwig S, MacCann C, Roberts RD. Development of the situational test of emotional understanding–brief (STEU-B) using item response theory. Personality and Individual Differences. 2014;65:3–7. doi:10.1016/j.paid.2014.01.051. [CrossRef] [Google Scholar]
  • Allen-Walker L, Beaton AA. Empathy and perception of emotion in eyes from the FEEST/Ekman and Friesen faces. Personality and Individual Differences. 2015;72:150–154. doi:10.1016/j.paid.2014.08.037. [CrossRef] [Google Scholar]
  • Barhoom H, Joshi MR, Schmidtmann G. The effect of response biases on resolution thresholds of Sloan letters in central and paracentral vision. Vision Research. 2021;187:110–119. doi:10.1016/j.visres.2021.06.002. [PubMed] [CrossRef] [Google Scholar]
  • Baron-Cohen S, Wheelwright S, Hill J, Raste Y, Plumb I. The “reading the mind in the eyes” test revised version: A study with normal adults, and adults with asperger syndrome or high-functioning autism. Journal of Child Psychology and Psychiatry, and Allied Disciplines. 2001;42(2):241–251. doi:10.1111/1469-7610.00715. [PubMed] [CrossRef] [Google Scholar]
  • Baron-Cohen S, Wheelwright S, Jolliffe AT. Is There a «language of the eyes»? Evidence from normal adults, and adults with autism or Asperger syndrome. Visual Cognition. 1997;4(3):311–331. doi:10.1080/713756761. [CrossRef] [Google Scholar]
  • Beaudoin, C., & Beauchamp, M. H. (2020). Social cognition. In Handbook of clinical neurology (Vol. 173, pp. 255–264). Elsevier. 10.1016/B978-0-444-64150-2.00022-8 [PubMed]
  • Biotti F, Cook R. Impaired perception of facial emotion in developmental prosopagnosia. Cortex. 2016;81:126–136. doi:10.1016/j.cortex.2016.04.008. [PubMed] [CrossRef] [Google Scholar]
  • Blazhenkova O, Dogerlioglu-Demir K, Booth RW. Masked emotions: Do face mask patterns and colors affect the recognition of emotions? Cognitive Research: Principles and Implications. 2022;7(1):33. [PMC free article] [PubMed] [Google Scholar]
  • Bryan VM, Mayer JD. Are people-centered intelligences psychometrically distinct from thing-centered intelligences? A Meta-Analysis. Journal of Intelligence. 2021;9(4):48. doi:10.3390/jintelligence9040048. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Campbell DT, Fiske DW. Convergent and discriminant validation by the multitrait-multimethod matrix. Psychological Bulletin. 1959;56(2):81–105. doi:10.1037/h0046016. [PubMed] [CrossRef] [Google Scholar]
  • Carbon CC, Serrano M. The impact of face masks on the emotional reading abilities of children: A lesson from a joint school–university project. i-Perception. 2021;12(4):20416695211038265. doi:10.1177/20416695211038265. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Chalmers RP. mirt: A multidimensional item response theory package for the R environment. Journal of Statistical Software. 2012;48:1–29. doi:10.18637/jss.v048.i06. [CrossRef] [Google Scholar]
  • Cho E. Reliability and omega hierarchical in multidimensional data: A comparison of various estimators. Advance Online Publication; 2022. [PubMed] [Google Scholar]
  • Dalili MN, Penton-Voak IS, Harmer CJ, Munafò MR. Meta-analysis of emotion recognition deficits in major depressive disorder. Psychological Medicine. 2015;45(6):1135–1144. doi:10.1017/S0033291714002591. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Demenescu LR, Kortekaas R, den Boer JA, Aleman A. Impaired attribution of emotion to facial expressions in anxiety and major depression. PLoS ONE. 2010;5(12):e15058. doi:10.1371/journal.pone.0015058. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Diconne K, Kountouriotis GK, Paltoglou AE, Parker A, Hostler TJ. Presenting KAPODI–the searchable database of emotional stimuli sets. Emotion Review. 2022;14(1):84–95. doi:10.1177/17540739211072803. [CrossRef] [Google Scholar]
  • Ekman P. An argument for basic emotions. Cognition & Emotion. 1992;6(3–4):169–200. doi:10.1080/02699939208411068. [CrossRef] [Google Scholar]
  • Fernández-Abascal EG, Cabello R, Fernández-Berrocal P, Baron-Cohen S. Test-retest reliability of the ‘reading the mind in the eyes’ test: A 1-year follow-up study. Molecular Autism. 2013;4:1–6. doi:10.1186/2040-2392-4-33. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Fitousi D, Rotschild N, Pnini C, Azizi O. Understanding the impact of face masks on the processing of facial identity, emotion, age, and gender. Frontiers in Psychology. 2021;12:4668. doi:10.3389/fpsyg.2021.743793. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Flora DB. Your coefficient alpha is probably wrong, but which coefficient omega is right? A tutorial on using R to obtain better reliability estimates. Advances in Methods and Practices in Psychological Science. 2020;3(4):484–501. doi:10.1177/2515245920951747. [CrossRef] [Google Scholar]
  • Frith C. Role of facial expressions in social interactions. Philosophical Transactions of the Royal Society B: Biological Sciences. 2009;364(1535):3453–3458. doi:10.1098/rstb.2009.0142. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Grahlow M, Rupp CI, Derntl B. The impact of face masks on emotion recognition performance and perception of threat. PLoS ONE. 2022;17(2):e0262840. doi:10.1371/journal.pone.0262840. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Grenville E, Dwyer DM. Face masks have emotion-dependent dissociable effects on accuracy and confidence in identifying facial expressions of emotion. Cognitive Research: Principles and Implications. 2022;7(1):15. [PMC free article] [PubMed] [Google Scholar]
  • Grundmann F, Epstude K, Scheibe S. Face masks reduce emotion-recognition accuracy and perceived closeness. PLoS ONE. 2021;16(4):e0249792. doi:10.1371/journal.pone.0249792. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Guarnera M, Magnano P, Pellerone M, Cascio MI, Squatrito V, Buccheri SL. Facial expressions and the ability to recognize emotions from the eyes or mouth: A comparison among old adults, young adults, and children. The Journal of Genetic Psychology. 2018;179(5):297–310. doi:10.1080/00221325.2018.1509200. [PubMed] [CrossRef] [Google Scholar]
  • Guo K. Holistic gaze strategy to categorize facial expression of varying intensities. PLoS ONE. 2012;7(8):e42585. doi:10.1371/journal.pone.0042585. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Haxby JV, Gobbini MI. Distributed neural systems for face perception. Oxford University Press. 2011 doi:10.1093/oxfordhb/9780199559053.013.0006. [CrossRef] [Google Scholar]
  • Henry JD, von Hippel W, Molenberghs P, Lee T, Sachdev PS. Clinical assessment of social cognitive function in neurological disorders. Nature Reviews Neurology. 2016;12(1):28–39. doi:10.1038/nrneurol.2015.229. [PubMed] [CrossRef] [Google Scholar]
  • Higgins, W. C., Ross, R. M., Langdon, R., & Polito, V. (2022). The “Reading the Mind in the Eyes” test shows poor psychometric properties in a large, demographically representative US sample. Assessment. Advance online publication 10.1177/10731911221124342 [PubMed]
  • Hopwood CJ, Donnellan MB. How should the internal structure of personality inventories be evaluated? Personality and Social Psychology Review. 2010;14(3):332–346. doi:10.1177/1088868310361240. [PubMed] [CrossRef] [Google Scholar]
  • Jankowiak-Siuda K, Baron-Cohen S, Białaszek W, Dopierała A, Kozłowska A, Rymarczyk K. Psychometric evaluation of the 'reading the mind in the eyes’ test with samples of different ages from a Polish population. Studia Psychologica. 2016;58:18–31. doi:10.21909/sp.2016.01.704. [CrossRef] [Google Scholar]
  • Kastendieck T, Zillmer S, Hess U. (Un) mask yourself! effects of face masks on facial mimicry and emotion perception during the COVID-19 pandemic. Cognition and Emotion. 2022;36(1):59–69. doi:10.1080/02699931.2021.1950639. [PubMed] [CrossRef] [Google Scholar]
  • Kim, H., Kaduthodil, J., Strong, R. W., Germine, L., Cohan, S., & Wilmer, J. B. (2022, August 27). Multiracial reading the mind in the eyes test (MRMET): An inclusive version of an influential measure. 10.31219/osf.io/y8djm
  • Kittel AFD, Olderbak S, Wilhelm O. Sty in the mind’s eye: A meta-analytic investigation of the nomological network and internal consistency of the “reading the mind in the eyes” test. Assessment. 2022;29(5):872–895. doi:10.1177/1073191121996469. [PubMed] [CrossRef] [Google Scholar]
  • Kohler CG, Walker JB, Martin EA, Healey KM, Moberg PJ. Facial emotion perception in schizophrenia: A meta-analytic review. Schizophrenia Bulletin. 2010;36(5):1009–1019. doi:10.1093/schbul/sbn192. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Konrath S, Corneille O, Bushman BJ, Luminet O. The relationship between narcissistic exploitativeness, dispositional empathy, and emotion recognition abilities. Journal of Nonverbal Behavior. 2014;38(1):129–143. doi:10.1007/s10919-013-0164-y. [CrossRef] [Google Scholar]
  • Koo SJ, Kim YJ, Han JH, Seo E, Park HY, Bang M, An SK. “Reading the mind in the eyes test”: Translated and Korean versions. Psychiatry Investigation. 2021;18(4):295. doi:10.30773/pi.2020.0289. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Kulke L, Langer T, Valuch C. The emotional lockdown: How social distancing and mask wearing influence mood and emotion recognition in adolescents and adults. Frontiers in Psychology. 2022;13:878002. doi:10.3389/fpsyg.2022.878002. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Lafond-Brina G, Bonnefond A. The stability of multidimensional subclinical apathy during a pandemic and its relations to psycho-behavioral factors. Scientific Reports. 2022;12(1):2931. doi:10.1038/s41598-022-06777-5. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Langbehn AT, Yermol DA, Zhao F, Thorstenson CA, Niedenthal PM. Wearing N95, surgical, and cloth face masks compromises the perception of emotion. Affective Science. 2022;3(1):105–117. doi:10.1007/s42761-021-00097-z. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Lau WK, Huckauf A. Effects of face masks on the appearance of emotional expressions and invariant characteristics. Open Psychology. 2021;3(1):87–102. doi:10.1515/psych-2020-0113. [CrossRef] [Google Scholar]
  • Lawrence K, Campbell R, Skuse D. Age, gender, and puberty influence the development of facial emotion recognition. Frontiers in Psychology. 2015 doi:10.3389/fpsyg.2015.00761. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Leitner MC, Meurer V, Hutzler F, Schuster S, Hawelka S. The effect of masks on the recognition of facial expressions: A true-to-life study on the perception of basic emotions. Frontiers in Psychology. 2022;13:933438. doi:10.3389/fpsyg.2022.933438. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Leppänen JM, Nelson CA. Tuning the developing brain to social signals of emotions. Nature Reviews Neuroscience. 2009;10(1):37–47. doi:10.1038/nrn2554. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Lewis M. Fixing the stimulus-as-a-fixed-effect-fallacy in forensically valid face-composite research. Journal of Applied Research in Memory and Cognition. 2023 doi:10.1037/mac0000128. [CrossRef] [Google Scholar]
  • Livingstone SR, Russo FA. The ryerson audio-visual database of emotional speech and song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE. 2018;13(5):e0196391. doi:10.1371/journal.pone.0196391. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Lozier LM, Vanmeter JW, Marsh AA. Impairments in facial affect recognition associated with autism spectrum disorders: A meta-analysis. Development and Psychopathology. 2014;26(4pt1):933–945. doi:10.1017/S0954579414000479. [PubMed] [CrossRef] [Google Scholar]
  • MacCann C, Roberts RD. New paradigms for assessing emotional intelligence: Theory and data. Emotion. 2008;8:540–551. doi:10.1037/a0012746. [PubMed] [CrossRef] [Google Scholar]
  • Maddaluno O, Aiello EN, Roncoroni C, Prunas A, Bolognini N. The reading the mind in the eyes test, Iowa gambling task and interpersonal reactivity index: Normative data in an Italian population sample. Archives of Clinical Neuropsychology. 2022;37(5):929–938. doi:10.1093/arclin/acab100. [PubMed] [CrossRef] [Google Scholar]
  • Mair P, Wilcox R. Robust statistical methods in R using the WRS2 package. Behavior Research Methods. 2020;52:464–488. doi:10.3758/s13428-019-01246-w. [PubMed] [CrossRef] [Google Scholar]
  • MarcóGarcía S, Ferrer Quintero M, Usall J, Ochoa S, Del Cacho N, Huerta Ramos E. Reconocimiento facial de emociones en trastornos neurológicos: Una revisión narrativa. Revista De Neurología. 2019;69(05):207. doi:10.33588/rn.6905.2019047. [PubMed] [CrossRef] [Google Scholar]
  • Messick S. Meaning and values in test validation: The science and ethics of assessment. Educational Researcher. 1989;18(2):5–11. doi:10.2307/1175249. [CrossRef] [Google Scholar]
  • Milders M, Fuchs S, Crawford JR. Neuropsychological impairments and changes in emotional and social behaviour following severe traumatic brain injury. Journal of Clinical and Experimental Neuropsychology. 2003;25(2):157–172. doi:10.1076/jcen.25.2.157.13642. [PubMed] [CrossRef] [Google Scholar]
  • Murphy BA, Lilienfeld SO. Are self-report cognitive empathy ratings valid proxies for cognitive empathy ability? Negligible meta-analytic relations with behavioral task performance. Psychological Assessment. 2019;31(8):1062–1072. doi:10.1037/pas0000732. [PubMed] [CrossRef] [Google Scholar]
  • Neubauer AC, Hofer G. (Retest-)reliable and valid despite low alphas? An example from a typical performance situational judgment test of emotional management. Personality and Individual Differences. 2022;189:111511. doi:10.1016/j.paid.2022.111511. [CrossRef] [Google Scholar]
  • Oakley BF, Brewer R, Bird G, Catmur C. Theory of mind is not theory of emotion: A cautionary note on the reading the mind in the eyes test. Journal of Abnormal Psychology. 2016;125(6):818–823. doi:10.1037/abn0000182. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Olderbak S, Riggenmann O, Wilhelm O, Doebler P. Reliability generalization of tasks and recommendations for assessing the ability to perceive facial expressions of emotion. Psychological Assessment. 2021;33(10):911–926. doi:10.1037/pas0001030. [PubMed] [CrossRef] [Google Scholar]
  • Olderbak S, Wilhelm O, Olaru G, Geiger M, Brenneman MW, Roberts RD. A psychometric analysis of the reading the mind in the eyes test: Toward a brief form for research and applied settings. Frontiers in Psychology. 2015;6:1503. doi:10.3389/fpsyg.2015.01503. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Ong JH, Liu F. Frequent experience with face coverings for 10 months improves emotion perception among individuals with high autistic traits: A repeated cross-sectional study. Quarterly Journal of Experimental Psychology. 2022;76(7):1599–1608. doi:10.1177/17470218221135585. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Parada-Fernández P, Herrero-Fernández D, Jorge R, Comesaña P. Wearing mask hinders emotion recognition, but enhances perception of attractiveness. Personality and Individual Differences. 2022;184:111195. doi:10.1016/j.paid.2021.111195. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Pavlova MA, Sokolov AA. Reading language of the eyes. Neuroscience & Biobehavioral Reviews. 2022;140:104755. doi:10.1016/j.neubiorev.2022.104755. [PubMed] [CrossRef] [Google Scholar]
  • Pazhoohi F, Forby L, Kingstone A. Facial masks affect emotion recognition in the general population and individuals with autistic traits. PLoS ONE. 2021;16(9):e0257740. doi:10.1371/journal.pone.0257740. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Pedrabissi L, Santinello M. Verifica della validità dello STAI forma y di Spielberger [Verification of the validity of the STAI, Form Y, by Spielberger] Giunti Organizzazioni Speciali. 1989;191–192:11–14. [Google Scholar]
  • Radakovic R, Abrahams S. Developing a new apathy measurement scale: Dimensional apathy scale. Psychiatry Research. 2014;219(3):658–663. doi:10.1016/j.psychres.2014.06.010. [PubMed] [CrossRef] [Google Scholar]
  • Revelle, W. (2023). Psych package Version 2.3.6. Retrieved & July 2023 from https://cran.r-project.org/web/packages/psych/index.html
  • Robert P, Lanctôt KL, Agüera-Ortiz L, Aalten P, Bremond F, Defrancesco M, Manera V. Is it time to revise the diagnostic criteria for apathy in brain disorders? The 2018 international consensus group. European Psychiatry. 2018;54:71–76. doi:10.1016/j.eurpsy.2018.07.008. [PubMed] [CrossRef] [Google Scholar]
  • Sakaluk JK, Short SD. A methodological review of exploratory factor analysis in sexuality research: Used practices, best practices, and data analysis resources. The Journal of Sex Research. 2017;54(1):1–9. doi:10.1080/00224499.2015.1137538. [PubMed] [CrossRef] [Google Scholar]
  • Santangelo G, Raimo S, Siciliano M, D’Iorio A, Piscopo F, Cuoco S, Trojano L. Assessment of apathy independent of physical disability: Validation of the dimensional apathy scale in Italian healthy sample. Neurological Sciences. 2017;38:303–309. doi:10.1007/s10072-016-2766-8. [PubMed] [CrossRef] [Google Scholar]
  • Schlegel K, Fontaine JR, Scherer KR. The nomological network of emotion recognition ability: Evidence from the Geneva emotion recognition test. European Journal of Psychological Assessment. 2019;35(3):352–363. doi:10.1027/1015-5759/a000396. [CrossRef] [Google Scholar]
  • Serafin M, Surian L. Il test degli Occhi: Uno strumento per valutare la “teoria della mente” Giornale Italiano Di Psicologia. 2004;31(4):839–862. [Google Scholar]
  • Shek DT. The factorial structure of the Chinese version of the State-Trait Anxiety Inventory: A confirmatory factor analysis. Educational and Psychological Measurement. 1991;51(4):985–997. doi:10.1177/001316449105100418. [CrossRef] [Google Scholar]
  • Silvera DH, Martinussen M, Dahl TI. The Tromsø social intelligence scale, a self-report measure of social intelligence. Scandinavian Journal of Psychology. 2001;42:313–319. doi:10.1111/1467-9450.00242. [PubMed] [CrossRef] [Google Scholar]
  • Simonet DV, Miller KE, Askew KL, Sumner KE, Mortillaro M, Schlegel K. How multidimensional is emotional intelligence? Bifactor modeling of global and broad emotional abilities of the Geneva emotional competence test. Journal of Intelligence. 2021;9(1):14. doi:10.3390/jintelligence9010014. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Spielberger CD, Gorsuch RL, Lushene R, Vagg PR, Jacobs GA, editors. Manual for the State-Trait Anxiety Inventory. Consulting Psychologists Press; 1983. [Google Scholar]
  • Steger D, Schroeders U, Wilhelm O. On the dimensionality of crystallized intelligence: A smartphone-based assessment. Intelligence. 2019;72:76–85. doi:10.1016/j.intell.2018.12.002. [CrossRef] [Google Scholar]
  • Sullivan S, Ruffman T, Hutton SB. Age differences in emotion recognition skills and the visual scanning of emotion faces. The Journals of Gerontology Series b, Psychological Sciences and Social Sciences. 2007;62(1):P53-60. doi:10.1093/geronb/62.1.p53. [PubMed] [CrossRef] [Google Scholar]
  • Swain RH, O’Hare AJ, Brandley K, Gardner AT. Individual differences in social intelligence and perception of emotion expression of masked and unmasked faces. Cognitive Research: Principles and Implications. 2022;7(1):1–9. [PMC free article] [PubMed] [Google Scholar]
  • Trainin N, Yeshurun Y. Reading the mind with a mask? Improvement in reading the mind in the eyes during the COVID-19 pandemic. Emotion. 2021;21(8):1801–1806. doi:10.1037/emo0001014. [PubMed] [CrossRef] [Google Scholar]
  • Tsantani M, Podgajecka V, Gray KL, Cook R. How does the presence of a surgical face mask impair the perceived intensity of facial emotions? PLoS ONE. 2022;17(1):e0262344. doi:10.1371/journal.pone.0262344. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Vassallo S, Cooper SL, Douglas JM. Visual scanning in the recognition of facial affect: Is there an observer sex difference? Journal of Vision. 2009;9(3):11. doi:10.1167/9.3.11. [PubMed] [CrossRef] [Google Scholar]
  • Vellante M, Baron-Cohen S, Melis M, Marrone M, Petretto DR, Masala C, Preti A. The “Reading the Mind in the Eyes” test: Systematic review of psychometric properties and a validation study in Italy. Cognitive Neuropsychiatry. 2013;18(4):326–354. doi:10.1080/13546805.2012.721728. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Verroca A, de Rienzo CM, Gambarota F, Sessa P. Mapping the perception-space of facial expressions in the era of face masks. Frontiers in Psychology. 2022;13:956832. doi:10.3389/fpsyg.2022.956832. [PMC free article] [PubMed] [CrossRef] [Google Scholar]
  • Westfall J, Kenny DA, Judd CM. Statistical power and optimal design in experiments in which samples of participants respond to samples of stimuli. Journal of Experimental Psychology: General. 2014;143(5):2020. doi:10.1037/xge0000014. [PubMed] [CrossRef] [Google Scholar]
  • Wong HK, Estudillo AJ. Face masks affect emotion categorisation, age estimation, recognition, and gender classification from faces. Cognitive Research: Principles and Implications. 2022;7(1):91. [PMC free article] [PubMed] [Google Scholar]
  • Yuvaraj R, Murugappan M, Norlinah MI, Sundaraj K, Khairiyah M. Review of emotion recognition in stroke patients. Dementia and Geriatric Cognitive Disorders. 2013;36(3–4):179–196. doi:10.1159/000353440. [PubMed] [CrossRef] [Google Scholar]

Articles from Cognitive Research: Principles and Implications are provided here courtesy of Springer

Seeing emotions in the eyes: a validated test to study individual differences in the perception of basic emotions (2024)

References

Top Articles
Latest Posts
Article information

Author: Rob Wisoky

Last Updated:

Views: 6374

Rating: 4.8 / 5 (68 voted)

Reviews: 83% of readers found this page helpful

Author information

Name: Rob Wisoky

Birthday: 1994-09-30

Address: 5789 Michel Vista, West Domenic, OR 80464-9452

Phone: +97313824072371

Job: Education Orchestrator

Hobby: Lockpicking, Crocheting, Baton twirling, Video gaming, Jogging, Whittling, Model building

Introduction: My name is Rob Wisoky, I am a smiling, helpful, encouraging, zealous, energetic, faithful, fantastic person who loves writing and wants to share my knowledge and understanding with you.