{"title":"Shared neural dynamics of facial expression processing.","authors":"Madeline Molly Ely, Géza Gergely Ambrus","doi":"10.1007/s11571-025-10230-4","DOIUrl":null,"url":null,"abstract":"<p><p>The ability to recognize and interpret facial expressions is fundamental to human social cognition, enabling navigation of complex interpersonal interactions and understanding of others' emotional states. The extent to which neural patterns associated with facial expression processing are shared between observers remains unexplored, and no study has yet examined the neural dynamics specific to different emotional expressions. Additionally, the neural processing dynamics of facial attributes such as sex and identity in relation to facial expressions have not been thoroughly investigated. In this study, we investigated the shared neural dynamics of emotional face processing using an explicit facial emotion recognition task, where participants made two-alternative forced choice (2AFC) decisions on the displayed emotion. Our data-driven approach employed cross-participant multivariate classification and representational dissimilarity analysis on EEG data. The results demonstrate that EEG signals can effectively decode the sex, emotional expression, and identity of face stimuli across different stimuli and participants, indicating shared neural codes for facial expression processing. Multivariate classification analyses revealed that sex is decoded first, followed by identity, and then emotion. Emotional expressions (angry, happy, sad) were decoded earlier when contrasted with neutral expressions. While identity and sex information were modulated by image-level stimulus features, the effects of emotion were independent of visual image properties. Importantly, our findings suggest enhanced processing of face identity and sex for emotional expressions, particularly for angry faces and, to a lesser extent, happy faces.</p><p><strong>Supplementary information: </strong>The online version contains supplementary material available at 10.1007/s11571-025-10230-4.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"45"},"PeriodicalIF":3.1000,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11880506/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cognitive Neurodynamics","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s11571-025-10230-4","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/4 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
The ability to recognize and interpret facial expressions is fundamental to human social cognition, enabling navigation of complex interpersonal interactions and understanding of others' emotional states. The extent to which neural patterns associated with facial expression processing are shared between observers remains unexplored, and no study has yet examined the neural dynamics specific to different emotional expressions. Additionally, the neural processing dynamics of facial attributes such as sex and identity in relation to facial expressions have not been thoroughly investigated. In this study, we investigated the shared neural dynamics of emotional face processing using an explicit facial emotion recognition task, where participants made two-alternative forced choice (2AFC) decisions on the displayed emotion. Our data-driven approach employed cross-participant multivariate classification and representational dissimilarity analysis on EEG data. The results demonstrate that EEG signals can effectively decode the sex, emotional expression, and identity of face stimuli across different stimuli and participants, indicating shared neural codes for facial expression processing. Multivariate classification analyses revealed that sex is decoded first, followed by identity, and then emotion. Emotional expressions (angry, happy, sad) were decoded earlier when contrasted with neutral expressions. While identity and sex information were modulated by image-level stimulus features, the effects of emotion were independent of visual image properties. Importantly, our findings suggest enhanced processing of face identity and sex for emotional expressions, particularly for angry faces and, to a lesser extent, happy faces.
Supplementary information: The online version contains supplementary material available at 10.1007/s11571-025-10230-4.
期刊介绍:
Cognitive Neurodynamics provides a unique forum of communication and cooperation for scientists and engineers working in the field of cognitive neurodynamics, intelligent science and applications, bridging the gap between theory and application, without any preference for pure theoretical, experimental or computational models.
The emphasis is to publish original models of cognitive neurodynamics, novel computational theories and experimental results. In particular, intelligent science inspired by cognitive neuroscience and neurodynamics is also very welcome.
The scope of Cognitive Neurodynamics covers cognitive neuroscience, neural computation based on dynamics, computer science, intelligent science as well as their interdisciplinary applications in the natural and engineering sciences. Papers that are appropriate for non-specialist readers are encouraged.
1. There is no page limit for manuscripts submitted to Cognitive Neurodynamics. Research papers should clearly represent an important advance of especially broad interest to researchers and technologists in neuroscience, biophysics, BCI, neural computer and intelligent robotics.
2. Cognitive Neurodynamics also welcomes brief communications: short papers reporting results that are of genuinely broad interest but that for one reason and another do not make a sufficiently complete story to justify a full article publication. Brief Communications should consist of approximately four manuscript pages.
3. Cognitive Neurodynamics publishes review articles in which a specific field is reviewed through an exhaustive literature survey. There are no restrictions on the number of pages. Review articles are usually invited, but submitted reviews will also be considered.