William Dupont, Nicolas Amiez, Richard Palluel-Germain, Alain Martin, Marcela Perrone-Bertolotti, Carol Madden-Lombardi, Florent Lebon
{"title":"Neural Activation Down to the Spinal Cord during Action Language? A Transcranial Magnetic Stimulation and Peripheral Nerve Stimulation Study.","authors":"William Dupont, Nicolas Amiez, Richard Palluel-Germain, Alain Martin, Marcela Perrone-Bertolotti, Carol Madden-Lombardi, Florent Lebon","doi":"10.1162/jocn.a.83","DOIUrl":"https://doi.org/10.1162/jocn.a.83","url":null,"abstract":"<p><p>Language comprehension is increasingly recognized as extending beyond the traditional linguistic system to engage motor and perceptual processes. This perspective is supported by numerous studies demonstrating that understanding action-related words often induces behavioral and neurophysiological changes in the motor system. However, it remains unclear whether the influence of action language on the motor system is restricted to cortical regions or whether it also extends to spinal structures, as observed during motor imagery. To address this, we used TMS and peripheral nerve stimulation to assess corticospinal excitability and cortico-motoneuronal transmission, respectively. Fifteen healthy and right-handed volunteers participated in four conditions: (i) rest, (ii) kinesthetic motor imagery of finger and wrist flexion, (iii) reading action sentences, and (iv) reading nonaction sentences. As anticipated, corticospinal excitability increased during both kinesthetic motor imagery and action reading compared to rest. Interestingly, although kinesthetic motor imagery also led to the expected increase in cortico-motoneuronal transmission, no such modulation occurred during action reading. These findings suggest that action reading do not modulate the excitability of high-threshold motoneurons at the spinal level, contrary to motor imagery. Further investigation is needed to test whether action reading activates lower-threshold spinal structures, such as interneurons involved in spinal presynaptic inhibition.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-11"},"PeriodicalIF":3.0,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144755072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Neural Evidence for Tonal Prediction: Multivariate Decoding of Predicted Tone Categories Using Functional Magnetic Resonance Imaging Data.","authors":"Shun Liu, Wenjia Zhang, Suiping Wang","doi":"10.1162/jocn.a.84","DOIUrl":"https://doi.org/10.1162/jocn.a.84","url":null,"abstract":"<p><p>Predictive processing plays a central role in language comprehension, allowing listeners to generate predictions about upcoming linguistic input. Although considerable evidence supports segmental prediction, less is known about whether listeners can form predictions about suprasegmental features such as lexical tone. This study investigates whether listeners can generate and neurally represent predicted tonal information in the absence of auditory input. Using a Mandarin Chinese tone sandhi paradigm, we established tonal predictions based on sentence and visual context, recording brain activity with functional magnetic resonance imaging. Multivariate pattern analysis showed that predicted tonal categories could be decoded from brain activity even without tonal stimuli present. These representations were localized in auditory areas, articulatory motor regions, and the right cerebellum. We also found that predicted tone representations had distinct neural substrates compared to perceived tone representations. The study provides direct neural evidence that listeners can form representations of lexical tone in predictions of auditory input. The findings expand our understanding of suprasegmental prediction in speech and highlight the cerebellum's role in linguistic prediction.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-16"},"PeriodicalIF":3.0,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144755073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Incidental Encoding of Objects during Search Is Stronger Than Intentional Memorization due to Increased Recollection Rather Than Familiarity.","authors":"Jason Helbing, Dejan Draschkow, Melissa L-H Võ","doi":"10.1162/jocn.a.80","DOIUrl":"https://doi.org/10.1162/jocn.a.80","url":null,"abstract":"<p><p>Most memory is not formed deliberately but as a by-product of natural behavior. These incidental representations, when generated during visual search, can be stronger than intentionally memorized content (search superiority effect). However, it is unknown if the search superiority effect is purely quantitative (stronger memory) or also driven by differences in the degree of recollection and familiarity, two hallmark processes supporting recognition memory. Here, we use signal detection modeling, introspective judgments, event-related EEG potentials, and eye tracking measures to answer this question. In a preregistered study, 30 participants searched for objects in scenes and intentionally memorized others before completing a surprise recognition memory test. Behavioral data from remember-know judgments and receiver operating characteristics indicate that search targets were more often recollected compared with intentionally memorized objects, whereas the two tasks did not lead to differences in familiarity. Surprisingly, the neural signatures did not fully align with the behavioral findings regarding recollection and familiarity. That is, both search targets and intentionally memorized objects elicited a more positive-going mid-frontal negativity peaking at around 400 msec post stimulus onset (FN400), which is associated with familiarity, as well as a more positive-going parietal late component (LPC), indicative of recollection. Both components showed no differences between tasks, indicating equal contributions of recollection and familiarity to remembering searched and memorized objects. Furthermore, the LPC was, as expected, sensitive to differences between recollected and familiar objects when these were intentionally memorized, but it was not affected by these differences for searched objects. Overall, our findings indicate that search superiority relies predominantly on increased recollection. The fact that established neural markers of recollection (LPC) behaved as anticipated for intentionally memorized objects but carried no predictive power for incidentally memorized objects implies that memories established in more ecologically valid tasks might involve neural processes different from those activated in commonly used settings that are more reductionist.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-20"},"PeriodicalIF":3.1,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Neural Bases of Graphical Perception: A Novel Instance of Cultural Recycling?","authors":"Lorenzo Ciccione, Stanislas Dehaene","doi":"10.1162/jocn.a.81","DOIUrl":"https://doi.org/10.1162/jocn.a.81","url":null,"abstract":"<p><p>Graphical representations of quantitative data abound in our culture, and yet the brain mechanisms of graphicacy, by which viewers quickly extract statistical information from a data graphic, are unknown. Here, using scatterplots as stimuli, we tested two hypotheses about the brain areas underlying graphicacy. First, at the perceptual level, we hypothesized that the visual processing of scatterplots and their main trend recycles cortical regions devoted to the perception of the principal axis of objects. Second, at a higher level, we speculated that the math-responsive network active during arithmetic and mathematical truth judgments should also be involved in graphical perception. Using fMRI, we indeed found that the judgment of the trend in a scatterplot recruits a right lateral occipital area involved in detecting the orientation of objects, as well as a right anterior intraparietal region also recruited during mathematical tasks. Both behavior and brain activity were driven by the t value that indexes the statistical correlation in the data, and right intraparietal activation covaried with participants' graphicacy level. On the basis of this first approach to the neural bases of graphical perception, we suggest that, like literacy and numeracy, graphicacy relies on the recycling of brain areas previously attuned to a similar problem, here the perception of object orientation.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-18"},"PeriodicalIF":3.1,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Monolingual, Non-tone Bilingual, and Tone Bilingual Infants: Language Experiences Alter Speech and Nonspeech Perception.","authors":"Liquan Liu, Varghese Peter, Zhen Zeng, Gabrielle Weidemann","doi":"10.1162/jocn.a.76","DOIUrl":"https://doi.org/10.1162/jocn.a.76","url":null,"abstract":"<p><p>Studies on first-year infants' pitch perception have witnessed shifts of perceptual focus from acoustic to linguistic information and from a wide range of contrasts to those relevant to their native language. Nevertheless, how linguistic experience interacts with this developmental process remains an open question. This study compared the neural discrimination of speech/lexical and nonspeech/violin tone contrasts by 5- to 6- and 11- to 12-month-old infants across three types of language backgrounds: monolingual infants learning a non-tone language (Mono), bilingual infants learning two non-tone languages (Bi-NT), and bilingual infants learning a non-tone and a tone language (Bi-Tone). Although Mono infants do not show significant responses to the lexical tone contrast, both Bi-NT and Bi-Tone infants showed positive mismatch responses at both ages, indicating an enhancement effect brought by a complex language environment as early as 5 months after birth. Regarding the violin tone perception, distinct patterns were observed across language backgrounds: a perceptual decrease for Mono infants, no significant response for Bi-NT infants, and a perceptual increase for Bi-Tone infants over the first year. These patterns suggest that pitch perception may be affected across domains by language experiences at this stage, where interactions in cognitive processing between speech and nonspeech prosodic information may occur.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-16"},"PeriodicalIF":3.1,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nancy A Dennis, Amy A Overman, Catherine M Carpenter, Alexa Becker, John T West, Spencer O Chase
{"title":"The Otter and the Cleaver: Exploring the Neural Underpinnings of Unitization Using the Gestalt Principle of Proximity.","authors":"Nancy A Dennis, Amy A Overman, Catherine M Carpenter, Alexa Becker, John T West, Spencer O Chase","doi":"10.1162/jocn.a.78","DOIUrl":"10.1162/jocn.a.78","url":null,"abstract":"<p><p>Associative memory requires the binding of multiple objects into a single representation in memory. As such, associative memory is viewed as harder and more resource demanding than item memory. One means of facilitating associative memory is through the process of unitization. Previous work has suggested that, once unitized, discrete stimuli are processed as a single ensemble. The present study aims to test whether the grouping principle of proximity enhances associative memory by creating unitized representations of item pairs that resemble how single items are processed in memory in younger adults. To examine the neural basis of perceptual unitization, young adult participants encoded unrelated object pairs either proximally arranged (unitized condition) or distally arranged (associative condition) as well as single objects. Behaviorally, results showed enhanced memory for proximally organized object pairs compared with distally organized object pairs. Examination of neural pattern similarity across conditions suggested that although regions critical to associative memory processed the proximal object pairs more similarly to the distal object pairs, clusters of neural activity throughout cortex did show greater similarity of neural patterns across proximal pairs and single objects during both encoding and retrieval. Results suggest that the simple act of configural placement is enough to initiate unitized-based encoding and maintain that representation at retrieval.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-21"},"PeriodicalIF":3.0,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Joan Tarrida, Manuel Moreno, Jordi Vidal, David Panyella, Josep Marco-Pallarés, Lluís Fuentemilla
{"title":"Odor-induced Sustained Neural Activity during Memory Encoding.","authors":"Joan Tarrida, Manuel Moreno, Jordi Vidal, David Panyella, Josep Marco-Pallarés, Lluís Fuentemilla","doi":"10.1162/jocn.a.77","DOIUrl":"https://doi.org/10.1162/jocn.a.77","url":null,"abstract":"<p><p>How long do the neural and cognitive effects of a brief odor experience last? This study investigated whether short exposures to pleasant and unpleasant odors can induce sustained changes in brain activity and influence memory formation for events occurring several seconds later. Using EEG, we combined univariate ERP analyses with time-resolved multivariate decoding to track neural responses during a 6-sec delay between odor presentation and visual memory encoding. We found that brief odor cues elicited sustained neural activity that persisted well beyond odor offset. Unpleasant odors, in particular, were associated with higher sustained ERP amplitudes compared with pleasant ones. Behaviorally, participants showed greater confidence in recognizing images that had been preceded by unpleasant odors, suggesting that even brief olfactory experiences can modulate memory encoding for temporally distant events. These findings demonstrate that brief olfactory cues have a lasting effect on both neural activity and subsequent memory performance.</p>","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":" ","pages":"1-10"},"PeriodicalIF":3.1,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Electroencephalographic Decoding of Conscious versus Unconscious Representations during Binocular Rivalry","authors":"Lara C. Krisst;Steven J. Luck","doi":"10.1162/jocn_a_02308","DOIUrl":"10.1162/jocn_a_02308","url":null,"abstract":"Theories of visual awareness often fall into two general categories, those assuming that awareness arises rapidly within visual cortex and those assuming that awareness arises more slowly as a result of interactions between visual cortex and frontoparietal regions. To test the plausibility of early theories of consciousness, we combined the temporal resolution of the EEG technique with multivariate pattern classification techniques to assess the latency at which decodable information about consciously perceived stimuli is enhanced relative to information about stimuli that are not consciously perceived. Competing red and green gratings were presented simultaneously to the two eyes, creating rivalry, and observers reported which one of the two colors was perceived on each trial. We then used the pattern of EEG over the scalp to decode the orientation of the grating that was perceived and the orientation of the grating that was suppressed by the rivalry and not perceived. This allowed us to determine when the content of the neural representations differed between the consciously perceived grating and the unconscious grating. Early theories predict that the difference between conscious and unconscious processing would occur within ∼200 msec of stimulus onset (e.g., at the time of the visual awareness negativity). We found that decoding accuracy was significantly greater for the consciously perceived orientation than for the unperceived orientation beginning 160 msec after stimulus onset, as predicted by theories that propose a rapid onset of visual awareness.","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":"37 8","pages":"1381-1390"},"PeriodicalIF":3.1,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143371347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Internal State Estimation via Physiological Data and Its Modulation by Environmental Context during Social Activity","authors":"Ayumu Yamashita;Hiroki Maeda;Jouh Yeong Chew;Kaoru Amano","doi":"10.1162/jocn_a_02315","DOIUrl":"10.1162/jocn_a_02315","url":null,"abstract":"Understanding individuals' internal cognitive states during group interactions is crucial for enhancing group dynamics and communication. This study investigated internal states by analyzing physiological data—EEG, electrocardiography, and pupil size—collected from high school students during group discussions. Using a data-driven clustering method, we identified four distinct internal states, each corresponding to the different power distributions in the four frequency bands of EEG activity, heart rate variability, and pupil size. These states were not only associated with verbal behavior such as “speaking”, they were also associated with nonverbal cues such as “gazing at faces” and other body languages. We also examined the influence of environmental factors on internal states, including the presence of a facilitator and the group size. The presence of a facilitator significantly increased the probability of participants remaining in the high alpha-power state, possibly reflecting a relaxed or moderately aroused state. This study provides insights into the physiological underpinnings of group interactions, which can be leveraged to improve educational settings and other group-based activities.","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":"37 8","pages":"1364-1380"},"PeriodicalIF":3.1,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143505869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Salient, Unexpected Omissions of Sounds Can Involuntarily Distract Attention","authors":"Valeria Baragona;Erich Schröger;Andreas Widmann","doi":"10.1162/jocn_a_02307","DOIUrl":"10.1162/jocn_a_02307","url":null,"abstract":"Salient unexpected and task-irrelevant sounds can act as distractors by capturing attention away from a task. Consequently, a performance impairment (e.g., prolonged RTs) is typically observed along with a pupil dilation response (PDR) and the P3a ERP component. Previous results showed prolonged RTs in response to task-relevant visual stimuli also following unexpected sound omissions. However, it was unclear whether this was due to the absence of the sound's warning effect or to distraction caused by the violation of a sensory prediction. In our paradigm, participants initiated a trial through a button press that elicited either a regular sound (80%), a deviant sound (10%), or no sound (10%). Thereafter, a digit was presented visually, and the participant had to classify it as even or odd. To dissociate warning and distraction effects, we additionally included a control condition in which a button press never generated a sound, and therefore no sound was expected. Results show that, compared with expected events, unexpected deviants and omissions lead to prolonged RTs (distraction effect), enlarged PDR, and a P3a-like ERP effect. Moreover, sound events, compared with no sound events, yielded faster RTs (warning effect), larger PDR, and increased P3a. Overall, we observed a co-occurrence of warning and distraction effects. This suggests that not only unexpected sounds but also unexpected sound omissions can act as salient distractors. This finding supports theories claiming that involuntary attention is based on prediction violation.","PeriodicalId":51081,"journal":{"name":"Journal of Cognitive Neuroscience","volume":"37 8","pages":"1291-1307"},"PeriodicalIF":3.1,"publicationDate":"2025-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11080645","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143371354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}