{"title":"Dissociating model architectures from inference computations.","authors":"Noor Sajid, Johan Medrano","doi":"10.1080/17588928.2025.2532604","DOIUrl":"https://doi.org/10.1080/17588928.2025.2532604","url":null,"abstract":"<p><p>Parr et al., 2025 examines how auto-regressive and deep temporal models differ in their treatment of non-Markovian sequence modelling. Building on this, we highlight the need for dissociating model architectures-i.e., how the predictive distribution factorises-from the computations invoked at inference. We demonstrate that deep temporal computations are mimicked by autoregressive models by structuring context access during iterative inference. Using a transformer trained on next-token prediction, we show that inducing hierarchical temporal factorisation during iterative inference maintains predictive capacity while instantiating fewer computations. This emphasises that processes for constructing and refining predictions are not necessarily bound to their underlying model architectures.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-3"},"PeriodicalIF":2.0,"publicationDate":"2025-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144648756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ROSE: A Universal Neural Grammar.","authors":"Elliot Murphy","doi":"10.1080/17588928.2025.2523875","DOIUrl":"https://doi.org/10.1080/17588928.2025.2523875","url":null,"abstract":"<p><p>Processing natural language syntax requires a negotiation between symbolic and subsymbolic representations. Building on the recent representation, operation, structure, encoding (ROSE) neurocomputational architecture for syntax that scales from single units to inter-areal dynamics, I discuss the prospects of reconciling the neural code for hierarchical syntax with predictive processes. Here, the higher levels of ROSE provide instructions for symbolic phrase structure representations (S/E), while the lower levels provide probabilistic aspects of linguistic processing (R/O), with different types of cross-frequency coupling being hypothesized to interface these domains. I argue that ROSE provides a possible infrastructure for flexibly implementing distinct types of minimalist grammar parsers for the real-time processing of language. This perspective helps furnish a more restrictive 'core language network' in the brain than contemporary approaches that isolate general sentence composition. I define the language network as being critically involved in executing specific parsing operations (i.e. establishing phrasal categories, tree-structure depth, resolving dependencies, and retrieving proprietary lexical representations), capturing these network-defining operations jointly with probabilistic aspects of parsing. ROSE offers a 'mesoscopic protectorate' for natural language; an intermediate level of emergent organizational complexity that demands multi-scale modeling. By drawing principled relations across computational, algorithmic and implementational Marrian levels, ROSE offers new constraints on what a unified neurocomputational settlement for natural language syntax might look like, providing a tentative scaffold for a 'Universal Neural Grammar' - a species-specific format for neurally organizing the construction of compositional syntactic structures, which matures in accordance with a genetically determined biological matrix.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-32"},"PeriodicalIF":2.0,"publicationDate":"2025-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144625415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Embeddings as Dirichlet counts: Attention is the tip of the iceberg.","authors":"Alexander Bernard Kiefer","doi":"10.1080/17588928.2025.2530430","DOIUrl":"https://doi.org/10.1080/17588928.2025.2530430","url":null,"abstract":"<p><p>Despite the overtly discrete nature of language, the use of semantic embedding spaces is pervasive in modern computational linguistics and machine learning for natural language. I argue that this is intelligible if language is viewed as an interface into a general-purpose system of concepts, in which metric spaces capture rich relationships. At the same time, language embeddings can be regarded, at least heuristically, as equivalent to parameters of distributions over word-word relationships.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-3"},"PeriodicalIF":2.0,"publicationDate":"2025-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144590540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Non-Markovian systems, phenomenology, and the challenges of capturing meaning and context - comment on Parr, Pezzulo, and Friston (2025).","authors":"Mahault Albarracin, Dalton A R Sakthivadivel","doi":"10.1080/17588928.2025.2523889","DOIUrl":"10.1080/17588928.2025.2523889","url":null,"abstract":"<p><p>Parr, et al., explore the problem of non-Markovian pro cesses, in which the future state of a system depends not only on its present state but also on its past states. The authors suggest that the success of transformer networks in dealing with sequential data, such as language, stems from their ability to address this non-Markovian nature through the use of attention mechanisms. This commentary builds on their discussion, aiming to link it to some notions in Husserlian phenomenology and explore the implications for understanding meaning, context, and the nature of knowledge.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-2"},"PeriodicalIF":2.0,"publicationDate":"2025-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144559411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ryan Singh, Alexander Tschantz, Christopher L Buckley
{"title":"Paying attention to process.","authors":"Ryan Singh, Alexander Tschantz, Christopher L Buckley","doi":"10.1080/17588928.2025.2520313","DOIUrl":"https://doi.org/10.1080/17588928.2025.2520313","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-2"},"PeriodicalIF":2.0,"publicationDate":"2025-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144552432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Beyond prediction: comments on the format of natural intelligence.","authors":"Elliot Murphy","doi":"10.1080/17588928.2025.2521403","DOIUrl":"https://doi.org/10.1080/17588928.2025.2521403","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-4"},"PeriodicalIF":2.0,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144324646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Beyond individuals: Collective predictive coding for memory, attention, and the emergence of language.","authors":"Tadahiro Taniguchi","doi":"10.1080/17588928.2025.2518942","DOIUrl":"https://doi.org/10.1080/17588928.2025.2518942","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-2"},"PeriodicalIF":2.0,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144316040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"How deep will you go? Hierarchy in predictive coding and transformers.","authors":"Jeffrey F Queißer, Henrique Oyama, Jun Tani","doi":"10.1080/17588928.2025.2518945","DOIUrl":"https://doi.org/10.1080/17588928.2025.2518945","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-3"},"PeriodicalIF":2.0,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144316041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Auditory facilitation in deterministic versus stochastic worlds.","authors":"Berfin Bastug, Urte Roeber, Erich Schröger","doi":"10.1080/17588928.2025.2497762","DOIUrl":"https://doi.org/10.1080/17588928.2025.2497762","url":null,"abstract":"<p><p>The brain learns statistical regularities in sensory sequences, enhancing behavioral performance for predictable stimuli while impairing behavioral performance for unpredictable stimuli. While previous research has shown that violations of non-informative regularities hinder task performance, it remains unclear whether predictable but task-irrelevant structures can facilitate performance. In a tone duration discrimination task, we manipulated the task-irrelevant pitch dimension by varying transition probabilities (TP) between successive tone frequencies. Participants judged duration, while pitch sequences were either deterministic (a rule-governed pitch pattern, TP = 1) or stochastic (no discernible pitch pattern, TP = 1/number of pitch levels). The tone pitch was task-irrelevant and it did not predict duration. Results showed that reaction times (RTs) were significantly faster for deterministic sequences, suggesting that predictability in a task-irrelevant dimension still facilitates task performance. RTs were also faster in two-tone sequences compared to eight-tone sequences, likely due to reduced memory load. These findings suggest that statistical learning benefits extend beyond task-relevant dimensions, supporting a predictive coding framework in which the brain integrates predictable sensory input to optimize cognitive processing.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-7"},"PeriodicalIF":2.0,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143969935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Beyond Markov: Transformers, memory, and attention.","authors":"Thomas Parr, Giovanni Pezzulo, Karl Friston","doi":"10.1080/17588928.2025.2484485","DOIUrl":"https://doi.org/10.1080/17588928.2025.2484485","url":null,"abstract":"<p><p>This paper asks what predictive processing models of brain function can learn from the success of transformer architectures. We suggest that the reason transformer architectures have been successful is that they implicitly commit to a non-Markovian generative model - in which we need memory to contextualize our current observations and make predictions about the future. Interestingly, both the notions of working memory in cognitive science and transformer architectures rely heavily upon the concept of attention. We will argue that the move beyond Markov is crucial in the construction of generative models capable of dealing with much of the sequential data - and certainly language - that our brains contend with. We characterize two broad approaches to this problem - deep temporal hierarchies and autoregressive models - with transformers being an example of the latter. Our key conclusions are that transformers benefit heavily from their use of embedding spaces that place strong metric priors on an implicit latent variable and utilize this metric to direct a form of attention that highlights the most relevant, and not only the most recent, previous elements in a sequence to help predict the next.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"1-19"},"PeriodicalIF":2.0,"publicationDate":"2025-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143984568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}