{"title":"Universal dimensions of visual representation","authors":"Zirui Chen, Michael F. Bonner","doi":"10.1126/sciadv.adw7697","DOIUrl":null,"url":null,"abstract":"<div >Do visual neural networks learn brain-aligned representations because they share architectural constraints and task objectives with biological vision or because they share universal features of natural image processing? We characterized the universality of hundreds of thousands of representational dimensions from networks with different architectures, tasks, and training data. We found that diverse networks learn to represent natural images using a shared set of latent dimensions, despite having highly distinct designs. Next, by comparing these networks with human brain representations measured with functional magnetic resonance imaging, we found that the most brain-aligned representations in neural networks are those that are universal and independent of a network’s specific characteristics. Each network can be reduced to fewer than 10 of its most universal dimensions with little impact on its representational similarity to the brain. These results suggest that the underlying similarities between artificial and biological vision are primarily governed by a core set of universal representations that are convergently learned by diverse systems.</div>","PeriodicalId":21609,"journal":{"name":"Science Advances","volume":"11 27","pages":""},"PeriodicalIF":12.5000,"publicationDate":"2025-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.science.org/doi/reader/10.1126/sciadv.adw7697","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Science Advances","FirstCategoryId":"103","ListUrlMain":"https://www.science.org/doi/10.1126/sciadv.adw7697","RegionNum":1,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
Do visual neural networks learn brain-aligned representations because they share architectural constraints and task objectives with biological vision or because they share universal features of natural image processing? We characterized the universality of hundreds of thousands of representational dimensions from networks with different architectures, tasks, and training data. We found that diverse networks learn to represent natural images using a shared set of latent dimensions, despite having highly distinct designs. Next, by comparing these networks with human brain representations measured with functional magnetic resonance imaging, we found that the most brain-aligned representations in neural networks are those that are universal and independent of a network’s specific characteristics. Each network can be reduced to fewer than 10 of its most universal dimensions with little impact on its representational similarity to the brain. These results suggest that the underlying similarities between artificial and biological vision are primarily governed by a core set of universal representations that are convergently learned by diverse systems.
期刊介绍:
Science Advances, an open-access journal by AAAS, publishes impactful research in diverse scientific areas. It aims for fair, fast, and expert peer review, providing freely accessible research to readers. Led by distinguished scientists, the journal supports AAAS's mission by extending Science magazine's capacity to identify and promote significant advances. Evolving digital publishing technologies play a crucial role in advancing AAAS's global mission for science communication and benefitting humankind.