是什么让一张脸看起来像一顶帽子用图像三胞胎解耦低级和高级视觉属性

Maytus Piriyajitakonkij, Sirawaj Itthipuripat, Ian Ballard, Ioannis Pappas
{"title":"是什么让一张脸看起来像一顶帽子用图像三胞胎解耦低级和高级视觉属性","authors":"Maytus Piriyajitakonkij, Sirawaj Itthipuripat, Ian Ballard, Ioannis Pappas","doi":"arxiv-2409.02241","DOIUrl":null,"url":null,"abstract":"In visual decision making, high-level features, such as object categories,\nhave a strong influence on choice. However, the impact of low-level features on\nbehavior is less understood partly due to the high correlation between high-\nand low-level features in the stimuli presented (e.g., objects of the same\ncategory are more likely to share low-level features). To disentangle these\neffects, we propose a method that de-correlates low- and high-level visual\nproperties in a novel set of stimuli. Our method uses two Convolutional Neural\nNetworks (CNNs) as candidate models of the ventral visual stream: the CORnet-S\nthat has high neural predictivity in high-level, IT-like responses and the\nVGG-16 that has high neural predictivity in low-level responses. Triplets\n(root, image1, image2) of stimuli are parametrized by the level of low- and\nhigh-level similarity of images extracted from the different layers. These\nstimuli are then used in a decision-making task where participants are tasked\nto choose the most similar-to-the-root image. We found that different networks\nshow differing abilities to predict the effects of low-versus-high-level\nsimilarity: while CORnet-S outperforms VGG-16 in explaining human choices based\non high-level similarity, VGG-16 outperforms CORnet-S in explaining human\nchoices based on low-level similarity. Using Brain-Score, we observed that the\nbehavioral prediction abilities of different layers of these networks\nqualitatively corresponded to their ability to explain neural activity at\ndifferent levels of the visual hierarchy. In summary, our algorithm for\nstimulus set generation enables the study of how different representations in\nthe visual stream affect high-level cognitive behaviors.","PeriodicalId":501517,"journal":{"name":"arXiv - QuanBio - Neurons and Cognition","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"What makes a face looks like a hat: Decoupling low-level and high-level Visual Properties with Image Triplets\",\"authors\":\"Maytus Piriyajitakonkij, Sirawaj Itthipuripat, Ian Ballard, Ioannis Pappas\",\"doi\":\"arxiv-2409.02241\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In visual decision making, high-level features, such as object categories,\\nhave a strong influence on choice. However, the impact of low-level features on\\nbehavior is less understood partly due to the high correlation between high-\\nand low-level features in the stimuli presented (e.g., objects of the same\\ncategory are more likely to share low-level features). To disentangle these\\neffects, we propose a method that de-correlates low- and high-level visual\\nproperties in a novel set of stimuli. Our method uses two Convolutional Neural\\nNetworks (CNNs) as candidate models of the ventral visual stream: the CORnet-S\\nthat has high neural predictivity in high-level, IT-like responses and the\\nVGG-16 that has high neural predictivity in low-level responses. Triplets\\n(root, image1, image2) of stimuli are parametrized by the level of low- and\\nhigh-level similarity of images extracted from the different layers. These\\nstimuli are then used in a decision-making task where participants are tasked\\nto choose the most similar-to-the-root image. We found that different networks\\nshow differing abilities to predict the effects of low-versus-high-level\\nsimilarity: while CORnet-S outperforms VGG-16 in explaining human choices based\\non high-level similarity, VGG-16 outperforms CORnet-S in explaining human\\nchoices based on low-level similarity. Using Brain-Score, we observed that the\\nbehavioral prediction abilities of different layers of these networks\\nqualitatively corresponded to their ability to explain neural activity at\\ndifferent levels of the visual hierarchy. In summary, our algorithm for\\nstimulus set generation enables the study of how different representations in\\nthe visual stream affect high-level cognitive behaviors.\",\"PeriodicalId\":501517,\"journal\":{\"name\":\"arXiv - QuanBio - Neurons and Cognition\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuanBio - Neurons and Cognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.02241\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuanBio - Neurons and Cognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.02241","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在视觉决策中,物体类别等高层次特征对选择有很大影响。然而,低层次特征对行为的影响却不那么为人所知,部分原因在于所呈现的刺激物中高低层次特征之间的高度相关性(例如,同一类别的物体更有可能具有相同的低层次特征)。为了将这些效应区分开来,我们提出了一种在一组新的刺激中消除低级和高级视觉特性相关性的方法。我们的方法使用两个卷积神经网络(CNN)作为腹侧视觉流的候选模型:CORnet-S 在高层次、类似 IT 的反应中具有较高的神经预测性,而 VGG-16 在低层次反应中具有较高的神经预测性。刺激的三胞胎(根、图像 1、图像 2)由从不同层提取的图像的低层和高层相似性水平参数化。然后将这些刺激用于决策任务中,让参与者选择与根图像最相似的图像。我们发现,不同的网络在预测低级与高级相似性的影响方面表现出不同的能力:CORnet-S 在解释人类基于高级相似性的选择方面优于 VGG-16,而 VGG-16 在解释人类基于低级相似性的选择方面优于 CORnet-S。利用 Brain-Score,我们观察到这些网络的不同层的行为预测能力与它们解释视觉层次结构中不同层次的神经活动的能力是定性对应的。总之,我们的刺激集生成算法可以研究视觉流中的不同表征是如何影响高层次认知行为的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
What makes a face looks like a hat: Decoupling low-level and high-level Visual Properties with Image Triplets
In visual decision making, high-level features, such as object categories, have a strong influence on choice. However, the impact of low-level features on behavior is less understood partly due to the high correlation between high- and low-level features in the stimuli presented (e.g., objects of the same category are more likely to share low-level features). To disentangle these effects, we propose a method that de-correlates low- and high-level visual properties in a novel set of stimuli. Our method uses two Convolutional Neural Networks (CNNs) as candidate models of the ventral visual stream: the CORnet-S that has high neural predictivity in high-level, IT-like responses and the VGG-16 that has high neural predictivity in low-level responses. Triplets (root, image1, image2) of stimuli are parametrized by the level of low- and high-level similarity of images extracted from the different layers. These stimuli are then used in a decision-making task where participants are tasked to choose the most similar-to-the-root image. We found that different networks show differing abilities to predict the effects of low-versus-high-level similarity: while CORnet-S outperforms VGG-16 in explaining human choices based on high-level similarity, VGG-16 outperforms CORnet-S in explaining human choices based on low-level similarity. Using Brain-Score, we observed that the behavioral prediction abilities of different layers of these networks qualitatively corresponded to their ability to explain neural activity at different levels of the visual hierarchy. In summary, our algorithm for stimulus set generation enables the study of how different representations in the visual stream affect high-level cognitive behaviors.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信