"Visual verbs": Dynamic event types are extracted spontaneously during visual perception.

IF 4.3 3区 材料科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Huichao Ji, Brian J Scholl
{"title":"\"Visual verbs\": Dynamic event types are extracted spontaneously during visual perception.","authors":"Huichao Ji, Brian J Scholl","doi":"10.1037/xge0001636","DOIUrl":null,"url":null,"abstract":"<p><p>During visual processing, input that is continuous in space and time is segmented, resulting in the representation of discrete tokens-objects or events. And there has been a great deal of research about how object representations are generalized into types-as when we see an object as an instance of a broader category (e.g., an animal or plant). There has been much less attention, however, to the possibility that vision represents dynamic information in terms of a small number of primitive event types (such as twisting or bouncing). (In models that posit a \"language of vision,\" these would be the foundational visual verbs.) Here we ask whether such event types are extracted spontaneously during visual perception, even when entirely task irrelevant during passive viewing. We exploited the phenomenon of categorical perception-wherein differences are more readily noticed when they are represented in terms of different underlying categories. Observers were better at detecting changes to images or short videos when the changes involved switches in the underlying event type-even when the changes that maintained the same event type were objectively larger (in terms of both brute image metrics and higher level feature change). We observed this categorical \"cross-event-type\" advantage for visual working memory for twisting versus rotating, scooping versus pouring, and rolling versus bouncing. Moreover, additional control experiments confirmed that such effects could not be explained by appeal to lower-level non-categorical stimulus differences. This spontaneous perception of \"visual verbs\" might promote both generalization and prediction about how events are likely to unfold. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":3,"journal":{"name":"ACS Applied Electronic Materials","volume":null,"pages":null},"PeriodicalIF":4.3000,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Applied Electronic Materials","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1037/xge0001636","RegionNum":3,"RegionCategory":"材料科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

During visual processing, input that is continuous in space and time is segmented, resulting in the representation of discrete tokens-objects or events. And there has been a great deal of research about how object representations are generalized into types-as when we see an object as an instance of a broader category (e.g., an animal or plant). There has been much less attention, however, to the possibility that vision represents dynamic information in terms of a small number of primitive event types (such as twisting or bouncing). (In models that posit a "language of vision," these would be the foundational visual verbs.) Here we ask whether such event types are extracted spontaneously during visual perception, even when entirely task irrelevant during passive viewing. We exploited the phenomenon of categorical perception-wherein differences are more readily noticed when they are represented in terms of different underlying categories. Observers were better at detecting changes to images or short videos when the changes involved switches in the underlying event type-even when the changes that maintained the same event type were objectively larger (in terms of both brute image metrics and higher level feature change). We observed this categorical "cross-event-type" advantage for visual working memory for twisting versus rotating, scooping versus pouring, and rolling versus bouncing. Moreover, additional control experiments confirmed that such effects could not be explained by appeal to lower-level non-categorical stimulus differences. This spontaneous perception of "visual verbs" might promote both generalization and prediction about how events are likely to unfold. (PsycInfo Database Record (c) 2024 APA, all rights reserved).

"视觉动词":动态事件类型是在视觉感知过程中自发提取的。
在视觉处理过程中,在空间和时间上连续的输入会被分割,从而形成离散的标记--物体或事件。关于物体表征如何被概括为类型的研究也非常多--比如当我们把一个物体看成一个更广泛类别(如动物或植物)的实例时。然而,对于视觉以少量原始事件类型(如扭曲或弹跳)来表征动态信息的可能性,人们的关注要少得多。(在假定有 "视觉语言 "的模型中,这些就是基本的视觉动词)。在这里,我们要问的是,这些事件类型是否会在视觉感知过程中被自发提取出来,即使在被动观看过程中与任务完全无关。我们利用了分类感知现象--当差异以不同的基本类别表示时,它们更容易被注意到。当图像或短视频的变化涉及底层事件类型的转换时,观察者能更好地检测到图像或短视频的变化--即使保持相同事件类型的变化在客观上更大(就原始图像指标和更高层次的特征变化而言)。我们在扭转与旋转、舀水与倒水、滚动与弹跳的视觉工作记忆中观察到了这种分类 "跨事件类型 "优势。此外,额外的对照实验证实,这种效应无法用低层次的非分类刺激差异来解释。这种对 "视觉动词 "的自发感知可能会促进对事件可能发生方式的概括和预测。(PsycInfo Database Record (c) 2024 APA, 版权所有)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.20
自引率
4.30%
发文量
567
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信