{"title":"\"视觉动词\":动态事件类型是在视觉感知过程中自发提取的。","authors":"Huichao Ji, Brian J Scholl","doi":"10.1037/xge0001636","DOIUrl":null,"url":null,"abstract":"<p><p>During visual processing, input that is continuous in space and time is segmented, resulting in the representation of discrete tokens-objects or events. And there has been a great deal of research about how object representations are generalized into types-as when we see an object as an instance of a broader category (e.g., an animal or plant). There has been much less attention, however, to the possibility that vision represents dynamic information in terms of a small number of primitive event types (such as twisting or bouncing). (In models that posit a \"language of vision,\" these would be the foundational visual verbs.) Here we ask whether such event types are extracted spontaneously during visual perception, even when entirely task irrelevant during passive viewing. We exploited the phenomenon of categorical perception-wherein differences are more readily noticed when they are represented in terms of different underlying categories. Observers were better at detecting changes to images or short videos when the changes involved switches in the underlying event type-even when the changes that maintained the same event type were objectively larger (in terms of both brute image metrics and higher level feature change). We observed this categorical \"cross-event-type\" advantage for visual working memory for twisting versus rotating, scooping versus pouring, and rolling versus bouncing. Moreover, additional control experiments confirmed that such effects could not be explained by appeal to lower-level non-categorical stimulus differences. This spontaneous perception of \"visual verbs\" might promote both generalization and prediction about how events are likely to unfold. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":15698,"journal":{"name":"Journal of Experimental Psychology: General","volume":"153 10","pages":"2441-2453"},"PeriodicalIF":3.7000,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"\\\"Visual verbs\\\": Dynamic event types are extracted spontaneously during visual perception.\",\"authors\":\"Huichao Ji, Brian J Scholl\",\"doi\":\"10.1037/xge0001636\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>During visual processing, input that is continuous in space and time is segmented, resulting in the representation of discrete tokens-objects or events. And there has been a great deal of research about how object representations are generalized into types-as when we see an object as an instance of a broader category (e.g., an animal or plant). There has been much less attention, however, to the possibility that vision represents dynamic information in terms of a small number of primitive event types (such as twisting or bouncing). (In models that posit a \\\"language of vision,\\\" these would be the foundational visual verbs.) Here we ask whether such event types are extracted spontaneously during visual perception, even when entirely task irrelevant during passive viewing. We exploited the phenomenon of categorical perception-wherein differences are more readily noticed when they are represented in terms of different underlying categories. Observers were better at detecting changes to images or short videos when the changes involved switches in the underlying event type-even when the changes that maintained the same event type were objectively larger (in terms of both brute image metrics and higher level feature change). We observed this categorical \\\"cross-event-type\\\" advantage for visual working memory for twisting versus rotating, scooping versus pouring, and rolling versus bouncing. Moreover, additional control experiments confirmed that such effects could not be explained by appeal to lower-level non-categorical stimulus differences. This spontaneous perception of \\\"visual verbs\\\" might promote both generalization and prediction about how events are likely to unfold. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>\",\"PeriodicalId\":15698,\"journal\":{\"name\":\"Journal of Experimental Psychology: General\",\"volume\":\"153 10\",\"pages\":\"2441-2453\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2024-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Experimental Psychology: General\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.1037/xge0001636\",\"RegionNum\":1,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY, EXPERIMENTAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Experimental Psychology: General","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1037/xge0001636","RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, EXPERIMENTAL","Score":null,"Total":0}
"Visual verbs": Dynamic event types are extracted spontaneously during visual perception.
During visual processing, input that is continuous in space and time is segmented, resulting in the representation of discrete tokens-objects or events. And there has been a great deal of research about how object representations are generalized into types-as when we see an object as an instance of a broader category (e.g., an animal or plant). There has been much less attention, however, to the possibility that vision represents dynamic information in terms of a small number of primitive event types (such as twisting or bouncing). (In models that posit a "language of vision," these would be the foundational visual verbs.) Here we ask whether such event types are extracted spontaneously during visual perception, even when entirely task irrelevant during passive viewing. We exploited the phenomenon of categorical perception-wherein differences are more readily noticed when they are represented in terms of different underlying categories. Observers were better at detecting changes to images or short videos when the changes involved switches in the underlying event type-even when the changes that maintained the same event type were objectively larger (in terms of both brute image metrics and higher level feature change). We observed this categorical "cross-event-type" advantage for visual working memory for twisting versus rotating, scooping versus pouring, and rolling versus bouncing. Moreover, additional control experiments confirmed that such effects could not be explained by appeal to lower-level non-categorical stimulus differences. This spontaneous perception of "visual verbs" might promote both generalization and prediction about how events are likely to unfold. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
期刊介绍:
The Journal of Experimental Psychology: General publishes articles describing empirical work that bridges the traditional interests of two or more communities of psychology. The work may touch on issues dealt with in JEP: Learning, Memory, and Cognition, JEP: Human Perception and Performance, JEP: Animal Behavior Processes, or JEP: Applied, but may also concern issues in other subdisciplines of psychology, including social processes, developmental processes, psychopathology, neuroscience, or computational modeling. Articles in JEP: General may be longer than the usual journal publication if necessary, but shorter articles that bridge subdisciplines will also be considered.