The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks.

Computational brain & behavior Pub Date : 2021-01-01 Epub Date: 2021-02-12 DOI:10.1007/s42113-021-00098-y
Xiaoliang Luo, Brett D Roads, Bradley C Love
{"title":"The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks.","authors":"Xiaoliang Luo,&nbsp;Brett D Roads,&nbsp;Bradley C Love","doi":"10.1007/s42113-021-00098-y","DOIUrl":null,"url":null,"abstract":"<p><p>People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention in categorisation models, we developed a goal-directed attention mechanism that can process naturalistic (photographic) stimuli. Our attention mechanism can be incorporated into any existing deep convolutional neural networks (DCNNs). The processing stages in DCNNs have been related to ventral visual stream. In that light, our attentional mechanism incorporates top-down influences from prefrontal cortex (PFC) to support goal-directed behaviour. Akin to how attention weights in categorisation models warp representational spaces, we introduce a layer of attention weights to the mid-level of a DCNN that amplify or attenuate activity to further a goal. We evaluated the attentional mechanism using photographic stimuli, varying the attentional target. We found that increasing goal-directed attention has benefits (increasing hit rates) and costs (increasing false alarm rates). At a moderate level, attention improves sensitivity (i.e. increases <math> <msup><mrow><mi>d</mi></mrow> <mrow><mi>'</mi></mrow> </msup> </math> ) at only a moderate increase in bias for tasks involving standard images, blended images and natural adversarial images chosen to fool DCNNs. These results suggest that goal-directed attention can reconfigure general-purpose DCNNs to better suit the current task goal, much like PFC modulates activity along the ventral stream. In addition to being more parsimonious and brain consistent, the mid-level attention approach performed better than a standard machine learning approach for transfer learning, namely retraining the final network layer to accommodate the new task.</p>","PeriodicalId":72660,"journal":{"name":"Computational brain & behavior","volume":"4 2","pages":"213-230"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s42113-021-00098-y","citationCount":"13","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computational brain & behavior","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s42113-021-00098-y","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2021/2/12 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 13

Abstract

People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention in categorisation models, we developed a goal-directed attention mechanism that can process naturalistic (photographic) stimuli. Our attention mechanism can be incorporated into any existing deep convolutional neural networks (DCNNs). The processing stages in DCNNs have been related to ventral visual stream. In that light, our attentional mechanism incorporates top-down influences from prefrontal cortex (PFC) to support goal-directed behaviour. Akin to how attention weights in categorisation models warp representational spaces, we introduce a layer of attention weights to the mid-level of a DCNN that amplify or attenuate activity to further a goal. We evaluated the attentional mechanism using photographic stimuli, varying the attentional target. We found that increasing goal-directed attention has benefits (increasing hit rates) and costs (increasing false alarm rates). At a moderate level, attention improves sensitivity (i.e. increases d ' ) at only a moderate increase in bias for tasks involving standard images, blended images and natural adversarial images chosen to fool DCNNs. These results suggest that goal-directed attention can reconfigure general-purpose DCNNs to better suit the current task goal, much like PFC modulates activity along the ventral stream. In addition to being more parsimonious and brain consistent, the mid-level attention approach performed better than a standard machine learning approach for transfer learning, namely retraining the final network layer to accommodate the new task.

Abstract Image

Abstract Image

Abstract Image

深度卷积神经网络中目标导向注意力的成本与收益。
人们利用自上而下、目标导向的注意力来完成任务,比如寻找丢失的钥匙。通过将视觉系统调整到相关的信息源,物体识别可以变得更有效(一个好处),更偏向于目标(一个潜在的成本)。在分类模型的选择性注意的激励下,我们开发了一个目标导向的注意机制,可以处理自然(摄影)刺激。我们的注意机制可以被整合到任何现有的深度卷积神经网络(DCNNs)中。DCNNs的加工阶段与腹侧视觉流有关。从这个角度来看,我们的注意机制结合了来自前额皮质(PFC)的自上而下的影响,以支持目标导向的行为。类似于分类模型中的注意力权重如何扭曲表征空间,我们在DCNN的中层引入了一层注意力权重,以放大或减弱活动以进一步实现目标。我们通过改变注意目标的摄影刺激来评估注意机制。我们发现,增加目标导向的注意力既有好处(提高命中率),也有代价(增加误报率)。在中等水平上,对于选择用于愚弄DCNNs的标准图像、混合图像和自然对抗图像的任务,注意仅在偏差适度增加的情况下提高灵敏度(即增加d ')。这些结果表明,目标导向注意力可以重新配置通用的DCNNs,以更好地适应当前的任务目标,就像PFC调节腹侧流的活动一样。除了更简洁和大脑一致之外,中级注意力方法比迁移学习的标准机器学习方法表现更好,即重新训练最终网络层以适应新任务。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
4.30
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信