Harmonizing the object recognition strategies of deep neural networks with humans.

Thomas Fel, Ivan Felipe, Drew Linsley, Thomas Serre
{"title":"Harmonizing the object recognition strategies of deep neural networks with humans.","authors":"Thomas Fel, Ivan Felipe, Drew Linsley, Thomas Serre","doi":"","DOIUrl":null,"url":null,"abstract":"<p><p>The many successes of deep neural networks (DNNs) over the past decade have largely been driven by computational scale rather than insights from biological intelligence. Here, we explore if these trends have also carried concomitant improvements in explaining the visual strategies humans rely on for object recognition. We do this by comparing two related but distinct properties of visual strategies in humans and DNNs: <i>where</i> they believe important visual features are in images and <i>how</i> they use those features to categorize objects. Across 84 different DNNs trained on ImageNet and three independent datasets measuring the <i>where</i> and the <i>how</i> of human visual strategies for object recognition on those images, we find a systematic trade-off between DNN categorization accuracy and alignment with human visual strategies for object recognition. <i>State-of-the-art DNNs are progressively becoming less aligned with humans as their accuracy improves</i>. We rectify this growing issue with our neural harmonizer: a general-purpose training routine that both aligns DNN and human visual strategies and improves categorization accuracy. Our work represents the first demonstration that the scaling laws [1-3] that are guiding the design of DNNs today have also produced worse models of human vision. We release our code and data at https://serre-lab.github.io/Harmonization to help the field build more human-like DNNs.</p>","PeriodicalId":72099,"journal":{"name":"Advances in neural information processing systems","volume":"35 ","pages":"9432-9446"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10353762/pdf/nihms-1856760.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in neural information processing systems","FirstCategoryId":"1085","ListUrlMain":"","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The many successes of deep neural networks (DNNs) over the past decade have largely been driven by computational scale rather than insights from biological intelligence. Here, we explore if these trends have also carried concomitant improvements in explaining the visual strategies humans rely on for object recognition. We do this by comparing two related but distinct properties of visual strategies in humans and DNNs: where they believe important visual features are in images and how they use those features to categorize objects. Across 84 different DNNs trained on ImageNet and three independent datasets measuring the where and the how of human visual strategies for object recognition on those images, we find a systematic trade-off between DNN categorization accuracy and alignment with human visual strategies for object recognition. State-of-the-art DNNs are progressively becoming less aligned with humans as their accuracy improves. We rectify this growing issue with our neural harmonizer: a general-purpose training routine that both aligns DNN and human visual strategies and improves categorization accuracy. Our work represents the first demonstration that the scaling laws [1-3] that are guiding the design of DNNs today have also produced worse models of human vision. We release our code and data at https://serre-lab.github.io/Harmonization to help the field build more human-like DNNs.

协调深度神经网络与人类的物体识别策略。
过去十年来,深度神经网络(DNN)取得了许多成功,其主要驱动力是计算规模,而不是生物智能的洞察力。在这里,我们要探讨的是,这些趋势在解释人类识别物体所依赖的视觉策略方面是否也带来了相应的改进。为此,我们比较了人类和 DNNs 视觉策略的两个相关但不同的特性:它们认为图像中重要的视觉特征在哪里,以及它们如何利用这些特征对物体进行分类。通过在 ImageNet 上训练的 84 种不同 DNN,以及测量人类视觉策略在这些图像中识别物体的位置和方式的三个独立数据集,我们发现在 DNN 的分类准确性和与人类识别物体的视觉策略的一致性之间存在系统性的权衡。随着精确度的提高,最先进的 DNN 与人类的一致性逐渐降低。我们的神经协调器纠正了这一日益严重的问题:它是一种通用训练程序,既能使 DNN 与人类视觉策略保持一致,又能提高分类准确性。我们的工作首次证明,如今指导 DNN 设计的缩放定律 [1-3] 也产生了更糟糕的人类视觉模型。我们在 https://serre-lab.github.io/Harmonization 上发布了我们的代码和数据,以帮助该领域构建更像人类的 DNN。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信