基于层激活优化的预训练神经网络行为分析

Melissa C Phillips, Rebecca Stein, Taeheon Park
{"title":"基于层激活优化的预训练神经网络行为分析","authors":"Melissa C Phillips, Rebecca Stein, Taeheon Park","doi":"10.1109/SIEDS49339.2020.9106628","DOIUrl":null,"url":null,"abstract":"Image classification and object recognition with neural networks could have applications in aesthetically-focused branches of the humanities, such as landscape architecture. However, such methods require either the assembly of a massive, domain specific labeled data set or use of network weights initialized on another data set, a technique known as transfer learning. Transfer learning research has established that a pre-trained convolutional neural network (CNN) can achieve high accuracy on new image recognition tasks with relatively few training images. In practice, pre-trained tends to mean pre-trained on ImageNet, the standard dataset for computer vision research. Experiments have shown that the dataset on which a pre-trained model was originally optimized can quantitatively bias it. The goal of this project was to design an experiment to qualitatively analyze how the dataset used to initialize a pre-trained classification system affects its behavior at progressive network layers using feature visualization strategies. We initialized two ResNet-18 CNNs with weights pre-trained on ImageNet and the Places365 dataset, respectively, and fine-tuned them for a new classification task on a landscape image dataset which we collected. Using class activation optimization methods taken from the deep visualization literature, we compared the network filters at several hidden layers and the final output layers. The class activation optimization results show that even at early stages in the networks, their neurons exhibit notably different behavior. Accordingly, we show both that feature visualization techniques can be used to qualitatively study the effect of original training data on transfer learning and, consequently, that the homogeneous use of ImageNet in computer vision experiments may have notable implications for model behavior.","PeriodicalId":331495,"journal":{"name":"2020 Systems and Information Engineering Design Symposium (SIEDS)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Analyzing Pre-Trained Neural Network Behavior with Layer Activation Optimization\",\"authors\":\"Melissa C Phillips, Rebecca Stein, Taeheon Park\",\"doi\":\"10.1109/SIEDS49339.2020.9106628\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Image classification and object recognition with neural networks could have applications in aesthetically-focused branches of the humanities, such as landscape architecture. However, such methods require either the assembly of a massive, domain specific labeled data set or use of network weights initialized on another data set, a technique known as transfer learning. Transfer learning research has established that a pre-trained convolutional neural network (CNN) can achieve high accuracy on new image recognition tasks with relatively few training images. In practice, pre-trained tends to mean pre-trained on ImageNet, the standard dataset for computer vision research. Experiments have shown that the dataset on which a pre-trained model was originally optimized can quantitatively bias it. The goal of this project was to design an experiment to qualitatively analyze how the dataset used to initialize a pre-trained classification system affects its behavior at progressive network layers using feature visualization strategies. We initialized two ResNet-18 CNNs with weights pre-trained on ImageNet and the Places365 dataset, respectively, and fine-tuned them for a new classification task on a landscape image dataset which we collected. Using class activation optimization methods taken from the deep visualization literature, we compared the network filters at several hidden layers and the final output layers. The class activation optimization results show that even at early stages in the networks, their neurons exhibit notably different behavior. Accordingly, we show both that feature visualization techniques can be used to qualitatively study the effect of original training data on transfer learning and, consequently, that the homogeneous use of ImageNet in computer vision experiments may have notable implications for model behavior.\",\"PeriodicalId\":331495,\"journal\":{\"name\":\"2020 Systems and Information Engineering Design Symposium (SIEDS)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 Systems and Information Engineering Design Symposium (SIEDS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SIEDS49339.2020.9106628\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 Systems and Information Engineering Design Symposium (SIEDS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SIEDS49339.2020.9106628","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

神经网络的图像分类和物体识别可以应用于人文学科中以美学为重点的分支,比如景观建筑学。然而,这种方法要么需要大量的、特定领域的标记数据集,要么需要使用在另一个数据集上初始化的网络权重,这种技术被称为迁移学习。迁移学习研究表明,预训练卷积神经网络(CNN)可以在训练图像相对较少的情况下,在新的图像识别任务上取得较高的准确率。在实践中,预训练往往意味着在ImageNet(计算机视觉研究的标准数据集)上进行预训练。实验表明,预训练模型最初优化的数据集可以定量地对其进行偏差。该项目的目标是设计一个实验来定性地分析用于初始化预训练分类系统的数据集如何使用特征可视化策略影响其在渐进式网络层的行为。我们分别在ImageNet和Places365数据集上初始化了两个权重预训练的ResNet-18 cnn,并对它们进行了微调,以便在我们收集的景观图像数据集上进行新的分类任务。使用来自深度可视化文献的类激活优化方法,我们比较了几个隐藏层和最终输出层的网络过滤器。类激活优化结果表明,即使在网络的早期阶段,它们的神经元也表现出明显不同的行为。因此,我们表明特征可视化技术可以用于定性地研究原始训练数据对迁移学习的影响,因此,在计算机视觉实验中均匀使用ImageNet可能对模型行为有显著的影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Analyzing Pre-Trained Neural Network Behavior with Layer Activation Optimization
Image classification and object recognition with neural networks could have applications in aesthetically-focused branches of the humanities, such as landscape architecture. However, such methods require either the assembly of a massive, domain specific labeled data set or use of network weights initialized on another data set, a technique known as transfer learning. Transfer learning research has established that a pre-trained convolutional neural network (CNN) can achieve high accuracy on new image recognition tasks with relatively few training images. In practice, pre-trained tends to mean pre-trained on ImageNet, the standard dataset for computer vision research. Experiments have shown that the dataset on which a pre-trained model was originally optimized can quantitatively bias it. The goal of this project was to design an experiment to qualitatively analyze how the dataset used to initialize a pre-trained classification system affects its behavior at progressive network layers using feature visualization strategies. We initialized two ResNet-18 CNNs with weights pre-trained on ImageNet and the Places365 dataset, respectively, and fine-tuned them for a new classification task on a landscape image dataset which we collected. Using class activation optimization methods taken from the deep visualization literature, we compared the network filters at several hidden layers and the final output layers. The class activation optimization results show that even at early stages in the networks, their neurons exhibit notably different behavior. Accordingly, we show both that feature visualization techniques can be used to qualitatively study the effect of original training data on transfer learning and, consequently, that the homogeneous use of ImageNet in computer vision experiments may have notable implications for model behavior.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信