Classification of Astronomical Bodies by Efficient Layer Fine-Tuning of Deep Neural Networks

Sabeesh Ethiraj, B. K. Bolla
{"title":"Classification of Astronomical Bodies by Efficient Layer Fine-Tuning of Deep Neural Networks","authors":"Sabeesh Ethiraj, B. K. Bolla","doi":"10.1109/CICT53865.2020.9672430","DOIUrl":null,"url":null,"abstract":"The SDSS-IV dataset contains information about various astronomical bodies such as Galaxies, Stars, and Quasars captured by observatories. Inspired by our work on deep multimodal learning, which utilized transfer learning to classify the SDSS-IV dataset, we further extended our research in the fine tuning of these architectures to study the effect in the classification scenario. Architectures such as Resnet-50, DenseNet-121 VGG-16, Xception, EfficientNetB2, MobileNetV2 and NasnetMobile have been built using layer wise fine tuning at different levels. Our findings suggest that freezing all layers with Imagenet weights and adding a final trainable layer may not be the optimal solution. Further, baseline models and models that have higher number of trainable layers performed similarly in certain architectures. Model need to be fine tuned at different levels and a specific training ratio is required for a model to be termed ideal. Different architectures had different responses to the change in the number of trainable layers w.r.t accuracies. While models such as DenseNet-121, Xception, EfficientNetB2 achieved peak accuracies that were relatively consistent with near perfect training curves, models such as Resnet-50,VGG-16, MobileNetV2 and NasnetMobile had lower, delayed peak accuracies with poorly fitting training curves. It was also found that though mobile neural networks have lesser parameters and model size, they may not always be ideal for deployment on a low computational device as they had consistently lower validation accuracies. Customized evaluation metrics such as Tuning Parameter Ratio and Tuning Layer Ratio are used for model evaluation.","PeriodicalId":265498,"journal":{"name":"2021 5th Conference on Information and Communication Technology (CICT)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 5th Conference on Information and Communication Technology (CICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CICT53865.2020.9672430","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

The SDSS-IV dataset contains information about various astronomical bodies such as Galaxies, Stars, and Quasars captured by observatories. Inspired by our work on deep multimodal learning, which utilized transfer learning to classify the SDSS-IV dataset, we further extended our research in the fine tuning of these architectures to study the effect in the classification scenario. Architectures such as Resnet-50, DenseNet-121 VGG-16, Xception, EfficientNetB2, MobileNetV2 and NasnetMobile have been built using layer wise fine tuning at different levels. Our findings suggest that freezing all layers with Imagenet weights and adding a final trainable layer may not be the optimal solution. Further, baseline models and models that have higher number of trainable layers performed similarly in certain architectures. Model need to be fine tuned at different levels and a specific training ratio is required for a model to be termed ideal. Different architectures had different responses to the change in the number of trainable layers w.r.t accuracies. While models such as DenseNet-121, Xception, EfficientNetB2 achieved peak accuracies that were relatively consistent with near perfect training curves, models such as Resnet-50,VGG-16, MobileNetV2 and NasnetMobile had lower, delayed peak accuracies with poorly fitting training curves. It was also found that though mobile neural networks have lesser parameters and model size, they may not always be ideal for deployment on a low computational device as they had consistently lower validation accuracies. Customized evaluation metrics such as Tuning Parameter Ratio and Tuning Layer Ratio are used for model evaluation.
基于深度神经网络有效层微调的天体分类
SDSS-IV数据集包含了天文台捕获的各种天体的信息,如星系、恒星和类星体。受我们在深度多模态学习(利用迁移学习对SDSS-IV数据集进行分类)方面的工作的启发,我们进一步扩展了这些架构的微调研究,以研究分类场景中的效果。Resnet-50、DenseNet-121 vgg16、Xception、EfficientNetB2、MobileNetV2和NasnetMobile等架构已经在不同级别上使用分层微调构建。我们的研究结果表明,冻结所有具有Imagenet权重的层并添加最终可训练层可能不是最佳解决方案。此外,基线模型和具有更多可训练层的模型在某些架构中执行相似。模型需要在不同的水平上进行微调,并且需要特定的训练比率才能将模型称为理想模型。不同的体系结构对可训练层数量的变化有不同的响应。虽然DenseNet-121、Xception、EfficientNetB2等模型的峰值精度与接近完美的训练曲线相对一致,但Resnet-50、vgg16、MobileNetV2和NasnetMobile等模型的峰值精度较低,延迟,训练曲线拟合较差。研究还发现,尽管移动神经网络具有较小的参数和模型大小,但它们可能并不总是理想的部署在低计算设备上,因为它们始终具有较低的验证准确性。自定义的评估指标,如调谐参数比率和调谐层比率用于模型评估。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信