Comparison of self-supervised in-domain and supervised out-domain transfer learning for bird species recognition

Houtan Ghaffari, Paul Devos
{"title":"Comparison of self-supervised in-domain and supervised out-domain transfer learning for bird species recognition","authors":"Houtan Ghaffari, Paul Devos","doi":"arxiv-2404.17252","DOIUrl":null,"url":null,"abstract":"Transferring the weights of a pre-trained model to assist another task has\nbecome a crucial part of modern deep learning, particularly in data-scarce\nscenarios. Pre-training refers to the initial step of training models outside\nthe current task of interest, typically on another dataset. It can be done via\nsupervised models using human-annotated datasets or self-supervised models\ntrained on unlabeled datasets. In both cases, many pre-trained models are\navailable to fine-tune for the task of interest. Interestingly, research has\nshown that pre-trained models from ImageNet can be helpful for audio tasks\ndespite being trained on image datasets. Hence, it's unclear whether in-domain\nmodels would be advantageous compared to competent out-domain models, such as\nconvolutional neural networks from ImageNet. Our experiments will demonstrate\nthe usefulness of in-domain models and datasets for bird species recognition by\nleveraging VICReg, a recent and powerful self-supervised method.","PeriodicalId":501178,"journal":{"name":"arXiv - CS - Sound","volume":"38 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Sound","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2404.17252","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Transferring the weights of a pre-trained model to assist another task has become a crucial part of modern deep learning, particularly in data-scarce scenarios. Pre-training refers to the initial step of training models outside the current task of interest, typically on another dataset. It can be done via supervised models using human-annotated datasets or self-supervised models trained on unlabeled datasets. In both cases, many pre-trained models are available to fine-tune for the task of interest. Interestingly, research has shown that pre-trained models from ImageNet can be helpful for audio tasks despite being trained on image datasets. Hence, it's unclear whether in-domain models would be advantageous compared to competent out-domain models, such as convolutional neural networks from ImageNet. Our experiments will demonstrate the usefulness of in-domain models and datasets for bird species recognition by leveraging VICReg, a recent and powerful self-supervised method.
比较自监督域内转移学习和监督域外转移学习在鸟类物种识别中的应用
将预先训练好的模型的权重转移到另一项任务上,已成为现代深度学习的重要组成部分,尤其是在数据稀缺的情况下。预训练指的是在当前任务之外训练模型的初始步骤,通常是在另一个数据集上。预训练可以通过使用人类标注数据集的监督模型或在无标注数据集上的自监督模型来完成。在这两种情况下,都有许多预先训练好的模型,可以针对感兴趣的任务进行微调。有趣的是,研究表明,来自 ImageNet 的预训练模型尽管是在图像数据集上训练的,但对音频任务也有帮助。因此,目前还不清楚域内模型与胜任的域外模型(如 ImageNet 的卷积神经网络)相比是否具有优势。我们的实验将通过利用 VICReg 这一最新的强大自监督方法,证明域内模型和数据集在鸟类物种识别方面的实用性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信