{"title":"Reimagining Linear Probing: Kolmogorov-Arnold Networks in Transfer Learning","authors":"Sheng Shen, Rabih Younes","doi":"arxiv-2409.07763","DOIUrl":null,"url":null,"abstract":"This paper introduces Kolmogorov-Arnold Networks (KAN) as an enhancement to\nthe traditional linear probing method in transfer learning. Linear probing,\noften applied to the final layer of pre-trained models, is limited by its\ninability to model complex relationships in data. To address this, we propose\nsubstituting the linear probing layer with KAN, which leverages spline-based\nrepresentations to approximate intricate functions. In this study, we integrate\nKAN with a ResNet-50 model pre-trained on ImageNet and evaluate its performance\non the CIFAR-10 dataset. We perform a systematic hyperparameter search,\nfocusing on grid size and spline degree (k), to optimize KAN's flexibility and\naccuracy. Our results demonstrate that KAN consistently outperforms traditional\nlinear probing, achieving significant improvements in accuracy and\ngeneralization across a range of configurations. These findings indicate that\nKAN offers a more powerful and adaptable alternative to conventional linear\nprobing techniques in transfer learning.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07763","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper introduces Kolmogorov-Arnold Networks (KAN) as an enhancement to
the traditional linear probing method in transfer learning. Linear probing,
often applied to the final layer of pre-trained models, is limited by its
inability to model complex relationships in data. To address this, we propose
substituting the linear probing layer with KAN, which leverages spline-based
representations to approximate intricate functions. In this study, we integrate
KAN with a ResNet-50 model pre-trained on ImageNet and evaluate its performance
on the CIFAR-10 dataset. We perform a systematic hyperparameter search,
focusing on grid size and spline degree (k), to optimize KAN's flexibility and
accuracy. Our results demonstrate that KAN consistently outperforms traditional
linear probing, achieving significant improvements in accuracy and
generalization across a range of configurations. These findings indicate that
KAN offers a more powerful and adaptable alternative to conventional linear
probing techniques in transfer learning.