Dan Luo , Kangfeng Zheng , Chunhua Wu , Xiujuan Wang
{"title":"MSLoRA: Meta-learned scaling for adaptive fine-tuning of LoRA","authors":"Dan Luo , Kangfeng Zheng , Chunhua Wu , Xiujuan Wang","doi":"10.1016/j.neucom.2025.130374","DOIUrl":null,"url":null,"abstract":"<div><div>Low rank adaptation (LoRA) methods have demonstrated strong capabilities in efficiently fine-tuning large models. However, existing LoRA-based approaches typically require manually setting the scaling factor, a process that involves extensive search efforts to find optimal values. To address this challenge, we first develop data-driven heuristic methods that automatically determine layer-wise scaling factors through either activation pattern analysis during forward propagation or gradient behavior monitoring during backward updates. However,their practical performance remains unsatisfactory in applications. Building upon these theoretical foundations, we present MSLoRA, a novel framework that reformulates scaling factor determination as a dynamic optimization problem in parameter-efficient fine-tuning. Our approach innovatively models scaling factors as self-adaptive meta-parameters whose optimal values emerge organically through the interplay between transformer architecture hierarchies and task-specific learning objectives. Extensive experiments conducted across both natural language understanding and generative tasks reveal that MSLoRA consistently outperforms baseline models. This highlights the effectiveness of MSLoRA’s dynamic, layer-specific adjustment mechanism in capturing the complex nature of task-specific activation patterns, making it a more robust and scalable solution for parameter-efficient fine-tuning of large models.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"643 ","pages":"Article 130374"},"PeriodicalIF":5.5000,"publicationDate":"2025-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S092523122501046X","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Low rank adaptation (LoRA) methods have demonstrated strong capabilities in efficiently fine-tuning large models. However, existing LoRA-based approaches typically require manually setting the scaling factor, a process that involves extensive search efforts to find optimal values. To address this challenge, we first develop data-driven heuristic methods that automatically determine layer-wise scaling factors through either activation pattern analysis during forward propagation or gradient behavior monitoring during backward updates. However,their practical performance remains unsatisfactory in applications. Building upon these theoretical foundations, we present MSLoRA, a novel framework that reformulates scaling factor determination as a dynamic optimization problem in parameter-efficient fine-tuning. Our approach innovatively models scaling factors as self-adaptive meta-parameters whose optimal values emerge organically through the interplay between transformer architecture hierarchies and task-specific learning objectives. Extensive experiments conducted across both natural language understanding and generative tasks reveal that MSLoRA consistently outperforms baseline models. This highlights the effectiveness of MSLoRA’s dynamic, layer-specific adjustment mechanism in capturing the complex nature of task-specific activation patterns, making it a more robust and scalable solution for parameter-efficient fine-tuning of large models.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.