{"title":"BEFM: A balanced and efficient fine-tuning model in class-incremental learning","authors":"Lize Liu, Jian Ji, Lei Zhao","doi":"10.1016/j.knosys.2025.113298","DOIUrl":null,"url":null,"abstract":"<div><div>The ultimate objective of class-incremental learning (CIL) is to solve the stability–plasticity dilemma by continuously learning new courses without losing what has been learnt from the previous ones. Typical CIL methods tend to favor sample replay and parameter regularization, but recent studies have shown that exploiting and adapting historical models can effectively improve model performance. We present A Balanced and Efficient Fine-tuning Model (BEFM) to improve memory consumption efficiency and prevent catastrophic forgetting in CIL by utilizing historical models. In order to fully utilize various normalization techniques and preserve the model’s high plasticity and stability throughout training, we first create a multi-normalization module to replace the original single batch normalization procedure. On the other hand, we construct a model expansion method with knowledge distillation and offer a logit fine-tuning strategy to increase memory and computational efficiency. The model expansion strategy effectively solves the memory problem caused by feature aggregation by expanding only the deeper models with greater influence when learning a new task, while the knowledge distillation strategy encourages the model to retain the memory of the old task and improves the stability of the model. As a way to address the issue of class imbalance, the logit fine-tuning technique optimizes the standard softmax cross entropy, improving classifier design without adding computational burden. We test our method on the CIFAR10, CIFAR100, and miniImageNet100 datasets under various conditions. According to experimental results, our strategy outperforms current approaches with notable gains in performance. The code is available at <span><span>https://github.com/Lize-Liu/BEFM-main</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"315 ","pages":"Article 113298"},"PeriodicalIF":7.2000,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125003454","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The ultimate objective of class-incremental learning (CIL) is to solve the stability–plasticity dilemma by continuously learning new courses without losing what has been learnt from the previous ones. Typical CIL methods tend to favor sample replay and parameter regularization, but recent studies have shown that exploiting and adapting historical models can effectively improve model performance. We present A Balanced and Efficient Fine-tuning Model (BEFM) to improve memory consumption efficiency and prevent catastrophic forgetting in CIL by utilizing historical models. In order to fully utilize various normalization techniques and preserve the model’s high plasticity and stability throughout training, we first create a multi-normalization module to replace the original single batch normalization procedure. On the other hand, we construct a model expansion method with knowledge distillation and offer a logit fine-tuning strategy to increase memory and computational efficiency. The model expansion strategy effectively solves the memory problem caused by feature aggregation by expanding only the deeper models with greater influence when learning a new task, while the knowledge distillation strategy encourages the model to retain the memory of the old task and improves the stability of the model. As a way to address the issue of class imbalance, the logit fine-tuning technique optimizes the standard softmax cross entropy, improving classifier design without adding computational burden. We test our method on the CIFAR10, CIFAR100, and miniImageNet100 datasets under various conditions. According to experimental results, our strategy outperforms current approaches with notable gains in performance. The code is available at https://github.com/Lize-Liu/BEFM-main.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.