Enhancing task incremental continual learning: integrating prompt-based feature selection with pre-trained vision-language model

IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Lie Yang , Haohan Yang , Xiangkun He , Wenhui Huang , Chen Lv
{"title":"Enhancing task incremental continual learning: integrating prompt-based feature selection with pre-trained vision-language model","authors":"Lie Yang ,&nbsp;Haohan Yang ,&nbsp;Xiangkun He ,&nbsp;Wenhui Huang ,&nbsp;Chen Lv","doi":"10.1016/j.knosys.2025.113704","DOIUrl":null,"url":null,"abstract":"<div><div>Task incremental continual learning is pivotal for the evolution of general artificial intelligence, enabling models to progressively acquire and integrate new knowledge. However, enhancing model plasticity while ensuring stability remains one of the most significant challenges in this field. In this study, we propose a task incremental continual learning method based on a vision-language model (TICL-VLM), which exhibits both high plasticity and good stability. First, the image encoder from a pre-trained vision-language model (VLM) is adopted for robust feature extraction, and a novel task-prompt-based feature selection module is designed to enhance the plasticity of the proposed model. Additionally, a class description constraint is introduced to further improve the performance of the method. To ensure excellent stability, we freeze the parameters of the VLM's image and text encoders and introduce distinct feature selection and classification modules for each incremental task. Furthermore, a specific dataset (LFDDE) is constructed to comprehensively evaluate the performance of task incremental continual learning algorithms. Extensive experiments have been conducted on both the LFDDE and the well-known CIFAR-100 datasets. The experimental results clearly demonstrate significant improvements in maintaining stability while efficiently incorporating new knowledge with our method.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"320 ","pages":"Article 113704"},"PeriodicalIF":7.2000,"publicationDate":"2025-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125007506","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Task incremental continual learning is pivotal for the evolution of general artificial intelligence, enabling models to progressively acquire and integrate new knowledge. However, enhancing model plasticity while ensuring stability remains one of the most significant challenges in this field. In this study, we propose a task incremental continual learning method based on a vision-language model (TICL-VLM), which exhibits both high plasticity and good stability. First, the image encoder from a pre-trained vision-language model (VLM) is adopted for robust feature extraction, and a novel task-prompt-based feature selection module is designed to enhance the plasticity of the proposed model. Additionally, a class description constraint is introduced to further improve the performance of the method. To ensure excellent stability, we freeze the parameters of the VLM's image and text encoders and introduce distinct feature selection and classification modules for each incremental task. Furthermore, a specific dataset (LFDDE) is constructed to comprehensively evaluate the performance of task incremental continual learning algorithms. Extensive experiments have been conducted on both the LFDDE and the well-known CIFAR-100 datasets. The experimental results clearly demonstrate significant improvements in maintaining stability while efficiently incorporating new knowledge with our method.
增强任务增量持续学习:将基于提示的特征选择与预训练的视觉语言模型相结合
任务增量式持续学习对于通用人工智能的发展至关重要,它使模型能够逐步获取和整合新知识。然而,在保证模型稳定性的同时提高模型的可塑性仍然是该领域最重大的挑战之一。在本研究中,我们提出了一种基于视觉语言模型(TICL-VLM)的任务增量持续学习方法,该方法具有较高的可塑性和良好的稳定性。首先,采用预训练视觉语言模型(VLM)的图像编码器进行鲁棒特征提取,并设计了基于任务提示的特征选择模块,增强了模型的可塑性;此外,还引入了类描述约束,进一步提高了方法的性能。为了确保出色的稳定性,我们冻结了VLM图像和文本编码器的参数,并为每个增量任务引入了不同的特征选择和分类模块。此外,还构建了一个特定的数据集(LFDDE)来综合评价任务增量式持续学习算法的性能。在LFDDE和著名的CIFAR-100数据集上进行了大量的实验。实验结果清楚地表明,在保持稳定性的同时,我们的方法有效地结合了新知识。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Knowledge-Based Systems
Knowledge-Based Systems 工程技术-计算机:人工智能
CiteScore
14.80
自引率
12.50%
发文量
1245
审稿时长
7.8 months
期刊介绍: Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信