Kiseong Hong , Hyundong Jin , Sungho Suh , Eunwoo Kim
{"title":"不断学习中的探索和利用","authors":"Kiseong Hong , Hyundong Jin , Sungho Suh , Eunwoo Kim","doi":"10.1016/j.neunet.2025.107444","DOIUrl":null,"url":null,"abstract":"<div><div>Continual learning (CL) has received a surge of interest, particularly in parameter isolation approaches, aiming to prevent catastrophic forgetting by assigning a disjoint parameter set to each task. Despite their effectiveness, existing approaches often neglect the task-specific differences, depending on predetermined allocation ratios of parameters. This can lead to suboptimal performance as it disregards the unique requirements of individual task traits. In this paper, we propose a novel <em>Exploration–Exploitation</em> approach to address this issue. Our goal is to adaptively distribute resources between acquiring new information (Exploration) and retaining previously learned knowledge (Exploitation) as new tasks emerge. This allows a continual learner to adaptively allocate parameters for every consecutive task by enabling them to compete for resources using exploration and exploitation. To achieve this, we introduce an allocation learner that learns the intricate interplay between exploration and exploitation across all layers of the continual learner. We demonstrate the proposed method under popular image classification benchmarks for diverse CL scenarios, including domain-shift task-incremental learning. Experimental results show that the proposed method outperforms other competitive continual learning approaches with an average margin of 5.3% across all scenarios.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"188 ","pages":"Article 107444"},"PeriodicalIF":6.0000,"publicationDate":"2025-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Exploration and exploitation in continual learning\",\"authors\":\"Kiseong Hong , Hyundong Jin , Sungho Suh , Eunwoo Kim\",\"doi\":\"10.1016/j.neunet.2025.107444\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Continual learning (CL) has received a surge of interest, particularly in parameter isolation approaches, aiming to prevent catastrophic forgetting by assigning a disjoint parameter set to each task. Despite their effectiveness, existing approaches often neglect the task-specific differences, depending on predetermined allocation ratios of parameters. This can lead to suboptimal performance as it disregards the unique requirements of individual task traits. In this paper, we propose a novel <em>Exploration–Exploitation</em> approach to address this issue. Our goal is to adaptively distribute resources between acquiring new information (Exploration) and retaining previously learned knowledge (Exploitation) as new tasks emerge. This allows a continual learner to adaptively allocate parameters for every consecutive task by enabling them to compete for resources using exploration and exploitation. To achieve this, we introduce an allocation learner that learns the intricate interplay between exploration and exploitation across all layers of the continual learner. We demonstrate the proposed method under popular image classification benchmarks for diverse CL scenarios, including domain-shift task-incremental learning. Experimental results show that the proposed method outperforms other competitive continual learning approaches with an average margin of 5.3% across all scenarios.</div></div>\",\"PeriodicalId\":49763,\"journal\":{\"name\":\"Neural Networks\",\"volume\":\"188 \",\"pages\":\"Article 107444\"},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2025-04-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0893608025003235\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025003235","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Exploration and exploitation in continual learning
Continual learning (CL) has received a surge of interest, particularly in parameter isolation approaches, aiming to prevent catastrophic forgetting by assigning a disjoint parameter set to each task. Despite their effectiveness, existing approaches often neglect the task-specific differences, depending on predetermined allocation ratios of parameters. This can lead to suboptimal performance as it disregards the unique requirements of individual task traits. In this paper, we propose a novel Exploration–Exploitation approach to address this issue. Our goal is to adaptively distribute resources between acquiring new information (Exploration) and retaining previously learned knowledge (Exploitation) as new tasks emerge. This allows a continual learner to adaptively allocate parameters for every consecutive task by enabling them to compete for resources using exploration and exploitation. To achieve this, we introduce an allocation learner that learns the intricate interplay between exploration and exploitation across all layers of the continual learner. We demonstrate the proposed method under popular image classification benchmarks for diverse CL scenarios, including domain-shift task-incremental learning. Experimental results show that the proposed method outperforms other competitive continual learning approaches with an average margin of 5.3% across all scenarios.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.