{"title":"Towards Generalizable Meta-Deep Reinforcement Learning Algorithm for Multiobjective Traveling Salesman Problems","authors":"Xiaoyu Fu;Shenshen Gu;Chee-Meng Chew;Tengfei Li","doi":"10.1109/TAI.2025.3614210","DOIUrl":null,"url":null,"abstract":"The multiobjective traveling salesman problem (MOTSP) is a representative class of multiobjective combinatorial optimization problems, with significant implications for both theoretical research and practical applications. Although deep reinforcement learning (DRL) has shown promise in solving MOTSPs, existing approaches often struggle with generalization to large-scale problem instances. To address this challenge, we propose a novel meta-deep reinforcement learning framework with preference-fused attention networks (MDRL-PFAN). This framework integrates a preference-fused mechanism to jointly encode problem instances and weight preferences into a unified feature space. Moreover, an ensemble meta-learning strategy is adopted to train the meta-model across tasks with varying scales, equipping MDRL-PFAN with robust solving and strong cross-scale generalization capabilities. During inference, a lightweight fine-tuning process on small-batch adaptation tasks is employed to further enhance optimization performance. Extensive experiments on diverse MOTSP instances demonstrate that MDRL-PFAN achieves superior performance compared to classic evolutionary algorithms and state-of-the-art DRL algorithms in terms of training efficiency, solution quality, and cross-scale generalization capability.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"7 4","pages":"2238-2252"},"PeriodicalIF":0.0000,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11181152/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/9/26 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The multiobjective traveling salesman problem (MOTSP) is a representative class of multiobjective combinatorial optimization problems, with significant implications for both theoretical research and practical applications. Although deep reinforcement learning (DRL) has shown promise in solving MOTSPs, existing approaches often struggle with generalization to large-scale problem instances. To address this challenge, we propose a novel meta-deep reinforcement learning framework with preference-fused attention networks (MDRL-PFAN). This framework integrates a preference-fused mechanism to jointly encode problem instances and weight preferences into a unified feature space. Moreover, an ensemble meta-learning strategy is adopted to train the meta-model across tasks with varying scales, equipping MDRL-PFAN with robust solving and strong cross-scale generalization capabilities. During inference, a lightweight fine-tuning process on small-batch adaptation tasks is employed to further enhance optimization performance. Extensive experiments on diverse MOTSP instances demonstrate that MDRL-PFAN achieves superior performance compared to classic evolutionary algorithms and state-of-the-art DRL algorithms in terms of training efficiency, solution quality, and cross-scale generalization capability.