{"title":"Generative Probabilistic Meta-Learning for Few-Shot Image Classification","authors":"Meijun Fu;Xiaomin Wang;Jun Wang;Zhang Yi","doi":"10.1109/TETCI.2024.3483255","DOIUrl":null,"url":null,"abstract":"Meta-learning, a rapidly advancing area in computational intelligence, leverages prior knowledge from related tasks to facilitate the swift adaptation to new tasks with limited data. A critical challenge in meta-learning is the quantification of model uncertainty. In this paper, we propose a novel meta-learning method, Generative Probabilistic Meta-Learning (GPML), designed for few-shot image classification. GPML extends the Probably Approximately Correct-Bayes (PAC-Bayes) framework, initially formulated for single-task scenarios, to meta-learning across multiple tasks. This extension not only provides theoretical generalization guarantees for meta-learning but also effectively captures model uncertainty through variational parameters. To enhance the expressiveness of approximated posteriors in Bayesian inference, GPML incorporates implicit modeling, which defines probability distributions over task-specific parameters in a data-driven manner. This is achieved by designing a generative model structure that integrates task-dependent prior knowledge into the model inference process. We conduct extensive multidimensional performance evaluations on few-shot image classification tasks across various benchmarks, demonstrating that GPML outperforms existing state-of-the-art meta-learning methods. Additionally, ablation studies focusing on model components, the PAC-Bayes framework, and implicit modeling validate the performance improvements attributed to the proposed generative model structure, learning framework, and modeling approach.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 2","pages":"1947-1960"},"PeriodicalIF":5.3000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10740484/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Meta-learning, a rapidly advancing area in computational intelligence, leverages prior knowledge from related tasks to facilitate the swift adaptation to new tasks with limited data. A critical challenge in meta-learning is the quantification of model uncertainty. In this paper, we propose a novel meta-learning method, Generative Probabilistic Meta-Learning (GPML), designed for few-shot image classification. GPML extends the Probably Approximately Correct-Bayes (PAC-Bayes) framework, initially formulated for single-task scenarios, to meta-learning across multiple tasks. This extension not only provides theoretical generalization guarantees for meta-learning but also effectively captures model uncertainty through variational parameters. To enhance the expressiveness of approximated posteriors in Bayesian inference, GPML incorporates implicit modeling, which defines probability distributions over task-specific parameters in a data-driven manner. This is achieved by designing a generative model structure that integrates task-dependent prior knowledge into the model inference process. We conduct extensive multidimensional performance evaluations on few-shot image classification tasks across various benchmarks, demonstrating that GPML outperforms existing state-of-the-art meta-learning methods. Additionally, ablation studies focusing on model components, the PAC-Bayes framework, and implicit modeling validate the performance improvements attributed to the proposed generative model structure, learning framework, and modeling approach.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.