Wenjie Liu, Lei Zhang, Xianliang Zhang, Xinyang Zhou, Xin Wei
{"title":"Be your own doctor: Temperature scaling self-knowledge distillation for medical image classification","authors":"Wenjie Liu, Lei Zhang, Xianliang Zhang, Xinyang Zhou, Xin Wei","doi":"10.1016/j.neucom.2025.130115","DOIUrl":null,"url":null,"abstract":"<div><div>Self-knowledge distillation (self-KD), which uses the student network as the teacher model, allows the model to learn knowledge by itself. It has been widely studied in various medical image tasks for constructing lightweight models to alleviate the limitations of computing resources. However, existing self-KD methods use a single temperature for distillation, ignoring the effect of temperature on different classes. In this paper, we investigate the effects of target class temperature and non-target class temperature on the performance of self-KD. Based on the above study, a temperature scaling self-knowledge distillation (TSS-KD) model is proposed, which can better balance the target class knowledge and non-target class knowledge. By adjusting the temperature scaling of different classes, the model can learn better representations by distilling the well-proportioned features. To make the network focus more on the local lesions of medical images, a regional gamma augmentation (RGA) method is proposed, which provides stronger perturbations to the same sample to generate more differentiated features. By self-regularizing the consistency of these features, the model can learn more local knowledge. To evaluate the effectiveness of the proposed method, extensive experiments are conducted on nine medical image classification tasks of eight public datasets. Experimental results show that the proposed method outperforms state-of-the-art self-KD models and has strong generality. The code is available at <span><span>https://github.com/JeaneyLau/TSS-KD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"638 ","pages":"Article 130115"},"PeriodicalIF":5.5000,"publicationDate":"2025-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225007878","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Self-knowledge distillation (self-KD), which uses the student network as the teacher model, allows the model to learn knowledge by itself. It has been widely studied in various medical image tasks for constructing lightweight models to alleviate the limitations of computing resources. However, existing self-KD methods use a single temperature for distillation, ignoring the effect of temperature on different classes. In this paper, we investigate the effects of target class temperature and non-target class temperature on the performance of self-KD. Based on the above study, a temperature scaling self-knowledge distillation (TSS-KD) model is proposed, which can better balance the target class knowledge and non-target class knowledge. By adjusting the temperature scaling of different classes, the model can learn better representations by distilling the well-proportioned features. To make the network focus more on the local lesions of medical images, a regional gamma augmentation (RGA) method is proposed, which provides stronger perturbations to the same sample to generate more differentiated features. By self-regularizing the consistency of these features, the model can learn more local knowledge. To evaluate the effectiveness of the proposed method, extensive experiments are conducted on nine medical image classification tasks of eight public datasets. Experimental results show that the proposed method outperforms state-of-the-art self-KD models and has strong generality. The code is available at https://github.com/JeaneyLau/TSS-KD.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.