{"title":"利用类关系的二维规范化知识蒸馏","authors":"Benhong Zhang, Yiren Song, Yidong Zhang, Xiang Bi","doi":"10.1016/j.jvcir.2025.104557","DOIUrl":null,"url":null,"abstract":"<div><div>Knowledge distillation (KD) as one of the important methods of model compression, has been widely used in tasks such as image classification and detection. Existing KD methods are mainly carried out at the instance level and often ignore the role of inter-class relational information. Additionally, when there is a significant gap between the student’s capacity and the teacher’s capacity, the two model cannot be matched precisely. To address these issues, this paper proposes a two-dimensional normalized knowledge distillation method, which integrates inter-class and intra-class correlations and rectifies logits in two dimensions. Through our approach, the student model is able to acquire contextual information between samples with the help of intra-class correlation and mitigate the effect of logits magnitude on the prediction results through normalized rectification. We conduct numerous experiments and results show that our method achieves higher accuracy and better training efficiency compared to traditional KD methods.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"112 ","pages":"Article 104557"},"PeriodicalIF":3.1000,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Two-dimensional normalized knowledge distillation leveraging class relations\",\"authors\":\"Benhong Zhang, Yiren Song, Yidong Zhang, Xiang Bi\",\"doi\":\"10.1016/j.jvcir.2025.104557\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Knowledge distillation (KD) as one of the important methods of model compression, has been widely used in tasks such as image classification and detection. Existing KD methods are mainly carried out at the instance level and often ignore the role of inter-class relational information. Additionally, when there is a significant gap between the student’s capacity and the teacher’s capacity, the two model cannot be matched precisely. To address these issues, this paper proposes a two-dimensional normalized knowledge distillation method, which integrates inter-class and intra-class correlations and rectifies logits in two dimensions. Through our approach, the student model is able to acquire contextual information between samples with the help of intra-class correlation and mitigate the effect of logits magnitude on the prediction results through normalized rectification. We conduct numerous experiments and results show that our method achieves higher accuracy and better training efficiency compared to traditional KD methods.</div></div>\",\"PeriodicalId\":54755,\"journal\":{\"name\":\"Journal of Visual Communication and Image Representation\",\"volume\":\"112 \",\"pages\":\"Article 104557\"},\"PeriodicalIF\":3.1000,\"publicationDate\":\"2025-08-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Visual Communication and Image Representation\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1047320325001713\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Communication and Image Representation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1047320325001713","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Two-dimensional normalized knowledge distillation leveraging class relations
Knowledge distillation (KD) as one of the important methods of model compression, has been widely used in tasks such as image classification and detection. Existing KD methods are mainly carried out at the instance level and often ignore the role of inter-class relational information. Additionally, when there is a significant gap between the student’s capacity and the teacher’s capacity, the two model cannot be matched precisely. To address these issues, this paper proposes a two-dimensional normalized knowledge distillation method, which integrates inter-class and intra-class correlations and rectifies logits in two dimensions. Through our approach, the student model is able to acquire contextual information between samples with the help of intra-class correlation and mitigate the effect of logits magnitude on the prediction results through normalized rectification. We conduct numerous experiments and results show that our method achieves higher accuracy and better training efficiency compared to traditional KD methods.
期刊介绍:
The Journal of Visual Communication and Image Representation publishes papers on state-of-the-art visual communication and image representation, with emphasis on novel technologies and theoretical work in this multidisciplinary area of pure and applied research. The field of visual communication and image representation is considered in its broadest sense and covers both digital and analog aspects as well as processing and communication in biological visual systems.