{"title":"继发性肺结核分类集成的知识提炼","authors":"Qinghua Zhou, Hengde Zhu, Xin Zhang, Yudong Zhang","doi":"10.1145/3492323.3495570","DOIUrl":null,"url":null,"abstract":"This paper focuses on a teacher-student scheme for knowledge distillation of a secondary pulmonary tuberculosis classification ensemble. As ensemble learning combines multiple neural networks, the combined ensemble often requires inference from each base network. Therefore, one of the challenges for ensemble learning is its size and efficiency in inference. This paper proposes knowledge distillation for ensemble learning via a teacher-student scheme, where a single noised student learns the concatenated representations generated by each base network. Comparing the ensemble of teacher networks and the single student, we showed that, with a performance penalty, the ensemble size and computational cost are significantly reduced.","PeriodicalId":440884,"journal":{"name":"Proceedings of the 14th IEEE/ACM International Conference on Utility and Cloud Computing Companion","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Knowledge distillation for secondary pulmonary tuberculosis classification ensemble\",\"authors\":\"Qinghua Zhou, Hengde Zhu, Xin Zhang, Yudong Zhang\",\"doi\":\"10.1145/3492323.3495570\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper focuses on a teacher-student scheme for knowledge distillation of a secondary pulmonary tuberculosis classification ensemble. As ensemble learning combines multiple neural networks, the combined ensemble often requires inference from each base network. Therefore, one of the challenges for ensemble learning is its size and efficiency in inference. This paper proposes knowledge distillation for ensemble learning via a teacher-student scheme, where a single noised student learns the concatenated representations generated by each base network. Comparing the ensemble of teacher networks and the single student, we showed that, with a performance penalty, the ensemble size and computational cost are significantly reduced.\",\"PeriodicalId\":440884,\"journal\":{\"name\":\"Proceedings of the 14th IEEE/ACM International Conference on Utility and Cloud Computing Companion\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 14th IEEE/ACM International Conference on Utility and Cloud Computing Companion\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3492323.3495570\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 14th IEEE/ACM International Conference on Utility and Cloud Computing Companion","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3492323.3495570","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Knowledge distillation for secondary pulmonary tuberculosis classification ensemble
This paper focuses on a teacher-student scheme for knowledge distillation of a secondary pulmonary tuberculosis classification ensemble. As ensemble learning combines multiple neural networks, the combined ensemble often requires inference from each base network. Therefore, one of the challenges for ensemble learning is its size and efficiency in inference. This paper proposes knowledge distillation for ensemble learning via a teacher-student scheme, where a single noised student learns the concatenated representations generated by each base network. Comparing the ensemble of teacher networks and the single student, we showed that, with a performance penalty, the ensemble size and computational cost are significantly reduced.