{"title":"通过Adversarial*的无数据知识蒸馏","authors":"Yu Jin, Zhaofan Qiu, GengQuan Xie, Jinye Cai, Chao Li, Liwen Shen","doi":"10.1109/ICCCS52626.2021.9449145","DOIUrl":null,"url":null,"abstract":"Network Compression is a challenging task, but it is crucial for using the deeper network in the low-performance device. If the original training datasets could be obtained, the traditional network compression approaches are useful for training a compact deep model. This paper proposes a novel framework for knowledge distillation without original training datasets via Generative Adversarial Network(GAN). We arrange the fixed pre-trained deeper network and the compact network as the discriminators to generate the training dataset. We also use the deeper network and the compact network as the generators, then introduce one simple full connection network as the discriminator to compress the complex network. We propose (i) a series of new images generation loss functions. (ii) a knowledge distillation method via generating adversarial networks. Finally, we show the superiority of our approach by contrasting with SOTA by benchmark datasets.","PeriodicalId":376290,"journal":{"name":"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Data-free Knowledge Distillation via Adversarial*\",\"authors\":\"Yu Jin, Zhaofan Qiu, GengQuan Xie, Jinye Cai, Chao Li, Liwen Shen\",\"doi\":\"10.1109/ICCCS52626.2021.9449145\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Network Compression is a challenging task, but it is crucial for using the deeper network in the low-performance device. If the original training datasets could be obtained, the traditional network compression approaches are useful for training a compact deep model. This paper proposes a novel framework for knowledge distillation without original training datasets via Generative Adversarial Network(GAN). We arrange the fixed pre-trained deeper network and the compact network as the discriminators to generate the training dataset. We also use the deeper network and the compact network as the generators, then introduce one simple full connection network as the discriminator to compress the complex network. We propose (i) a series of new images generation loss functions. (ii) a knowledge distillation method via generating adversarial networks. Finally, we show the superiority of our approach by contrasting with SOTA by benchmark datasets.\",\"PeriodicalId\":376290,\"journal\":{\"name\":\"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)\",\"volume\":\"3 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-04-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCCS52626.2021.9449145\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCS52626.2021.9449145","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Network Compression is a challenging task, but it is crucial for using the deeper network in the low-performance device. If the original training datasets could be obtained, the traditional network compression approaches are useful for training a compact deep model. This paper proposes a novel framework for knowledge distillation without original training datasets via Generative Adversarial Network(GAN). We arrange the fixed pre-trained deeper network and the compact network as the discriminators to generate the training dataset. We also use the deeper network and the compact network as the generators, then introduce one simple full connection network as the discriminator to compress the complex network. We propose (i) a series of new images generation loss functions. (ii) a knowledge distillation method via generating adversarial networks. Finally, we show the superiority of our approach by contrasting with SOTA by benchmark datasets.