Xiaobo Guo, Pei Wang, Neng Gao, Xin Wang, Wenying Feng
{"title":"面向高维知识图嵌入的单个高级教师知识蒸馏模型","authors":"Xiaobo Guo, Pei Wang, Neng Gao, Xin Wang, Wenying Feng","doi":"10.1109/icicse55337.2022.9828905","DOIUrl":null,"url":null,"abstract":"An important role of Knowledge Graph Embedding (KGE) is to automatically complete the missing fact in a knowledge base. It is well-known that human society is constantly developing and the knowledge generated by human society will always being increasing. The increasing scale of the knowledge base is a great challenge to the storage and computing resources of downstream applications. At present, the dimensions of most mainstream knowledge graph embedding models are between 200-1000. For a large-scale knowledge base with millions of entities, these embeddings with hundreds of dimensional values are not conducive to rapid and frequent deployment on many kinds of artificial intelligent applications with limited storage and computing resources. To solve this problem, we propose a single-senior-teacher knowledge distillation model for high-dimensional knowledge graph embeddings named STDE, which constructs a low-dimensional student from a trained high-dimensional teacher. In STDE, the senior teacher can help the student learn key knowledge from correct knowledge and indistinguishable wrong knowledge with use of high-quality negative samples of triplets. We apply STDE to four typical KGE models on two famous data sets. Experimental results show that STDE can compress the embedding parameters of high-dimensional KGE models to 1/8 or 1/16 of their original scales. We further verify the effectiveness of \"senior teacher\" through ablation experiments.","PeriodicalId":177985,"journal":{"name":"2022 IEEE 2nd International Conference on Information Communication and Software Engineering (ICICSE)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"STDE: A Single-Senior-Teacher Knowledge Distillation Model for High-Dimensional Knowledge Graph Embeddings\",\"authors\":\"Xiaobo Guo, Pei Wang, Neng Gao, Xin Wang, Wenying Feng\",\"doi\":\"10.1109/icicse55337.2022.9828905\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An important role of Knowledge Graph Embedding (KGE) is to automatically complete the missing fact in a knowledge base. It is well-known that human society is constantly developing and the knowledge generated by human society will always being increasing. The increasing scale of the knowledge base is a great challenge to the storage and computing resources of downstream applications. At present, the dimensions of most mainstream knowledge graph embedding models are between 200-1000. For a large-scale knowledge base with millions of entities, these embeddings with hundreds of dimensional values are not conducive to rapid and frequent deployment on many kinds of artificial intelligent applications with limited storage and computing resources. To solve this problem, we propose a single-senior-teacher knowledge distillation model for high-dimensional knowledge graph embeddings named STDE, which constructs a low-dimensional student from a trained high-dimensional teacher. In STDE, the senior teacher can help the student learn key knowledge from correct knowledge and indistinguishable wrong knowledge with use of high-quality negative samples of triplets. We apply STDE to four typical KGE models on two famous data sets. Experimental results show that STDE can compress the embedding parameters of high-dimensional KGE models to 1/8 or 1/16 of their original scales. We further verify the effectiveness of \\\"senior teacher\\\" through ablation experiments.\",\"PeriodicalId\":177985,\"journal\":{\"name\":\"2022 IEEE 2nd International Conference on Information Communication and Software Engineering (ICICSE)\",\"volume\":\"70 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-03-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 2nd International Conference on Information Communication and Software Engineering (ICICSE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/icicse55337.2022.9828905\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 2nd International Conference on Information Communication and Software Engineering (ICICSE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icicse55337.2022.9828905","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
STDE: A Single-Senior-Teacher Knowledge Distillation Model for High-Dimensional Knowledge Graph Embeddings
An important role of Knowledge Graph Embedding (KGE) is to automatically complete the missing fact in a knowledge base. It is well-known that human society is constantly developing and the knowledge generated by human society will always being increasing. The increasing scale of the knowledge base is a great challenge to the storage and computing resources of downstream applications. At present, the dimensions of most mainstream knowledge graph embedding models are between 200-1000. For a large-scale knowledge base with millions of entities, these embeddings with hundreds of dimensional values are not conducive to rapid and frequent deployment on many kinds of artificial intelligent applications with limited storage and computing resources. To solve this problem, we propose a single-senior-teacher knowledge distillation model for high-dimensional knowledge graph embeddings named STDE, which constructs a low-dimensional student from a trained high-dimensional teacher. In STDE, the senior teacher can help the student learn key knowledge from correct knowledge and indistinguishable wrong knowledge with use of high-quality negative samples of triplets. We apply STDE to four typical KGE models on two famous data sets. Experimental results show that STDE can compress the embedding parameters of high-dimensional KGE models to 1/8 or 1/16 of their original scales. We further verify the effectiveness of "senior teacher" through ablation experiments.