{"title":"用于核心训练集构建的数据集蒸馏","authors":"Yuna Jeong, Myunggwon Hwang, Won-Kyoung Sung","doi":"10.1145/3426020.3426051","DOIUrl":null,"url":null,"abstract":"Machine learning is a widely adopted solution to complex and non-linear problems, but it takes considerable labor and time to develop an optimal model with high reliability. The costs increase even more as the model deepens and training data grows. This paper presents a method in which, a technique known as dataset distillation, can be implemented in data selection to reduce the training time. We first train the model with distilled images, and then, predict original train data to measure training contribution as sampling weight of selection. Our method enables the fast and easy calculation of weights even in the case of redesigning a network.","PeriodicalId":305132,"journal":{"name":"The 9th International Conference on Smart Media and Applications","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Dataset Distillation for Core Training Set Construction\",\"authors\":\"Yuna Jeong, Myunggwon Hwang, Won-Kyoung Sung\",\"doi\":\"10.1145/3426020.3426051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine learning is a widely adopted solution to complex and non-linear problems, but it takes considerable labor and time to develop an optimal model with high reliability. The costs increase even more as the model deepens and training data grows. This paper presents a method in which, a technique known as dataset distillation, can be implemented in data selection to reduce the training time. We first train the model with distilled images, and then, predict original train data to measure training contribution as sampling weight of selection. Our method enables the fast and easy calculation of weights even in the case of redesigning a network.\",\"PeriodicalId\":305132,\"journal\":{\"name\":\"The 9th International Conference on Smart Media and Applications\",\"volume\":\"44 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The 9th International Conference on Smart Media and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3426020.3426051\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The 9th International Conference on Smart Media and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3426020.3426051","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Dataset Distillation for Core Training Set Construction
Machine learning is a widely adopted solution to complex and non-linear problems, but it takes considerable labor and time to develop an optimal model with high reliability. The costs increase even more as the model deepens and training data grows. This paper presents a method in which, a technique known as dataset distillation, can be implemented in data selection to reduce the training time. We first train the model with distilled images, and then, predict original train data to measure training contribution as sampling weight of selection. Our method enables the fast and easy calculation of weights even in the case of redesigning a network.