Akihiro Goda, Yukio Miyasaka, A. M. Gharehbaghi, M. Fujita
{"title":"考虑通信约束的并行算法的综合与推广","authors":"Akihiro Goda, Yukio Miyasaka, A. M. Gharehbaghi, M. Fujita","doi":"10.1109/ISQED48828.2020.9137022","DOIUrl":null,"url":null,"abstract":"Recently, the opportunities of parallel computing are expanding rapidly in various applications including neural networks and machine learning. It is, however, not at all straightforward to develop an efficient algorithm for each parallel computing environment since communications always introduce overhead in computation. In this paper, we propose a design method of optimum parallel computing under user-specified communication constraints. The basic strategy is to automatically generate optimum scheduling from small instances of the target problem and then they are semi-automatically generalized to much larger problems. Several experiments targeting matrix vector multiplication and convolutional neural networks have been conducted. Their results show the correctness and usefulness of the proposed method as well as its scalability.","PeriodicalId":225828,"journal":{"name":"2020 21st International Symposium on Quality Electronic Design (ISQED)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Synthesis and Generalization of Parallel Algorithms Considering Communication Constraints\",\"authors\":\"Akihiro Goda, Yukio Miyasaka, A. M. Gharehbaghi, M. Fujita\",\"doi\":\"10.1109/ISQED48828.2020.9137022\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recently, the opportunities of parallel computing are expanding rapidly in various applications including neural networks and machine learning. It is, however, not at all straightforward to develop an efficient algorithm for each parallel computing environment since communications always introduce overhead in computation. In this paper, we propose a design method of optimum parallel computing under user-specified communication constraints. The basic strategy is to automatically generate optimum scheduling from small instances of the target problem and then they are semi-automatically generalized to much larger problems. Several experiments targeting matrix vector multiplication and convolutional neural networks have been conducted. Their results show the correctness and usefulness of the proposed method as well as its scalability.\",\"PeriodicalId\":225828,\"journal\":{\"name\":\"2020 21st International Symposium on Quality Electronic Design (ISQED)\",\"volume\":\"33 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 21st International Symposium on Quality Electronic Design (ISQED)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISQED48828.2020.9137022\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 21st International Symposium on Quality Electronic Design (ISQED)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISQED48828.2020.9137022","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Synthesis and Generalization of Parallel Algorithms Considering Communication Constraints
Recently, the opportunities of parallel computing are expanding rapidly in various applications including neural networks and machine learning. It is, however, not at all straightforward to develop an efficient algorithm for each parallel computing environment since communications always introduce overhead in computation. In this paper, we propose a design method of optimum parallel computing under user-specified communication constraints. The basic strategy is to automatically generate optimum scheduling from small instances of the target problem and then they are semi-automatically generalized to much larger problems. Several experiments targeting matrix vector multiplication and convolutional neural networks have been conducted. Their results show the correctness and usefulness of the proposed method as well as its scalability.