{"title":"Database caching for job-level computing","authors":"H. Chiang, Ting-Han Wei, I-Chen Wu","doi":"10.1109/TAAI.2016.7880170","DOIUrl":null,"url":null,"abstract":"This paper improves upon Job-Level (JL) computing, a general distributed computing approach. In JL computing, a client maintains the overall search tree and parcels the overall search into coarse-grained jobs, which are then each calculated by pre-existing game-playing programs. In order to support large-scale problems such as solving 7×7 killall-Go, or building opening books for 9×9 Go or Connect6, JL computing is modified so that the entire search tree is stored in a database, as opposed to simply being stored in the client process' memory. However, the time cost of accessing this database becomes a bottleneck on performance when using a large number of computing resources. This paper proposes a cache mechanism for JL search trees. Instead of the previous approach, where the entire search tree is stored in the database, we maintain parts of the search tree in the memory of the client process to reduce the number of database accesses. Our method significantly improves the performance of job operations. Assuming that each job requires 30 seconds on average, the JL application with this cache mechanism can allow for the use of 5036 distributed computing resources in parallel without database accesses becoming the performance bottleneck.","PeriodicalId":159858,"journal":{"name":"2016 Conference on Technologies and Applications of Artificial Intelligence (TAAI)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 Conference on Technologies and Applications of Artificial Intelligence (TAAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAAI.2016.7880170","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper improves upon Job-Level (JL) computing, a general distributed computing approach. In JL computing, a client maintains the overall search tree and parcels the overall search into coarse-grained jobs, which are then each calculated by pre-existing game-playing programs. In order to support large-scale problems such as solving 7×7 killall-Go, or building opening books for 9×9 Go or Connect6, JL computing is modified so that the entire search tree is stored in a database, as opposed to simply being stored in the client process' memory. However, the time cost of accessing this database becomes a bottleneck on performance when using a large number of computing resources. This paper proposes a cache mechanism for JL search trees. Instead of the previous approach, where the entire search tree is stored in the database, we maintain parts of the search tree in the memory of the client process to reduce the number of database accesses. Our method significantly improves the performance of job operations. Assuming that each job requires 30 seconds on average, the JL application with this cache mechanism can allow for the use of 5036 distributed computing resources in parallel without database accesses becoming the performance bottleneck.