{"title":"On minimizing memory and computation overheads for binary-tree based data replication","authors":"S. Souravlas, Angelo Sifaleras","doi":"10.1109/ISCC.2017.8024703","DOIUrl":null,"url":null,"abstract":"Data replication is used to track the most popular files (i.e., the ones with most requests) and replicate them in selected nodes. In this way, more requests for such popular files can be completed over a period of time and bandwidth consumption is reduced, since these files do not need to be transferred from remote nodes. In this article, we extend our previous work [1] to make it more efficient in terms of memory and total computation cost, so that it becomes more efficient and suitable for larger grids. To reduce the memory costs, we present a centralized strategy which estimates the potential for selected batches of files. The computations required for these estimations are executed in a pipelined way, so their cost is also reduced.","PeriodicalId":106141,"journal":{"name":"2017 IEEE Symposium on Computers and Communications (ISCC)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE Symposium on Computers and Communications (ISCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISCC.2017.8024703","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Data replication is used to track the most popular files (i.e., the ones with most requests) and replicate them in selected nodes. In this way, more requests for such popular files can be completed over a period of time and bandwidth consumption is reduced, since these files do not need to be transferred from remote nodes. In this article, we extend our previous work [1] to make it more efficient in terms of memory and total computation cost, so that it becomes more efficient and suitable for larger grids. To reduce the memory costs, we present a centralized strategy which estimates the potential for selected batches of files. The computations required for these estimations are executed in a pipelined way, so their cost is also reduced.