文件束缓存的最优在线算法及其向分布式缓存的推广

Tiancheng Qin, S. Etesami
{"title":"文件束缓存的最优在线算法及其向分布式缓存的推广","authors":"Tiancheng Qin, S. Etesami","doi":"10.1145/3445028","DOIUrl":null,"url":null,"abstract":"We consider a generalization of the standard cache problem called file-bundle caching, where different queries (tasks), each containing l ≥ 1 files, sequentially arrive. An online algorithm that does not know the sequence of queries ahead of time must adaptively decide on what files to keep in the cache to incur the minimum number of cache misses. Here a cache miss refers to the case where at least one file in a query is missing among the cache files. In the special case where l = 1, this problem reduces to the standard cache problem. We first analyze the performance of the classic least recently used (LRU) algorithm in this setting and show that LRU is a near-optimal online deterministic algorithm for file-bundle caching with regard to competitive ratio. We then extend our results to a generalized (h,k)-paging problem in this file-bundle setting, where the performance of the online algorithm with a cache size k is compared to an optimal offline benchmark of a smaller cache size h < k. In this latter case, we provide a randomized O(l ln k/k-h)-competitive algorithm for our generalized (h, k)-paging problem, which can be viewed as an extension of the classic marking algorithm. We complete this result by providing a matching lower bound for the competitive ratio, indicating that the performance of this modified marking algorithm is within a factor of 2 of any randomized online algorithm. Finally, we look at the distributed version of the file-bundle caching problem where there are m ≥ 1 identical caches in the system. In this case, we show that for m = l + 1 caches, there is a deterministic distributed caching algorithm that is (l2 + l)-competitive and a randomized distributed caching algorithm that is O(l ln (2l + 1)-competitive when l ≥ 2. We also provide a general framework to devise other efficient algorithms for the distributed file-bundle caching problem and evaluate the performance of our results through simulations.","PeriodicalId":105474,"journal":{"name":"ACM Transactions on Modeling and Performance Evaluation of Computing Systems (TOMPECS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Optimal Online Algorithms for File-Bundle Caching and Generalization to Distributed Caching\",\"authors\":\"Tiancheng Qin, S. Etesami\",\"doi\":\"10.1145/3445028\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We consider a generalization of the standard cache problem called file-bundle caching, where different queries (tasks), each containing l ≥ 1 files, sequentially arrive. An online algorithm that does not know the sequence of queries ahead of time must adaptively decide on what files to keep in the cache to incur the minimum number of cache misses. Here a cache miss refers to the case where at least one file in a query is missing among the cache files. In the special case where l = 1, this problem reduces to the standard cache problem. We first analyze the performance of the classic least recently used (LRU) algorithm in this setting and show that LRU is a near-optimal online deterministic algorithm for file-bundle caching with regard to competitive ratio. We then extend our results to a generalized (h,k)-paging problem in this file-bundle setting, where the performance of the online algorithm with a cache size k is compared to an optimal offline benchmark of a smaller cache size h < k. In this latter case, we provide a randomized O(l ln k/k-h)-competitive algorithm for our generalized (h, k)-paging problem, which can be viewed as an extension of the classic marking algorithm. We complete this result by providing a matching lower bound for the competitive ratio, indicating that the performance of this modified marking algorithm is within a factor of 2 of any randomized online algorithm. Finally, we look at the distributed version of the file-bundle caching problem where there are m ≥ 1 identical caches in the system. In this case, we show that for m = l + 1 caches, there is a deterministic distributed caching algorithm that is (l2 + l)-competitive and a randomized distributed caching algorithm that is O(l ln (2l + 1)-competitive when l ≥ 2. We also provide a general framework to devise other efficient algorithms for the distributed file-bundle caching problem and evaluate the performance of our results through simulations.\",\"PeriodicalId\":105474,\"journal\":{\"name\":\"ACM Transactions on Modeling and Performance Evaluation of Computing Systems (TOMPECS)\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-11-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ACM Transactions on Modeling and Performance Evaluation of Computing Systems (TOMPECS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3445028\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Modeling and Performance Evaluation of Computing Systems (TOMPECS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3445028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

我们考虑标准缓存问题的一个泛化,称为文件包缓存,其中不同的查询(任务),每个包含l≥1个文件,顺序到达。不知道查询顺序的在线算法必须自适应地决定在缓存中保留哪些文件以导致最小的缓存丢失次数。这里的缓存缺失是指在缓存文件中查询中至少缺少一个文件的情况。在l = 1的特殊情况下,这个问题简化为标准缓存问题。我们首先分析了经典的最近最少使用(least recently used, LRU)算法在这种情况下的性能,并表明LRU在竞争比方面是一个近乎最优的文件包缓存在线确定性算法。然后,我们将结果扩展到该文件束设置中的广义(h,k)分页问题,其中将缓存大小为k的在线算法的性能与较小缓存大小为h < k的最优离线基准进行比较。在后一种情况下,我们为广义(h,k)分页问题提供了一个随机的O(l ln k/k-h)竞争算法,该算法可以视为经典标记算法的扩展。我们通过为竞争比提供一个匹配的下界来完成这个结果,表明这种改进的评分算法的性能在任何随机在线算法的2倍之内。最后,我们来看一下文件包缓存问题的分布式版本,其中系统中有m≥1个相同的缓存。在这种情况下,我们证明了对于m = l + 1个缓存,存在(l2 + l)竞争的确定性分布式缓存算法和当l≥2时O(l ln (2l + 1)竞争的随机分布式缓存算法。我们还提供了一个通用框架来设计分布式文件包缓存问题的其他有效算法,并通过模拟评估我们的结果的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Optimal Online Algorithms for File-Bundle Caching and Generalization to Distributed Caching
We consider a generalization of the standard cache problem called file-bundle caching, where different queries (tasks), each containing l ≥ 1 files, sequentially arrive. An online algorithm that does not know the sequence of queries ahead of time must adaptively decide on what files to keep in the cache to incur the minimum number of cache misses. Here a cache miss refers to the case where at least one file in a query is missing among the cache files. In the special case where l = 1, this problem reduces to the standard cache problem. We first analyze the performance of the classic least recently used (LRU) algorithm in this setting and show that LRU is a near-optimal online deterministic algorithm for file-bundle caching with regard to competitive ratio. We then extend our results to a generalized (h,k)-paging problem in this file-bundle setting, where the performance of the online algorithm with a cache size k is compared to an optimal offline benchmark of a smaller cache size h < k. In this latter case, we provide a randomized O(l ln k/k-h)-competitive algorithm for our generalized (h, k)-paging problem, which can be viewed as an extension of the classic marking algorithm. We complete this result by providing a matching lower bound for the competitive ratio, indicating that the performance of this modified marking algorithm is within a factor of 2 of any randomized online algorithm. Finally, we look at the distributed version of the file-bundle caching problem where there are m ≥ 1 identical caches in the system. In this case, we show that for m = l + 1 caches, there is a deterministic distributed caching algorithm that is (l2 + l)-competitive and a randomized distributed caching algorithm that is O(l ln (2l + 1)-competitive when l ≥ 2. We also provide a general framework to devise other efficient algorithms for the distributed file-bundle caching problem and evaluate the performance of our results through simulations.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信