{"title":"具有对数复杂性和遗憾保证的基于梯度的在线缓存策略","authors":"Damiano Carra, Giovanni Neglia","doi":"arxiv-2405.01263","DOIUrl":null,"url":null,"abstract":"The commonly used caching policies, such as LRU or LFU, exhibit optimal\nperformance only for specific traffic patterns. Even advanced Machine\nLearning-based methods, which detect patterns in historical request data,\nstruggle when future requests deviate from past trends. Recently, a new class\nof policies has emerged that makes no assumptions about the request arrival\nprocess. These algorithms solve an online optimization problem, enabling\ncontinuous adaptation to the context. They offer theoretical guarantees on the\nregret metric, which is the gap between the gain of the online policy and the\ngain of the optimal static cache allocation in hindsight. Nevertheless, the\nhigh computational complexity of these solutions hinders their practical\nadoption. In this study, we introduce a groundbreaking gradient-based online\ncaching policy, the first to achieve logarithmic computational complexity\nrelative to catalog size along with regret guarantees. This means our algorithm\ncan efficiently handle large-scale data while minimizing the performance gap\nbetween real-time decisions and optimal hindsight choices. As requests arrive,\nour policy dynamically adjusts the probabilities of including items in the\ncache, which drive cache update decisions. Our algorithm's streamlined\ncomplexity is a key advantage, enabling its application to real-world traces\nfeaturing millions of requests and items. This is a significant achievement, as\ntraces of this scale have been out of reach for existing policies with regret\nguarantees. To the best of our knowledge, our experimental results show for the\nfirst time that the regret guarantees of gradient-based caching policies bring\nsignificant benefits in scenarios of practical interest.","PeriodicalId":501333,"journal":{"name":"arXiv - CS - Operating Systems","volume":"837 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Online Gradient-Based Caching Policy with Logarithmic Complexity and Regret Guarantees\",\"authors\":\"Damiano Carra, Giovanni Neglia\",\"doi\":\"arxiv-2405.01263\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The commonly used caching policies, such as LRU or LFU, exhibit optimal\\nperformance only for specific traffic patterns. Even advanced Machine\\nLearning-based methods, which detect patterns in historical request data,\\nstruggle when future requests deviate from past trends. Recently, a new class\\nof policies has emerged that makes no assumptions about the request arrival\\nprocess. These algorithms solve an online optimization problem, enabling\\ncontinuous adaptation to the context. They offer theoretical guarantees on the\\nregret metric, which is the gap between the gain of the online policy and the\\ngain of the optimal static cache allocation in hindsight. Nevertheless, the\\nhigh computational complexity of these solutions hinders their practical\\nadoption. In this study, we introduce a groundbreaking gradient-based online\\ncaching policy, the first to achieve logarithmic computational complexity\\nrelative to catalog size along with regret guarantees. This means our algorithm\\ncan efficiently handle large-scale data while minimizing the performance gap\\nbetween real-time decisions and optimal hindsight choices. As requests arrive,\\nour policy dynamically adjusts the probabilities of including items in the\\ncache, which drive cache update decisions. Our algorithm's streamlined\\ncomplexity is a key advantage, enabling its application to real-world traces\\nfeaturing millions of requests and items. This is a significant achievement, as\\ntraces of this scale have been out of reach for existing policies with regret\\nguarantees. To the best of our knowledge, our experimental results show for the\\nfirst time that the regret guarantees of gradient-based caching policies bring\\nsignificant benefits in scenarios of practical interest.\",\"PeriodicalId\":501333,\"journal\":{\"name\":\"arXiv - CS - Operating Systems\",\"volume\":\"837 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Operating Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2405.01263\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Operating Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.01263","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Online Gradient-Based Caching Policy with Logarithmic Complexity and Regret Guarantees
The commonly used caching policies, such as LRU or LFU, exhibit optimal
performance only for specific traffic patterns. Even advanced Machine
Learning-based methods, which detect patterns in historical request data,
struggle when future requests deviate from past trends. Recently, a new class
of policies has emerged that makes no assumptions about the request arrival
process. These algorithms solve an online optimization problem, enabling
continuous adaptation to the context. They offer theoretical guarantees on the
regret metric, which is the gap between the gain of the online policy and the
gain of the optimal static cache allocation in hindsight. Nevertheless, the
high computational complexity of these solutions hinders their practical
adoption. In this study, we introduce a groundbreaking gradient-based online
caching policy, the first to achieve logarithmic computational complexity
relative to catalog size along with regret guarantees. This means our algorithm
can efficiently handle large-scale data while minimizing the performance gap
between real-time decisions and optimal hindsight choices. As requests arrive,
our policy dynamically adjusts the probabilities of including items in the
cache, which drive cache update decisions. Our algorithm's streamlined
complexity is a key advantage, enabling its application to real-world traces
featuring millions of requests and items. This is a significant achievement, as
traces of this scale have been out of reach for existing policies with regret
guarantees. To the best of our knowledge, our experimental results show for the
first time that the regret guarantees of gradient-based caching policies bring
significant benefits in scenarios of practical interest.