协调弹性Web缓存的成本和性能目标

Farhana Kabir, David Chiu
{"title":"协调弹性Web缓存的成本和性能目标","authors":"Farhana Kabir, David Chiu","doi":"10.1109/CSC.2012.21","DOIUrl":null,"url":null,"abstract":"Web and service applications are generally I/O bound and follow a Zipf-like request distribution, ushering in potential for significant latency reduction by caching and reusing results. However, such web caches require manual resource allocation, and when deployed in the cloud, costs may further complicate the provisioning process. We propose a fully autonomous, self-scaling, and cost-aware cloud cache with the objective of accelerating data-intensive applications. Our system, which is distributed over multiple cloud nodes, intelligently provisions resources at runtime based on user's cost and performance expectations, while abstracting the various low-level decisions regarding efficient cloud resource management and data placement within the cloud from the user. Our prediction model lends the system the capability to auto-configure the optimal resource requirement to automatically scale itself up (or down) to accommodate demand peaks while staying within certain cost constraints while fulfilling the performance expectations. Our evaluation shows a 5.5 time speedup for a typical web workload, while staying under cost constraints.","PeriodicalId":183800,"journal":{"name":"2012 International Conference on Cloud and Service Computing","volume":"138 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Reconciling Cost and Performance Objectives for Elastic Web Caches\",\"authors\":\"Farhana Kabir, David Chiu\",\"doi\":\"10.1109/CSC.2012.21\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Web and service applications are generally I/O bound and follow a Zipf-like request distribution, ushering in potential for significant latency reduction by caching and reusing results. However, such web caches require manual resource allocation, and when deployed in the cloud, costs may further complicate the provisioning process. We propose a fully autonomous, self-scaling, and cost-aware cloud cache with the objective of accelerating data-intensive applications. Our system, which is distributed over multiple cloud nodes, intelligently provisions resources at runtime based on user's cost and performance expectations, while abstracting the various low-level decisions regarding efficient cloud resource management and data placement within the cloud from the user. Our prediction model lends the system the capability to auto-configure the optimal resource requirement to automatically scale itself up (or down) to accommodate demand peaks while staying within certain cost constraints while fulfilling the performance expectations. Our evaluation shows a 5.5 time speedup for a typical web workload, while staying under cost constraints.\",\"PeriodicalId\":183800,\"journal\":{\"name\":\"2012 International Conference on Cloud and Service Computing\",\"volume\":\"138 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-11-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 International Conference on Cloud and Service Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CSC.2012.21\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 International Conference on Cloud and Service Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSC.2012.21","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

摘要

Web和服务应用程序通常是I/O绑定的,并遵循类似zipf的请求分布,通过缓存和重用结果可以显著减少延迟。然而,这样的web缓存需要手动资源分配,并且当部署在云中时,成本可能会使配置过程进一步复杂化。我们提出了一个完全自主、自扩展和成本敏感的云缓存,目标是加速数据密集型应用程序。我们的系统分布在多个云节点上,在运行时根据用户的成本和性能期望智能地提供资源,同时从用户那里抽象出有关高效云资源管理和云内数据放置的各种低级决策。我们的预测模型赋予系统自动配置最优资源需求的能力,使其能够自动向上(或向下)扩展以适应需求高峰,同时在满足性能期望的同时保持一定的成本约束。我们的评估显示,对于典型的web工作负载,在保持成本限制的情况下,时间加快了5.5。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Reconciling Cost and Performance Objectives for Elastic Web Caches
Web and service applications are generally I/O bound and follow a Zipf-like request distribution, ushering in potential for significant latency reduction by caching and reusing results. However, such web caches require manual resource allocation, and when deployed in the cloud, costs may further complicate the provisioning process. We propose a fully autonomous, self-scaling, and cost-aware cloud cache with the objective of accelerating data-intensive applications. Our system, which is distributed over multiple cloud nodes, intelligently provisions resources at runtime based on user's cost and performance expectations, while abstracting the various low-level decisions regarding efficient cloud resource management and data placement within the cloud from the user. Our prediction model lends the system the capability to auto-configure the optimal resource requirement to automatically scale itself up (or down) to accommodate demand peaks while staying within certain cost constraints while fulfilling the performance expectations. Our evaluation shows a 5.5 time speedup for a typical web workload, while staying under cost constraints.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信