Investigation of a prefetch model for low bandwidth networks

N. Tuah, Mohan J. Kumar, S. Venkatesh
{"title":"Investigation of a prefetch model for low bandwidth networks","authors":"N. Tuah, Mohan J. Kumar, S. Venkatesh","doi":"10.1145/288338.288368","DOIUrl":null,"url":null,"abstract":"We investigate speculative prefetching under a model in which prefetching is neither aborted nor preempted by demand fetch but instead gets equal priority in network bandwidth utilisation. We argue that the non-abortive assumption is appropriate for wireless networks where bandwidth is low and latency is high, and the non-preemptive assumption is appropriate for Internet where prioritization is not always possible. This paper assumes the existence of an access model to provide some knowledge about future accesses and investigates analytically the performance of a prefetcher that utilises this knowledge. In mobile computing, because resources are severely constrained, performance prediction is as important as access prediction. For uniform retrieval time, we derive a theoretical limit of improvement in access time due to prefetching. This leads to the formulation of an optimal algorithrn for prefetching one access ahead. For non-uniform retrieval time, two different types of prefetching of multiple documents, namely mainline and branch prefetch, are evaluated against prefetch of single document. In mainline prefetch, the most probable sequence of future accesses is prefetched. In branch prefetch, a set of different alternatives for future accesses is prefetched. Under some conditions, mainline prefetch may give slight improvement in user-perceived access time over single prefetch with nominal extra retrieval cost, where retrieval cost is defined as the expected network time wasted in non-useful prefetch. Branch prefetch performs better than mainline prefetch but incurs more retrieval cost.","PeriodicalId":91426,"journal":{"name":"World of wireless mobile and multimedia networks. IEEE International Symposium on a World of Wireless Mobile and Multimedia Networks","volume":"41 1","pages":"38-47"},"PeriodicalIF":0.0000,"publicationDate":"1998-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"16","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"World of wireless mobile and multimedia networks. IEEE International Symposium on a World of Wireless Mobile and Multimedia Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/288338.288368","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 16

Abstract

We investigate speculative prefetching under a model in which prefetching is neither aborted nor preempted by demand fetch but instead gets equal priority in network bandwidth utilisation. We argue that the non-abortive assumption is appropriate for wireless networks where bandwidth is low and latency is high, and the non-preemptive assumption is appropriate for Internet where prioritization is not always possible. This paper assumes the existence of an access model to provide some knowledge about future accesses and investigates analytically the performance of a prefetcher that utilises this knowledge. In mobile computing, because resources are severely constrained, performance prediction is as important as access prediction. For uniform retrieval time, we derive a theoretical limit of improvement in access time due to prefetching. This leads to the formulation of an optimal algorithrn for prefetching one access ahead. For non-uniform retrieval time, two different types of prefetching of multiple documents, namely mainline and branch prefetch, are evaluated against prefetch of single document. In mainline prefetch, the most probable sequence of future accesses is prefetched. In branch prefetch, a set of different alternatives for future accesses is prefetched. Under some conditions, mainline prefetch may give slight improvement in user-perceived access time over single prefetch with nominal extra retrieval cost, where retrieval cost is defined as the expected network time wasted in non-useful prefetch. Branch prefetch performs better than mainline prefetch but incurs more retrieval cost.
低带宽网络预取模型的研究
我们在一个模型下研究推测预取,在这个模型中,预取既不被中止,也不被需求取抢占,而是在网络带宽利用中获得同等优先级。我们认为,非流产假设适用于带宽低、延迟高的无线网络,而非抢占假设适用于优先级并不总是可能的互联网。本文假设存在一个访问模型来提供关于未来访问的一些知识,并分析地研究了利用这些知识的预取器的性能。在移动计算中,由于资源受到严重限制,因此性能预测与访问预测同样重要。对于统一的检索时间,我们导出了由于预取而提高访问时间的理论极限。这导致一个最优算法的公式预取一个访问提前。对于非均匀检索时间,将多文档的两种不同类型的预取,即主线预取和分支预取,与单个文档的预取进行比较。在主线预取中,预取最可能的未来访问顺序。在分支预取中,为将来的访问预取一组不同的备选方案。在某些情况下,与单次预取相比,主线预取可能会在用户感知的访问时间上有轻微的改善,并且会带来名义上的额外检索成本,其中检索成本定义为在无用预取中浪费的预期网络时间。分支预取比主线预取性能好,但会增加检索成本。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信