{"title":"在线性网络中通过多服务器编码缓存进行内容传递","authors":"S. P. Shariatpanahi, A. Motahari, B. Khalaj","doi":"10.1109/ITWF.2015.7360777","DOIUrl":null,"url":null,"abstract":"We consider a content delivery network where multiple servers are connected to multiple cache-enabled clients. Clients request their corresponding contents from the servers and servers collaboratively transmit packets to fulfill all the requests. It is assumed that some contents are stored in the caches in off-peak time of the network without knowing the actual requests, the so called cache content placement phase. The goal is to minimize the worst case delay in the content delivery phase. Considering a random linear network, we propose a coding strategy which exploits servers' multiplexing gains as well as caches' global and local coding gains. The main idea in our coding scheme is to expand the number of users benefiting from a single packet by using zero-forcing techniques. This results in an increase in multi-casting gain which in turn provides faster delivery of the contents to the users. In addition, we show that our scheme is optimal for a certain regime of parameters.","PeriodicalId":281890,"journal":{"name":"2015 IEEE Information Theory Workshop - Fall (ITW)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Content delivery via multi-server coded caching in linear networks\",\"authors\":\"S. P. Shariatpanahi, A. Motahari, B. Khalaj\",\"doi\":\"10.1109/ITWF.2015.7360777\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We consider a content delivery network where multiple servers are connected to multiple cache-enabled clients. Clients request their corresponding contents from the servers and servers collaboratively transmit packets to fulfill all the requests. It is assumed that some contents are stored in the caches in off-peak time of the network without knowing the actual requests, the so called cache content placement phase. The goal is to minimize the worst case delay in the content delivery phase. Considering a random linear network, we propose a coding strategy which exploits servers' multiplexing gains as well as caches' global and local coding gains. The main idea in our coding scheme is to expand the number of users benefiting from a single packet by using zero-forcing techniques. This results in an increase in multi-casting gain which in turn provides faster delivery of the contents to the users. In addition, we show that our scheme is optimal for a certain regime of parameters.\",\"PeriodicalId\":281890,\"journal\":{\"name\":\"2015 IEEE Information Theory Workshop - Fall (ITW)\",\"volume\":\"35 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-12-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 IEEE Information Theory Workshop - Fall (ITW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITWF.2015.7360777\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE Information Theory Workshop - Fall (ITW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITWF.2015.7360777","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Content delivery via multi-server coded caching in linear networks
We consider a content delivery network where multiple servers are connected to multiple cache-enabled clients. Clients request their corresponding contents from the servers and servers collaboratively transmit packets to fulfill all the requests. It is assumed that some contents are stored in the caches in off-peak time of the network without knowing the actual requests, the so called cache content placement phase. The goal is to minimize the worst case delay in the content delivery phase. Considering a random linear network, we propose a coding strategy which exploits servers' multiplexing gains as well as caches' global and local coding gains. The main idea in our coding scheme is to expand the number of users benefiting from a single packet by using zero-forcing techniques. This results in an increase in multi-casting gain which in turn provides faster delivery of the contents to the users. In addition, we show that our scheme is optimal for a certain regime of parameters.