{"title":"Packet dropping for H.264 videos considering both coding and packet-loss artifacts","authors":"Yuxia Wang, Ting-Lan Lin, P. Cosman","doi":"10.1109/PV.2010.5706834","DOIUrl":"https://doi.org/10.1109/PV.2010.5706834","url":null,"abstract":"In the context of both compression artifacts and packet loss artifacts, we use generalized linear models to predict VQM quality scores. Using a network-based model, a router can estimate the visual importance of each incoming packet and decide which packet to drop when congestion happens. Considering a wide variety of bit reduction rates, we perform packet dropping experiments for combinations of video streams and examine the effects of video contents and different bit rates. By comparing with randomly dropping B slices or B frames, we conclude that our model gives a good performance on objective evaluation of packet importance.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"582 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116296535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A speech-video synchrony quality metric using CoIA","authors":"Yaodu Wei, Xiang Xie, Jingming Kuang, Xinlu Han","doi":"10.1109/PV.2010.5706835","DOIUrl":"https://doi.org/10.1109/PV.2010.5706835","url":null,"abstract":"A quality model was built to assess the influence of speech-video asynchrony on the audio-visual quality perception. The audio-visual contents were separated into two categories: “speaker inside” and “speaker outside”, depending on whether the speaker is inside the video. For the first category, speech was shifted in a small scale. DCT and MFCC coefficients were calculated from video and speech separately. A Co-inertia Analysis (CoIA) was used to decide the speech-video correlation, and as the speech progressively shifts, a correlation curve emerged. The curve was modeled by an Gaussian function, and then the function was used to predict the perceptual quality. On the other hand, a Gaussian curve was used to predict the perceptual quality of the “speaker outside” category. A subjective test proved the effectiveness of the proposed method.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132027272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pengye Xia, S. Chan, M. Chiang, G. Shi, Hongbo Zhang, Liufei Wen, Z. Yan
{"title":"Distributed joint optimization of traffic engineering and server selection","authors":"Pengye Xia, S. Chan, M. Chiang, G. Shi, Hongbo Zhang, Liufei Wen, Z. Yan","doi":"10.1109/PV.2010.5706824","DOIUrl":"https://doi.org/10.1109/PV.2010.5706824","url":null,"abstract":"Internet service providers (ISP) apply traffic engineering (TE) in the underlay network to avoid congestion. On the other hand, content providers (CP) use different server selection (SS) strategies in the overlay network to reduce delay. It has been shown that a joint optimization of TE and SS is beneficial to the performance from both ISP's and CP's perspectives. One challenging issue in such a network is to design a distributed protocol which achieves optimality while revealing as little information as possible between ISP and CP. To address this problem, we propose a distributed protocol termed PETS, in which each router of ISP makes independent traffic engineering decision and each server of CP makes independent server selection decision. We prove that PETS can achieve optimality for the joint optimization of TE and SS. We also show that PETS can significantly reduce message passing and enables ISP to hide important underlay network information (e.g., topology) from CP. Furthermore, PETS can be easily extended to handle the case of multiple CPs in the network.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131413340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An optimized resource allocation algorithm for scalable video delivery over wireless multicast links","authors":"Seong-Ping Chuah, Zhenzhong Chen, Yap-Peng Tan","doi":"10.1109/PV.2010.5706818","DOIUrl":"https://doi.org/10.1109/PV.2010.5706818","url":null,"abstract":"In this paper, we investigate resource allocation of video multicast over resource-constrained wireless networks. The video multicast is a best-effort service which suffers from limited transmission energy and channel access time. To cater for multi-resolution video demands of heterogeneous mobile clients, we consider scalable video coding (SVC) which offers spatial, temporal and fidelity scalabilities. We formulate multicast strategy that maximizes video quality under transmission energy and channel access constraints. We show that the problem is a joint optimization of mode selection and transmission power allocation. It is a mixed-integer programming problem which is in general NP-hard. To solve the optimization problem, we first reduce the problem into a binary integer programming problem. By exploiting inter-layer dependency of SVC packets, we propose a novel solution based on the dynamic programming approach. Experiment results demonstrate that the proposed method archives notable improvement of average video quality over a conventional method.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121308435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Frame layer rate control for dual frame motion compensation","authors":"Da Liu, Debin Zhao, Siwei Ma, Wen Gao","doi":"10.1109/PV.2010.5706823","DOIUrl":"https://doi.org/10.1109/PV.2010.5706823","url":null,"abstract":"Rate control plays an important role in video coding. When jump update dual frame motion compensation (JU-DFMC) is utilized in video coding, two kinds of frames are existed according to bit allocation, one is low quality frame (LQF), the other is high quality frame (HQF). For each current frame (LQF or HQF), one short term reference frame (STR) and one long term reference frame (LTR) are utilized for motion compensation. Owning to this kind of coding structure, rate control for JU-DFMC is different from traditional methods. In this paper, a rate control scheme for JU-DFMC is proposed. Firstly, a linear bit allocation method is proposed for LOFs to obtain smooth quality. Secondly, different source rate and quantization stepsize (Qstep) models are proposed to different frames (LQFs or HQFs). Thirdly, overhead prediction and MAD prediction are presented. Experimental results show that the proposed method can bring accurate rate control for JU-DFMC, at the same time the PSNR is better than the other methods.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134100010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Quality of Experience estimation using frame loss pattern and video encoding characteristics in DVB-H networks","authors":"K. Singh, G. Rubino","doi":"10.1109/PV.2010.5706832","DOIUrl":"https://doi.org/10.1109/PV.2010.5706832","url":null,"abstract":"Automatic estimation of Quality of Experience (QoE) is of key importance for mobile television networks such as DVB-H. These networks can install network probes in order to monitor QoE. The QoE feedback can be used to take some corrective measures, in case the quality drops, to bring back QoE to satisfactory level. In this paper, we extend a previously proposed noreference QoE monitoring module for H.264 video over DVB-H networks. We consider an additional parameter called quantisation parameter (QP) and consider frame loss pattern, instead of packet loss pattern, apart from the parameters used in the earlier work such as motion activity and loss rank in a Group of Pictures (GOP). The earlier work is restricted to a fixed encoding bitrate. By considering QP this restriction is removed because QP determines the bitrate as well as the resulting video quality. The results show that our estimation module based on Random Neural Networks (RNN) captures the non-linear relationship between these parameters and QoE. Moreover, our consideration of additional parameters leads to significant improvement in QoE estimation accuracy.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123803918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tamar Barzuza, Sagee Ben Zedeff, Ori Modai, Leonid Vainbrand, Yair Wiener, Einat Yellin
{"title":"Trend: A dynamic bandwidth estimation and adaptation algorithm for real-time video calling","authors":"Tamar Barzuza, Sagee Ben Zedeff, Ori Modai, Leonid Vainbrand, Yair Wiener, Einat Yellin","doi":"10.1109/PV.2010.5706829","DOIUrl":"https://doi.org/10.1109/PV.2010.5706829","url":null,"abstract":"Video calling over the internet is becoming more and more popular in recent years. The quality of video in such applications is a major contributor to the overall user experience during the call. As received video quality degrades significantly when the network path used is either under-utilized or over-utilized, a method for estimating network bandwidth availability for adapting video bit rate in video calls is highly desirable. This paper presents TREND, a novel technique for dynamic bandwidth estimation and adaptation designed for real-time, interactive video calling. By detecting the delay in the received video frames, TREND estimates available network bandwidth in the path and adapts the encoded bitrate sent to best fit it before packets are lost. TREND uses only standard protocols for delay detection and flow control, and can therefore be easily and effectively used with traditional visual communications systems. After testing TREND under various use cases we have found that it efficiently estimates available bandwidth while outperforming competing tools in terms of detection time and accuracy. By implementing TREND in its video conferencing solutions, RADVISION was able to show a superior quality of experience for its video calls over other solutions available today.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"382 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123690116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Joint forward backward error concealment of redundantly coded video","authors":"Xiaopeng Fan, O. Au, Debin Zhao, Wen Gao","doi":"10.1109/PV.2010.5706816","DOIUrl":"https://doi.org/10.1109/PV.2010.5706816","url":null,"abstract":"Error concealment at the video decoder is to recover erroneous picture region based on correctly decoded region in the same frame or the neighboring frames. However, the existing error concealment methods cannot give satisfactory result in some situations, e.g. when a whole frame is lost. In this paper, we propose a novel joint forward-backward concealment (JFBC) method. The key observation is that, a future frame can be recovered without using its reference frame when some error resilience tools, such as redundant picture, are used. Based on this observation, the proposed JFBC performs temporal error concealment along not only forward direction but also backward direction. Furthermore, the JFBC combines the forward concealment result and the backward concealment result by using linear minimum mean square error (LMMSE) estimation. The JFBC estimates for each pixel an error level such that the weights in LMMSE estimation are adaptive in pixel-level. In the experiments, the JFBC achieves considerable gain over the state-of-the-art temporal error concealment method.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128238042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Forward Error Protection for low-delay packet video","authors":"Zhi Li, A. Khisti, B. Girod","doi":"10.1109/PV.2010.5706813","DOIUrl":"https://doi.org/10.1109/PV.2010.5706813","url":null,"abstract":"We study different forward error correction (FEC) codes for packet video streaming over erasure channels with strict delay constraints. Our study includes traditional maximum distance separable (MDS) codes and streaming burst erasure codes with optimal delay performance. We develop a continuous-times model to calculate burst error correction capabilities of these codes with a delay constraint. Our analysis also incorporates Systematic Lossy Error Protection (SLEP) that achieves stronger error protection in exchange for a slight drop in video quality when error correction is needed. We provide simulation results for transmitting H.264/AVC encoded video over a bursty packet erasure channel and show that the combination of streaming erasure codes and SLEP greatly outperforms conventional MDS FEC for video streaming with a tight delay constraint.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126559332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Aditya Mavlankar, P. Agrawal, Derek Pang, Sherif A. Halawa, Ngai-Man Cheung, B. Girod
{"title":"An interactive region-of-interest video streaming system for online lecture viewing","authors":"Aditya Mavlankar, P. Agrawal, Derek Pang, Sherif A. Halawa, Ngai-Man Cheung, B. Girod","doi":"10.1109/PV.2010.5706821","DOIUrl":"https://doi.org/10.1109/PV.2010.5706821","url":null,"abstract":"ClassX is an interactive online lecture viewing system developed at Stanford University. Unlike existing solutions that restrict the user to watch only a pre-defined view, ClassX allows interactive pan/tilt/zoom while watching the video. The interactive video streaming paradigm avoids sending the entire field-of-view in the recorded high resolution, thus reducing the required data rate. To alleviate the navigation burden on the part of the online viewer, ClassX offers automatic tracking of the lecturer. ClassX also employs slide recognition technology, which allows automatic synchronization of digital presentation slides with those appearing in the lecture video. This paper presents a design overview of the ClassX system and the evaluation results of a 3-month pilot deployment at Stanford University. The results demonstrate that our system is a low-cost, efficient and pragmatic solution to interactive online lecture viewing.","PeriodicalId":339319,"journal":{"name":"2010 18th International Packet Video Workshop","volume":"282 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124508419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}