2011 IEEE 13th International Workshop on Multimedia Signal Processing最新文献

筛选
英文 中文
Image quality assessment based on multiple watermarking approach 基于多重水印的图像质量评估方法
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093787
N. Baaziz, Dong Zheng, Demin Wang
{"title":"Image quality assessment based on multiple watermarking approach","authors":"N. Baaziz, Dong Zheng, Demin Wang","doi":"10.1109/MMSP.2011.6093787","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093787","url":null,"abstract":"Automatic monitoring of image/video quality is very important in modern multimedia communication services. We are interested in digital watermarking as a promising approach to image quality assessment without reference to the original image. The proposed methodology makes use of wavelet-based embedding of multiple watermarks with robustness control in order to capture the degree of the degradation undergone by a received image. The watermark robustness is controlled through 1) embedding and detection of multiple watermarks, 2) multi-resolution and directional subband selection, 3) perceptual watermark weighting and 4) fine watermark strength adjustment process. At the receiver end, the detection or lack of detection of the watermarks in a received image are used to estimate image's PSNR range and determine its associated quality attribute. Simulation results show the efficiency of such watermarking scheme in assessing the quality level of test images under JPEG compression.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124895521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Angular intra prediction in High Efficiency Video Coding (HEVC) 高效视频编码(HEVC)中的角内预测
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093806
J. Lainema, K. Ugur
{"title":"Angular intra prediction in High Efficiency Video Coding (HEVC)","authors":"J. Lainema, K. Ugur","doi":"10.1109/MMSP.2011.6093806","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093806","url":null,"abstract":"New video coding solutions, such as the HEVC (High Efficiency Video Coding) standard being developed by JCT-VC (Joint Collaborative Team on Video Coding), are typically designed for high resolution video content. Increasing video resolution creates two basic requirements for practical video codecs; those need to be able to provide compression efficiency superior to prior video coding solutions and the computational requirements need to be aligned with the foreseeable hardware platforms. This paper proposes an intra prediction method which is designed to provide high compression efficiency and which can be implemented effectively in resource constrained environments making it applicable to wide range of use cases. When designing the method, special attention was given to the algorithmic definition of the prediction sample generation, in order to be able to utilize the same reconstruction process at different block sizes. The proposed method outperforms earlier variations of the same family of technologies significantly and consistently across different classes of video material, and has recently been adopted as the directional intra prediction method for the draft HEVC standard. Experimental results show that the proposed method outperforms the H.264/AVC intra prediction approach on average by 4.8 %. For sequences with dominant directional structures, the coding efficiency gains become more significant and exceed 10 %.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123026585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
ViewMark: An interactive videoconferencing system for mobile devices ViewMark:针对移动设备的交互式视频会议系统
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093792
Shu Shi, Zhengyou Zhang
{"title":"ViewMark: An interactive videoconferencing system for mobile devices","authors":"Shu Shi, Zhengyou Zhang","doi":"10.1109/MMSP.2011.6093792","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093792","url":null,"abstract":"ViewMark, a server-client based interactive mobile videoconferencing system is proposed in this paper to enhance the remote meeting experience for mobile users. Compared with the state-of-the-art mobile videoconferencing technology, ViewMark is novel in allowing a mobile user to interactively change the viewpoint of the remote video, create viewmarks, and hear with spatial audio. In addition, ViewMark also streams the screen of the presentation slides to mobile devices. In this paper, we introduce the system design of ViewMark in details, compare the devices that can be used to implement interactive videoconferencing, and demonstrate the prototype system we have built on Windows Mobile platform.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128127843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
GPU based fast algorithm for tanner graph based image interpolation 基于GPU的tanner图图像插值快速算法
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093783
Wei Lei, Ruiqin Xiong, Siwei Ma, Luhong Liang
{"title":"GPU based fast algorithm for tanner graph based image interpolation","authors":"Wei Lei, Ruiqin Xiong, Siwei Ma, Luhong Liang","doi":"10.1109/MMSP.2011.6093783","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093783","url":null,"abstract":"In image/video processing software and hardware products, low complexity interpolation algorithms, such as cubic and splines methods, are commonly used. However, these methods tend to blur textures and produce jaggy effect compared with other adaptive methods such as NEDI, SAI. Tanner graph based image interpolation algorithm has better effect in dealing with edge and texture, but with high computation complexity. Thanks to the high performance parallel processing capability of today's GPU, use of complex algorithms for real time application is becoming possible. In this paper, we present a fast algorithm for tanner graph based image interpolation and it's implementation on GPU. In our algorithm, the image model training process of tanner graph based image interpolation is greatly simplified. Experimental results show that the GPU implementation can be more than 47 times as fast as the CPU implementation.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130349420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Separation of speech sources using an Acoustic Vector Sensor 使用声矢量传感器分离语音源
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093797
M. Shujau, C. Ritz, I. Burnett
{"title":"Separation of speech sources using an Acoustic Vector Sensor","authors":"M. Shujau, C. Ritz, I. Burnett","doi":"10.1109/MMSP.2011.6093797","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093797","url":null,"abstract":"This paper investigates how the directional characteristics of an Acoustic Vector Sensor (AVS) can be used to separate speech sources. The technique described in this work takes advantage of the frequency domain direction of arrival estimates to identify the location, relative to the AVS array, of each individual speaker in a group of speakers and separate them accordingly into individual speech signals. Results presented in this work show that the technique can be used for real-time separation of speech sources using a single 20ms frame of speech, furthermore the results presented show that there is an average improvement in the Signal to Interference Ratio (SIR) for the proposed algorithm over the unprocessed recording of 15.1 dB and an average improvement of 5.4 dB in terms of Signal to Distortion Ratio (SDR) over the unprocessed recordings. In addition to the SIR and SDR results, Perceptual Evaluation of Speech Quality (PESQ) and listening tests both show an improvement in perceptual quality of 1 Mean Opinion Score (MOS) over unprocessed recordings.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"29 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114102158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Region of interest determination using human computation 用人工计算确定感兴趣的区域
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093839
Flavio P. Ribeiro, D. Florêncio
{"title":"Region of interest determination using human computation","authors":"Flavio P. Ribeiro, D. Florêncio","doi":"10.1109/MMSP.2011.6093839","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093839","url":null,"abstract":"The ability to identify and track visually interesting regions has many practical applications — for example, in image and video compression, visual marketing and foveal machine vision. Due to challenges in modeling the peculiarities of human physiological and psychological responses, automatic detection of fixation points is an open problem. Indeed, no objective methods are currently capable of fully modeling the human perception of regions of interest (ROIs). Thus, research often relies on user studies with eye tracking systems. In this paper we propose a cost-effective and convenient alternative, obtained by having internet workers annotate videos with ROI coordinates. The workers use an interactive video player with a simulated mouse-driven fovea, which models the fall-off in resolution of the human visual system. Since this approach is not supervised, we implement methods for identifying inaccurate or malicious results. Using this proposal, one can collect ROI data in an automated fashion, and at a much lower cost than laboratory studies.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114567645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Low-complexity, near-lossless coding of depth maps from kinect-like depth cameras 低复杂度,近乎无损的深度图编码,来自类似kinect的深度相机
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093803
S. Mehrotra, Zhengyou Zhang, Q. Cai, Cha Zhang, P. Chou
{"title":"Low-complexity, near-lossless coding of depth maps from kinect-like depth cameras","authors":"S. Mehrotra, Zhengyou Zhang, Q. Cai, Cha Zhang, P. Chou","doi":"10.1109/MMSP.2011.6093803","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093803","url":null,"abstract":"Depth cameras are gaining interest rapidly in the market as depth plus RGB is being used for a variety of applications ranging from foreground/background segmentation, face tracking, activity detection, and free viewpoint video rendering. In this paper, we present a low-complexity, near-lossless codec for coding depth maps. This coding requires no buffering of video frames, is table-less, can encode or decode a frame in close to 5ms with little code optimization, and provides between 7:1 to 16:1 compression ratio for near-lossless coding of 16-bit depth maps generated by the Kinect camera.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130059674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Block-based codebook model with oriented-gradient feature for real-time foreground detection 面向前景实时检测的基于块的梯度特征码本模型
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093830
Jiu Xu, Ning Jiang, S. Goto
{"title":"Block-based codebook model with oriented-gradient feature for real-time foreground detection","authors":"Jiu Xu, Ning Jiang, S. Goto","doi":"10.1109/MMSP.2011.6093830","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093830","url":null,"abstract":"In this paper, a novel approach is proposed to achieve the foreground objects detection in video surveillance system using codebook method. The block-based background model upgrades the pixel-based codebook model to block level which can utilize the dependency and find relationships between neighbouring pixels, thus improving the processing speed and reducing memory during model construction and foreground detection. Moreover, by adding the orientation and magnitude of the block gradient, the codebook model contains not only information of color, but also the texture feature. The texture information can further reduce noises and refine more entire foreground regions. Experimental results prove that our method has better performance compared with the standard codebook and some other former algorithms.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115940863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
An MDC-based video streaming architecture for mobile networks 基于mdc的移动网络视频流架构
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093829
C. Greco, G. Petrazzuoli, Marco Cagnazzo, B. Pesquet-Popescu
{"title":"An MDC-based video streaming architecture for mobile networks","authors":"C. Greco, G. Petrazzuoli, Marco Cagnazzo, B. Pesquet-Popescu","doi":"10.1109/MMSP.2011.6093829","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093829","url":null,"abstract":"Multiple description coding (MDC) is a framework designed to improve the robustness of video content transmission in lossy environments. In this work, we propose an MDC technique using a legacy coder to produce two descriptions, based on separation of even and odd frames. If only one description is received, the missing frames are reconstructed using temporal high-order motion interpolation (HOMI), a technique originally proposed for distributed video coding. If both descriptions are received, the frames are reconstructed as a block-wise linear combination of the two descriptions, with the coefficient computed at the encoder in a RD-optimised fashion, encoded with a context-adaptive arithmetic coder, and sent as side information. We integrated the proposed technique in a mobile ad-hoc streaming protocol, and tested it using a group mobility model. The results show a non-negligible gain for the expected video quality, with respect to the reference technique.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128188788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
L1-norm multi-frame super-resolution from images with zooming motion l1范数多帧超分辨率的图像缩放运动
2011 IEEE 13th International Workshop on Multimedia Signal Processing Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093847
Yushuang Tian, Kim-Hui Yap, Li Chen
{"title":"L1-norm multi-frame super-resolution from images with zooming motion","authors":"Yushuang Tian, Kim-Hui Yap, Li Chen","doi":"10.1109/MMSP.2011.6093847","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093847","url":null,"abstract":"This paper proposes a new image super-resolution (SR) approach to reconstruct a high-resolution (HR) image by fusing multiple low-resolution (LR) images with zooming motion. Most conventional SR image reconstruction methods assume that the motion among different images consists of only translation and possibly rotation. This in-plane motion model, however, is not practical in some applications, when relative zooming exists among the acquired LR images. In view of this, this paper presents a new SR method that addresses a motion model including both in-plane motion (e.g. translation and rotation) and zooming motion. Based on this model, a maximum a posteriori (MAP) based SR algorithm using L1-norm optimization is proposed. Experimental results show that the proposed algorithm based on the new motion model performs well in terms of visual evaluation and quantitative measurement.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"7 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133090968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信