2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)最新文献

筛选
英文 中文
Visual conditioning for augmented-reality-assisted video conferencing 增强现实辅助视频会议的视觉调节
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-12-31 DOI: 10.1109/MMSP.2012.6343418
O. Guleryuz, T. Kalker
{"title":"Visual conditioning for augmented-reality-assisted video conferencing","authors":"O. Guleryuz, T. Kalker","doi":"10.1109/MMSP.2012.6343418","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343418","url":null,"abstract":"Typical video conferencing scenarios bring together individuals from disparate environments. Unless one commits to expensive tele-presence rooms, conferences involving many individuals result in a cacophony of visuals and backgrounds. Ideally one would like to separate participant visuals from their respective environments and render them over visually pleasing backgrounds that enhance immersion for all. Yet available image/video segmentation techniques are limited and result in significant artifacts even with recently popular commodity depth sensors. In this paper we present a technique that accomplishes robust and visually pleasing rendering of segmented participants over adaptively-designed virtual backgrounds. Our method works by determining virtual backgrounds that match and highlight participant visuals and uses directional textures to hide segmentation artifacts due to noisy segmentation boundaries, missing regions, etc. Taking advantage of simple computations and look-up-tables, our work leads to fast, real-time implementations that can run on mobile and other computationally-limited platforms.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130174511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A P300-based BCI classification algorithm using median filtering and Bayesian feature extraction 一种基于p300的中值滤波和贝叶斯特征提取的BCI分类算法
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343459
Xiaoou Li, Feng Wang, Xu Chen, R. Ward
{"title":"A P300-based BCI classification algorithm using median filtering and Bayesian feature extraction","authors":"Xiaoou Li, Feng Wang, Xu Chen, R. Ward","doi":"10.1109/MMSP.2012.6343459","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343459","url":null,"abstract":"A brain computer interface (BCI) system translates a person's brain activity into useful control or communication signals. In this paper, an effective P300-based BCI identification algorithm using median filtering and Bayesian classifier is proposed to improve the classification accuracy and computation efficiency of P300-based BCI. Median filtering is firstly applied to remove noises and Bayesian Linear Discriminant Analysis (BLDA) is then employed for classification. Testing on the P300 speller paradigm in dataset II of 2004 BCI Competition III, we show that a 90% average classification accuracy can be achieved and the highest accuracy is 100%. The proposed method is also computationally efficient and thus it represents a practical implementation for man-computer communication control, especially for on-line applications.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115386608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Lagrangian framework for video analytics 视频分析的拉格朗日框架
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343474
A. Kuhn, T. Senst, I. Keller, T. Sikora, H. Theisel
{"title":"A Lagrangian framework for video analytics","authors":"A. Kuhn, T. Senst, I. Keller, T. Sikora, H. Theisel","doi":"10.1109/MMSP.2012.6343474","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343474","url":null,"abstract":"The extraction of motion patterns from image sequences based on the optical flow methodology is an important and timely topic among visual multi media applications. In this work we will present a novel framework that combines the optical flow methodology from image processing with methods developed for the Lagrangian analysis of time-dependent vector fields. The Lagrangian approach has been proven to be a valuable and powerful tool to capture the complex dynamic motion behavior within unsteady vector fields. To come up with a compact and applicable framework, this paper will provide concepts on how to compute trajectory-based Lagrangian measures in series of optical flow fields, a set of basic measures to capture the essence of the motion behavior within the image, and a compact hierarchical, feature-based description of the resulting motion features. The resulting framework will bee shown to be suitable for an automated image analysis as well as compact visual analysis of image sequences in its spatio-temporal context. We show its applicability for the task of motion feature description and extraction on different temporal scales, crowd motion analysis, and automated detection of abnormal events within video sequences.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125159090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
An optimal client buffer model for multiplexing HTTP streams 用于多路复用HTTP流的最佳客户端缓冲模型
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343455
Saayan Mitra, Viswanathan Swaminathan
{"title":"An optimal client buffer model for multiplexing HTTP streams","authors":"Saayan Mitra, Viswanathan Swaminathan","doi":"10.1109/MMSP.2012.6343455","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343455","url":null,"abstract":"The basic tenet of HTTP streaming is to deliver fragments of video and audio that are individually addressable chunks of content over HTTP. Some media players consume incoming video and audio data only in a time ordered multiplexed format. If alternate tracks need to be added post packaging of the media, it has to be repackaged that involves duplication resulting in multiple multiplexed files. Additionally for adaptive streaming, a set of all those files need to be added for each bitrate. Alternatively, it is more efficient to store component tracks separately, fetching only the required tracks and multiplexing audio and video in the client before sending the data to the decoder. To deliver an optimal viewing experience, the client has to take care of the seemingly conflicting constraints viz., handling the network jitter, minimizing the time to switch to an alternate track and minimizing the live latency. For instance, to absorb more network jitter more data should be available in the buffers but this would increase the switching latency. We introduce a formal buffer model for a client that gathers video and audio fragments and multiplexes them on the fly. This model uses separate video and audio buffers, a multiplexed buffer in the application, and decoding buffer associated with the decoder. We model the buffer sizes, their thresholds to request data from the network, and the rate of transfer of data between buffers. We show that these buffers can be designed varying these parameters to optimize for the above constraints. This buffer model can also be leveraged for deciding when to switch in adaptive bitrate streaming. We further validate these by experimental results from our implementation.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114279101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Human action recognition using Lagrangian descriptors 使用拉格朗日描述符的人类动作识别
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343469
Esra Acar, T. Senst, A. Kuhn, I. Keller, H. Theisel, S. Albayrak, T. Sikora
{"title":"Human action recognition using Lagrangian descriptors","authors":"Esra Acar, T. Senst, A. Kuhn, I. Keller, H. Theisel, S. Albayrak, T. Sikora","doi":"10.1109/MMSP.2012.6343469","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343469","url":null,"abstract":"Human action recognition requires the description of complex motion patterns in image sequences. In general, these patterns span varying temporal scales. In this context, Lagrangian methods have proven to be valuable for crowd analysis tasks such as crowd segmentation. In this paper, we show that, besides their potential in describing large scale motion patterns, Lagrangian methods are also well suited to model complex individual human activities over variable time intervals. We use Finite Time Lyapunov Exponents and time-normalized arc length measures in a linear SVM classification scheme. We evaluated our method on the Weizmann and KTH datasets. The results demonstrate that our approach is promising and that human action recognition performance is improved by fusing Lagrangian measures.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122032581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Interactive mobile visual search for social activities completion using query image contextual model 交互式移动视觉搜索的社交活动完成使用查询图像上下文模型
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343447
Ning Zhang, Tao Mei, Xiansheng Hua, L. Guan, Shipeng Li
{"title":"Interactive mobile visual search for social activities completion using query image contextual model","authors":"Ning Zhang, Tao Mei, Xiansheng Hua, L. Guan, Shipeng Li","doi":"10.1109/MMSP.2012.6343447","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343447","url":null,"abstract":"Mobile devices are ubiquitous. People use their phones as a personal concierge not only discovering information but also searching for particular interest on-the-go and making decisions. This brings a new horizon for multimedia retrieval on mobile. While existing efforts have predominantly focused on understanding textual or a voice query, this paper presents a new perspective which understands visual queries captured by the built-in camera such that mobile-based social activities can be recommended for users to complete. In this work, a query image-based contextual model is proposed for visual search. A mobile user can take a photo and naturally indicate an object-of-interest within the photo via circle based gesture called “O” gesture. Both selected object-of-interest region as well as surrounding visual context in photo are used in achieving a search-based recognition by retrieving similar images based on a large-scale of visual vocabulary tree. Consequently, social activities such as visiting contextually relevant entities (i.e., local businesses) are recommended to the users based on their visual queries and GPS location. Along with the proposed method, an exemplary real application has been developed on Windows Phone 7 devices and evaluated with a wide variety of scenarios on million-scale image database. To test the performance of proposed mobile visual search model, extensive experimentation has been conducted and compared with state-of-the-art algorithms in content-based image retrieval (CBIR) domain.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116879288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Video jitter analysis for automatic bootleg detection 视频抖动分析自动盗版检测
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343423
M. V. Scarzanella, P. Dragotti
{"title":"Video jitter analysis for automatic bootleg detection","authors":"M. V. Scarzanella, P. Dragotti","doi":"10.1109/MMSP.2012.6343423","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343423","url":null,"abstract":"This paper presents a novel technique for the automatic detection of recaptured videos with applications to video forensics. The proposed technique uses scene jitter as a cue for classification: when recapturing planar surfaces approximately parallel to the imaging plane, any added motion due to jitter will result in approximately uniform high-frequency 2D motion fields. The inter-frame motion trajectories are retrieved with feature tracking techniques, while local and global feature motion are decoupled through a 2-level wavelet decomposition. A normalised cross-correlation matrix is then populated with the similarities between the high-frequency components of the tracked features' trajectories. The correlation distribution is then compared with trained models for classification. Experiments with original and recaptured standard datasets show the validity of the proposed technique.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128386191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Improved seam carving for semantic video cod 改进了语义视频编码的接缝雕刻
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343415
M. Decombas, F. Dufaux, E. Renan, B. Pesquet-Popescu, F. Capman
{"title":"Improved seam carving for semantic video cod","authors":"M. Decombas, F. Dufaux, E. Renan, B. Pesquet-Popescu, F. Capman","doi":"10.1109/MMSP.2012.6343415","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343415","url":null,"abstract":"Traditional video codecs like H.264/AVC encode video sequences to minimize the Mean Squared Error (MSE)at a given bitrate. Seam carving is a content-aware resizing method. In this paper, we propose a semantic video compression scheme based on seam carving. Its principle is to suppress non salient parts of the video by seam carving. The reduced sequence is then encoded with H.264/AVC and the seams are represented and encoded with our proposed approach. The main idea is to encode the seams by regrouping them. Compared to our earlier work, the main contributions of this paper are: a new energy map with better temporal robustness, a new way to define groups of seams using k-median clustering, and an improved background synthesis. Experiments show that, compared to a traditional H.264/AVC encoding, we reach a bitrate saving between 10% and 24% % with the same quality of the salient objects.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121471831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Learning based screen image compression 基于学习的屏幕图像压缩
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343419
Huan Yang, Weisi Lin, Chenwei Deng
{"title":"Learning based screen image compression","authors":"Huan Yang, Weisi Lin, Chenwei Deng","doi":"10.1109/MMSP.2012.6343419","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343419","url":null,"abstract":"There are usually two components in computer screen images: textual and pictorial parts. The pictorial part can be compressed efficiently by classical coding approaches (e.g. JPEG, JPEG2000), while the compression of the textual part is still far away from being satisfactory for the reason that the textual content is usually of high-frequency. In this paper, a learning approach is used to construct a tailored dictionary for text representation. Based on the learned dictionary, a novel screen image compression algorithm is proposed through adopting different basis functions for the textual and pictorial components respectively. The screen images are firstly segmented into textual and pictorial parts. Then we employ traditional discrete cosine transformation (DCT) to facilitate the compression of pictorial part, while the learned dictionary is used to represent the textual part in screen images. Experimental results demonstrate the effectiveness of the proposed compression algorithm.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127649971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Block-based compressed sampling with non-linear coding for image transmission 基于块的非线性编码压缩采样图像传输
2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2012-11-12 DOI: 10.1109/MMSP.2012.6343416
B. Liu, Wei Qiao, Zixiang Xiong, G. Arce, J. Garcia-Frías
{"title":"Block-based compressed sampling with non-linear coding for image transmission","authors":"B. Liu, Wei Qiao, Zixiang Xiong, G. Arce, J. Garcia-Frías","doi":"10.1109/MMSP.2012.6343416","DOIUrl":"https://doi.org/10.1109/MMSP.2012.6343416","url":null,"abstract":"We propose a novel block-based image transmission system, which exploits the a prior information existing in the DCT domain of images and combines both linear and non-linear coding schemes accommodated to a block-based DCT domain compressed sampling method. An image is firstly divided into blocks and each block is separately sampled in DCT domain. Different coding schemes are used to transmit the samples based on their properties. With block-based strategy, each image block can be processed and transmitted separately, which reduces a lot of latency. Besides, an efficient system optimization algorithm is proposed by jointly optimizing the power allocation scheme and the transmission parameters to search for the maximum peak signal-to-noise ratio (PSNR) of the reconstructed image. Simulation results show that the proposed system provides a good performance with less latency.","PeriodicalId":325274,"journal":{"name":"2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114802921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信