2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)最新文献

筛选
英文 中文
Autonomous virtual humans and social robots in telepresence 远程呈现中的自主虚拟人和社交机器人
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958836
N. Magnenat-Thalmann, Zerrin Yumak, Aryel Beck
{"title":"Autonomous virtual humans and social robots in telepresence","authors":"N. Magnenat-Thalmann, Zerrin Yumak, Aryel Beck","doi":"10.1109/MMSP.2014.6958836","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958836","url":null,"abstract":"Telepresence refers to the possibility of feeling present in a remote location through the use of technology. This can be achieved by immersing a user to a place reconstructed in 3D. The reconstructed place can be captured from the real world or can be completely virtual. Another way to realize telepresence is by using robots and virtual avatars that act as proxies for real people. In case a human-mediated interaction is not needed or not possible, the virtual human and the robot can rely on artificial intelligence to act and interact autonomously. In this paper, these forms of telepresence are discussed, how they are related and different from each other and how autonomy takes place in telepresence. The paper concludes with an overview of the ongoing research on autonomous virtual humans and social robots conducted in the BeingThere centre.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129919739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Grabcut-based abandoned object detection 基于grabcut的废弃物体检测
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958806
K. Muchtar, Chih-Yang Lin, C. Yeh
{"title":"Grabcut-based abandoned object detection","authors":"K. Muchtar, Chih-Yang Lin, C. Yeh","doi":"10.1109/MMSP.2014.6958806","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958806","url":null,"abstract":"This paper presents a detection-based method to subtract abandoned object from a surveillance scene. Unlike tracking-based approaches that are commonly complicated and unreliable on a crowded scene, the proposed method employs background (BG) modelling and focus only on immobile objects. The main contribution of our work is to build abandoned object detection system which is robust and can resist interference (shadow, illumination changes and occlusion). In addition, we introduce the MRF model and shadow removal to our system. MRF is a promising way to model neighbours' information when labeling the pixel that is either set to background or abandoned object. It represents the correlation and dependency in a pixel and its neighbours. By incorporating the MRF model, as shown in the experimental part, our method can efficiently reduce the false alarm. To evaluate the system's robustness, several dataset including CAVIAR datasets and outdoor test cases are both tested in our experiments.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131077296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
QoE-driven performance analysis of cloud gaming services 基于qos的云游戏服务性能分析
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958835
Zi-Yi Wen, Hsu-Feng Hsiao
{"title":"QoE-driven performance analysis of cloud gaming services","authors":"Zi-Yi Wen, Hsu-Feng Hsiao","doi":"10.1109/MMSP.2014.6958835","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958835","url":null,"abstract":"With the popularity of cloud computing services and the endorsement from the video game industry, cloud gaming services have emerged promisingly. In a cloud gaming service, the contents of games can be delivered to the clients through either video streaming or file streaming. Due to the strict constraint on the end-to-end latency for real-time interaction in a game, there are still challenges in designing a successful cloud gaming system, which needs to deliver satisfying quality of experience to the customers. In this paper, the methodology for subjective and objective evaluation as well as the analysis of cloud gaming services was developed. The methodology is based on a nonintrusive approach, and therefore, it can be used on different kinds of cloud gaming systems. There are challenges in such objective measurements of important QoS factors, due to the fact that most of the commercial cloud gaming systems are proprietary and closed. In addition, satisfactory QoE is one of the crucial ingredients in the success of cloud gaming services. By combining subjective and objective evaluation results, cloud gaming system developers can infer possible results of QoE levels based on the measured QoS factors. It can also be used in an expert system for choosing the list of games that customers can appreciate at a given environment, as well as for deciding the upper bound of the number of users in a system.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115608697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Estimating spatial layout of rooms from RGB-D videos 从RGB-D视频中估计房间的空间布局
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958786
Anran Wang, Jiwen Lu, Jianfei Cai, G. Wang, Tat-Jen Cham
{"title":"Estimating spatial layout of rooms from RGB-D videos","authors":"Anran Wang, Jiwen Lu, Jianfei Cai, G. Wang, Tat-Jen Cham","doi":"10.1109/MMSP.2014.6958786","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958786","url":null,"abstract":"Spatial layout estimation of indoor rooms plays an important role in many visual analysis applications such as robotics and human-computer interaction. While many methods have been proposed for recovering spatial layout of rooms in recent years, their performance is still far from satisfactory due to high occlusion caused by the presence of objects that clutter the scene. In this paper, we propose a new approach to estimate the spatial layout of rooms from RGB-D videos. Unlike most existing methods which estimate the layout from still images, RGB-D videos provide more spatial-temporal and depth information, which are helpful to improve the estimation performance because more contextual information can be exploited in RGB-D videos. Given a RGB-D video, we first estimate the spatial layout of the scene in each single frame and compute the camera trajectory using the simultaneous localization and mapping (SLAM) algorithm. Then, the estimated spatial layouts of different frames are integrated to infer temporally consistent layouts of the room throughout the whole video. Our method is evaluated on the NYU RGB-D dataset, and the experimental results show the efficacy of the proposed approach.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126857351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
2D/3D AudioVisual content analysis & description 2D/3D视听内容分析和描述
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958837
I. Pitas, K. Papachristou, N. Nikolaidis, M. Liuni, L. Benaroya, G. Peeters, A. Röbel, A. Linnemann, Mohan Liu, S. Gerke
{"title":"2D/3D AudioVisual content analysis & description","authors":"I. Pitas, K. Papachristou, N. Nikolaidis, M. Liuni, L. Benaroya, G. Peeters, A. Röbel, A. Linnemann, Mohan Liu, S. Gerke","doi":"10.1109/MMSP.2014.6958837","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958837","url":null,"abstract":"In this paper, we propose a way of using the Audio-Visual Description Profile (AVDP) of the MPEG-7 standard for 2D or stereo video and multichannel audio content description. Our aim is to provide means of using AVDP in such a way, that 3D video and audio content can be correctly and consistently described. Since AVDP semantics do not include ways for dealing with 3D audiovisual content, a new semantic framework within AVDP is proposed and examples of using AVDP to describe the results of analysis algorithms on stereo video and multichannel audio content are presented.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123007725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Block-based compressive sensing of video using local sparsifying transform 基于局部稀疏化变换的视频分块压缩感知
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958826
Chien Van Trinh, V. Nguyen, B. Jeon
{"title":"Block-based compressive sensing of video using local sparsifying transform","authors":"Chien Van Trinh, V. Nguyen, B. Jeon","doi":"10.1109/MMSP.2014.6958826","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958826","url":null,"abstract":"Block-based compressive sensing is attractive for sensing natural images and video because it makes large-sized image/video tractable. However, its reconstruction performance is yet to be improved much. This paper proposes a new block-based compressive video sensing recovery scheme which can reconstruct video sequences with high quality. It generates initial key frames by incorporating the augmented Lagrangian total variation with a nonlocal means filter which is well known for being good at preserving edges and reducing noise. Additionally, local principal component analysis (PCA) transform is employed to enhance the detailed information. The non-key frames are initially predicted by their measurements and reconstructed key frames. Furthermore, regularization with PCA transform-aided side information iteratively seeks better reconstructed solution. Simulation results manifest effectiveness of the proposed scheme.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125302934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Soccer video summarization based on cinematography and motion analysis 基于电影摄影和运动分析的足球视频摘要
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958804
Ngoc Nguyen, A. Yoshitaka
{"title":"Soccer video summarization based on cinematography and motion analysis","authors":"Ngoc Nguyen, A. Yoshitaka","doi":"10.1109/MMSP.2014.6958804","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958804","url":null,"abstract":"Summarization of soccer videos has been widely studied due to its worldwide viewers and potential commercial applications. Most existing methods focus on searching for highlight events in soccer videos such as goals, penalty kicks and generating a summary as a list of such events. However, besides highlight events, scenes of intensive competition between players of two teams and emotional moments are also interesting. In this paper, we propose a soccer summarization system which is able to capture highlight events, scenes of intensive competition, and emotional moments. Based on the flow of soccer games, we organize a video summary as follows: first, scenes of intensive competition, second, what events happened, third, who were involved in the events, and finally how players or audience reacted to the events. With this structure, the generated summary is more complete and interesting because it provides both game play and emotional moments. Our system takes broadcast video as input, and divides it into multiple clips based on cinematographic features such as sport video production techniques, the transition of shots, and camera motions. Then, the system evaluates the interest level of each clip to generate a summary. Experimental results and subjective evaluation are carried out to evaluate the quality of the generated summary and the effectiveness of our proposed interest level measure.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131959582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Gaze direction estimation from static images 静态图像的凝视方向估计
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958803
Krystian Radlak, M. Kawulok, B. Smolka, Natalia Radlak
{"title":"Gaze direction estimation from static images","authors":"Krystian Radlak, M. Kawulok, B. Smolka, Natalia Radlak","doi":"10.1109/MMSP.2014.6958803","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958803","url":null,"abstract":"This study presents a novel multilevel algorithm for gaze direction recognition from static images. Proposed solution consists of three stages: (i) eye pupil localization using a multistage ellipse detector combined with a support vector machines verifier, (ii) eye bounding box localization calculated using a hybrid projection function and (iii) gaze direction classification using support vector machines and random forests. The proposed method has been tested on Eye-Chimera database with very promising results. Extensive tests show that eye bounding box localization allows us to achieve highly accurate results both in terms of eye location and gaze direction classification.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124604446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Shot type characterization in 2D and 3D video content 2D和3D视频内容的镜头类型表征
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958788
Ioannis Tsingalis, A. Tefas, N. Nikolaidis, I. Pitas
{"title":"Shot type characterization in 2D and 3D video content","authors":"Ioannis Tsingalis, A. Tefas, N. Nikolaidis, I. Pitas","doi":"10.1109/MMSP.2014.6958788","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958788","url":null,"abstract":"Due to the enormous increase of video and image content on the web in the last decades, automatic video annotation became a necessity. The successful annotation of video and image content facilitate a successful indexing and retrieval in search databases. In this work we study a variety of possible shot type characterizations that can be assigned in a single video frame or still image. Possible ways to propagate these characterizations to a video segment (or to an entire shot) are also discussed. A method for the detection of Over-the-Shoulder shots in 3D (stereo) video is also proposed.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114330985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A hybrid approach to animating the murals with Dunhuang style 采用混合的方法使壁画具有敦煌风格
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP) Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958789
Bingwen Jin, Linglong Feng, Gang Liu, Huaqing Luo, Wei-dong Geng
{"title":"A hybrid approach to animating the murals with Dunhuang style","authors":"Bingwen Jin, Linglong Feng, Gang Liu, Huaqing Luo, Wei-dong Geng","doi":"10.1109/MMSP.2014.6958789","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958789","url":null,"abstract":"In order to animate the valuable murals of Dunhuang Mogao Grottoes, we propose a hybrid approach to creating the animation with the artistic style in the murals. Its key point is the fusion of 2D and 3D animation assets, for which a hybrid model is constructed from a 2.5D model, a 3D model, and registration information. The 2.5D model, created from 2D multi-view drawings, is composed of 2.5D strokes. For each 2.5D stroke, we let the user draw corresponding strokes on the surface of the 3D model in multiple views. Then the method automatically generates registration information, which enables 3D animation assets to animate the 2.5D model. At last, the animated line drawings are produced from 2.5D and 3D models respectively and blended under the control of per-stroke weights. The user can manually modify the weights to get the desired animation style.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134635608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信