2011 IEEE International Symposium on Multimedia最新文献

筛选
英文 中文
A Feasibility Study of Collaborative Stream Routing in Peer-to-Peer Multiparty Video Conferencing 协同流路由在点对点多方视频会议中的可行性研究
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.45
Han Zhao, D. Smilkov, P. Dettori, J. Nogima, F. Schaffa, P. Westerink, C. Wu
{"title":"A Feasibility Study of Collaborative Stream Routing in Peer-to-Peer Multiparty Video Conferencing","authors":"Han Zhao, D. Smilkov, P. Dettori, J. Nogima, F. Schaffa, P. Westerink, C. Wu","doi":"10.1109/ISM.2011.45","DOIUrl":"https://doi.org/10.1109/ISM.2011.45","url":null,"abstract":"Video transmission in multiparty video conferencing is challenging due to the demanding bandwidth usage and stringent latency requirement. In this paper, we systematically analyze the problem of collaborative stream routing using one-hop forwarding assistance in a bandwidth constraint environment. We model the problem as a multi-source degree-constrained multicast tree construction problem, and investigate heuristic algorithms to construct bandwidth-feasible shared multicast trees. The contribution of this work is primarily two-fold: (1) we study the solution space of finding a feasible bandwidth configuration for stream routing in a peer-to-peer (P2P) setting, and propose two heuristic algorithms that can quickly produce a bandwidth-feasible solution, making them suitable for large-scale conference sessions, (2) we conduct an empirical study using a realistic dataset and show the effectiveness of our heuristic algorithms. Various QoS metrics are taken into account to evaluate the performance of our algorithms. Finally, we discuss open issues for further exploration. The feasibility study presented in this paper will shed light on the design and implementation of practical P2P multiparty video conferencing applications.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132133956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Passive Forensics Method to Detect Tampering for Double JPEG Compression Image 双JPEG压缩图像篡改检测的被动取证方法
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.37
Zhenli Liu, Xiaofeng Wang, Jing Chen
{"title":"Passive Forensics Method to Detect Tampering for Double JPEG Compression Image","authors":"Zhenli Liu, Xiaofeng Wang, Jing Chen","doi":"10.1109/ISM.2011.37","DOIUrl":"https://doi.org/10.1109/ISM.2011.37","url":null,"abstract":"A passive forensics method to detect tampering for double JPEG compression image is proposed. In the proposed method, inconsistency of quality factors is used to detect double JPEG compression, and then a passive forensics approach to detect tampering and locate tampered area for tampered JPEG images is proposed. Comparing with existing methods, the main advantages of the proposed method are as follows: (1) It can detect rotation, scaling and tampering in small area. (2) It has a high computing efficiency.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116608664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multimodal Temporal Panorama for Moving Vehicle Detection and Reconstruction 多模态时间全景移动车辆检测与重建
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.101
Tao Wang, Zhigang Zhu, Clark N. Taylor
{"title":"Multimodal Temporal Panorama for Moving Vehicle Detection and Reconstruction","authors":"Tao Wang, Zhigang Zhu, Clark N. Taylor","doi":"10.1109/ISM.2011.101","DOIUrl":"https://doi.org/10.1109/ISM.2011.101","url":null,"abstract":"In this work, we present a multimodal temporal panorama (MTP) representation that synchronizes visual, motion, and acoustic signatures of moving vehicles in the time axis. The MTP representation includes two layers: a synopsis layer and a snapshot layer. The temporal synopsis consists of 1) a panoramic view image (PVI) to represent vehicles' presence, which is constructed from 1D vertical detecting lines of a selected column location of all video frames, 2) an epipolar plane image (EPI) to characterize their motion (speeds and directions), generated from 1D horizontal scanning lines along the vehicles' moving paths, and 3) an audio wave scroll for visualizing moving vehicles' acoustic signatures. The MTP synopsis not only synchronizes all the three modalities (visual, motion and acoustic) of the vehicles, but also provides information that can perform automatic detection tasks including moving vehicle visual detection, motion estimation, and acoustic signature retrieval. Then in the snapshot layer, the occlusion-free, motion-blur-free, and view-invariant reconstruction of each vehicle (with both shape and motion information) and its acoustic signatures (e.g. spectrogram) are embedded. The MTP provides a very effective approach to (semi-)automatically labeling the multimodal data of uncontrolled traffic scenes in real time for further vehicle classification, check-point inspection and traffic analysis. The concept of MTP may not be only limited to visual, motion and audio modalities, it could also be applicable to other sensing modalities that can obtain data in the temporal domain.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123312149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Hybrid Video Compression Using Selective Keyframe Identification and Patch-Based Super-Resolution 使用选择性关键帧识别和基于补丁的超分辨率混合视频压缩
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.25
J. Glaister, Calvin Chan, M. Frankovich, Adrian Tang, A. Wong
{"title":"Hybrid Video Compression Using Selective Keyframe Identification and Patch-Based Super-Resolution","authors":"J. Glaister, Calvin Chan, M. Frankovich, Adrian Tang, A. Wong","doi":"10.1109/ISM.2011.25","DOIUrl":"https://doi.org/10.1109/ISM.2011.25","url":null,"abstract":"This paper details a novel video compression pipeline using selective key frame identification to encode video and patch-based super-resolution to decode for playback. Selective key frame identification uses shot boundary detection and frame differencing methods to identify representative frames which are subsequently kept in high resolution within the compressed container. All other non-key frames are downscaled for compression purposes. Patch-based super-resolution finds similar patches between an up scaled non-key frame and the associated, high-resolution key frame to regain lost detail via a super-resolution process. The algorithm was integrated into the H.264 video compression pipeline tested on web cam, cartoon and live-action video for both streaming and storage purposes. Experimental results show that the proposed hybrid video compression pipeline successfully achieved higher compression ratios than standard H.264, while achieving superior video quality than low resolution H.264 at similar compression ratios.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114579023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An Adaptive Approach for Authoring Interactivity for Rich Multimedia Content 富多媒体内容交互性创作的自适应方法
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.39
M. Palviainen, S. Dutton
{"title":"An Adaptive Approach for Authoring Interactivity for Rich Multimedia Content","authors":"M. Palviainen, S. Dutton","doi":"10.1109/ISM.2011.39","DOIUrl":"https://doi.org/10.1109/ISM.2011.39","url":null,"abstract":"This paper describes an adaptive content authoring approach, the LIMO Authoring Tool to support the usage of the approach, and two examples in which an editor is adapted for presentation skeletons. The adapted editor aids the users as they create content to be attached to the presentation skeleton that specifies a ready-made baseline (e.g. skeleton, layout, and code libraries) for the presentation. The adapted editor does not just facilitate content creation but can also reduce errors and provide more robust, error free content.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"722 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122996037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Clustering-based Algorithm for Predicting File Size and Structural Similarity of Transcoded JPEG Images 基于聚类的高效JPEG转码图像文件大小和结构相似性预测算法
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.30
S. Pigeon, S. Coulombe
{"title":"Efficient Clustering-based Algorithm for Predicting File Size and Structural Similarity of Transcoded JPEG Images","authors":"S. Pigeon, S. Coulombe","doi":"10.1109/ISM.2011.30","DOIUrl":"https://doi.org/10.1109/ISM.2011.30","url":null,"abstract":"The problem of adapting JPEG images to satisfy constraints such as file size and resolution arises in a number of applications, from universal media access to multimedia messaging services. Visually optimized adaptation, however, commands a non-negligible computational cost which we aim to minimize using predictors. In previous works, we presented predictors and systems to achieve low-cost near-optimal adaptation of JPEG images. In this work, we propose a new approach to file size and quality prediction resulting from the Transco ding of a JPEG image subject to changes in quality factor and resolution. We show that the new predictor significantly outperforms the previously proposed solutions in accuracy.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127924400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Improved Multi-Rate Video Encoding 改进的多速率视频编码
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.53
Dag Haavi Finstad, H. Stensland, H. Espeland, P. Halvorsen
{"title":"Improved Multi-Rate Video Encoding","authors":"Dag Haavi Finstad, H. Stensland, H. Espeland, P. Halvorsen","doi":"10.1109/ISM.2011.53","DOIUrl":"https://doi.org/10.1109/ISM.2011.53","url":null,"abstract":"Adaptive HTTP streaming is frequently used for both live and on-Demand video delivery over the Internet. Adaptive ness is often achieved by encoding the video stream in multiple qualities (and thus bit rates), and then transparently switching between the qualities according to the bandwidth fluctuations and the amount of resources available for decoding the video content on the end device. For this kind of video delivery over the Internet, H.264 is currently the most used codec, but VP8 is an emerging open-source codec expected to compete with H.264 in the streaming scenario. The challenge is that, when encoding video for adaptive video streaming, both VP8 and H.264 run once for each quality layer, i.e., consuming both time and resources, especially important in a live video delivery scenario. In this paper, we address the resource consumption issues by proposing a method for reusing redundant steps in a video encoder, emitting multiple outputs with varying bit rates and qualities. It shares and reuses the computational heavy analysis step, notably macro-block mode decision, intra prediction and inter prediction between the instances, and outputs video in several rates. The method has been implemented in the VP8 reference encoder, and experimental results show that we can encode the different quality layers at the same rates and qualities compared to the VP8 reference encoder, while reducing the encoding time significantly.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132760508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Skin Region Extraction and Person-Independent Deformable Face Templates for Fast Video Indexing 用于快速视频索引的皮肤区域提取和独立于人的可变形人脸模板
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.75
S. Clippingdale, Mahito Fujii
{"title":"Skin Region Extraction and Person-Independent Deformable Face Templates for Fast Video Indexing","authors":"S. Clippingdale, Mahito Fujii","doi":"10.1109/ISM.2011.75","DOIUrl":"https://doi.org/10.1109/ISM.2011.75","url":null,"abstract":"We describe a face tracking and recognition system for video and multimedia indexing that handles face regions at variable face poses (left-right and up-down), and deformations due to facial expressions and speech, by employing person-independent deformable templates at multiple poses on the view-sphere. An earlier version of the system handled variable poses (left-right only) by employing person-specific templates registered for each target individual at multiple poses. The new system speeds up processing by (i) extracting and restricting attention to skin-color regions, (ii) performing recognition using person-specific templates at near-frontal poses only, and (iii) tracking at non-frontal poses using the person-independent templates. Registration is also simplified, since multiple views of each target individual are no longer required, at the cost of a loss of recognition functionality at poses far from frontal (the system instead \"remembers\" the identity of each individual from near-frontal matches and tracks between them). We describe the skin region extraction process and the process by which the person-independent templates are constructed off-line from \"bootstrap\" face images of multiple non-target individuals, and we present experimental results showing the system in operation. Finally we discuss remaining issues in the practical application of the system to video and multimedia archive indexing.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133339278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Utilization of Co-occurrence Relationships between Semantic Concepts in Re-ranking for Information Retrieval 语义概念共现关系在信息检索重排序中的应用
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.18
Chao Chen, Lin Lin, M. Shyu
{"title":"Utilization of Co-occurrence Relationships between Semantic Concepts in Re-ranking for Information Retrieval","authors":"Chao Chen, Lin Lin, M. Shyu","doi":"10.1109/ISM.2011.18","DOIUrl":"https://doi.org/10.1109/ISM.2011.18","url":null,"abstract":"Semantic information retrieval is a popular research topic in the multimedia area. The goal of the retrieval is to provide the end users with as relevant results as possible. Many research efforts have been done to build ranking models for different semantic concepts (or classes). While some of them have been proven to be effective, others are still far from satisfactory. Our observation that certain target semantic concepts have high co-occurrence relationships with those easy-to-retrieve semantic concepts (or called reference semantics) has motivated us to utilize such co-occurrence relationships between semantic concepts in information retrieval and re-ranking. In this paper, we propose a novel semantic retrieval and re-ranking framework that takes advantage of the co-occurrence relationships between a target semantic concept and a reference semantic concept to re-rank the retrieved results. The proposed framework discretizes the training data into a set of feature-value pairs and employs Multiple Correspondence Analysis (MCA) to capture the correlation in terms of the impact weight between feature-value pairs and the positive-positive class in which the data instances belong to both the target semantic concept and the reference semantic concept. A combination of all these impact weights is utilized to re-rank the retrieved results for the target semantic concept. Comparative experiments are designed and evaluated on TRECVID 2005 and TRECVID 2010 video collections with public-available ranking scores. Experimental results on different retrieval scales demonstrate that our proposed framework can enhance the retrieval results for the target semantic concepts in terms of average precision, and the improvements for some semantic concepts are promising.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131839894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Layered Approach for Fast Multi-view Stereo Panorama Generation 一种快速多视角立体全景生成的分层方法
2011 IEEE International Symposium on Multimedia Pub Date : 2011-12-05 DOI: 10.1109/ISM.2011.104
E. Molina, Zhigang Zhu, Clark N. Taylor
{"title":"A Layered Approach for Fast Multi-view Stereo Panorama Generation","authors":"E. Molina, Zhigang Zhu, Clark N. Taylor","doi":"10.1109/ISM.2011.104","DOIUrl":"https://doi.org/10.1109/ISM.2011.104","url":null,"abstract":"In this paper we propose a fast method for constructing multi-view stereo panoramas using a layering approach. Constructing panoramas requires accurate camera pose estimation and will often require an image blending or interpolation method to generate seamless results. We use a registration error correction method that provides globally corrected and fast results for paths that create cycles such as circular paths, back and forth straight sweeps, and even a single sweep. Then we apply our layering approach to generate multi-view stereo panoramas quickly for time sensitive applications that require immediate results and 3D perception.","PeriodicalId":339410,"journal":{"name":"2011 IEEE International Symposium on Multimedia","volume":"11 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132207038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信