2011 IEEE International Conference on Multimedia and Expo最新文献

筛选
英文 中文
Credit-title detection of video contents based on estimation of superimposed region using character density distribution 基于字符密度分布估计叠加区域的视频内容字幕检测
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012129
R. Mase, R. Oami, T. Nomura
{"title":"Credit-title detection of video contents based on estimation of superimposed region using character density distribution","authors":"R. Mase, R. Oami, T. Nomura","doi":"10.1109/ICME.2011.6012129","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012129","url":null,"abstract":"We propose a credit-title detection method of video contents based on estimation of superimposed region using character density distribution. Copyright information of video contents is manually extracted for the secondary use of those contents, and its cost is highly expensive. Therefore, automatic detection of credit titles that contain copyright information is highly demanded. However, accuracy of conventional methods is usually insufficient for this purpose. Our method first estimates credit-title-superimposed region based on character density distribution calculated in advance by using many video contents. Then, credit titles are detected in the estimated region. The experiment results show that proposed method improves both recall and precision rates compared to a conventional method. Furthermore, the processing time of the proposed method is less than half that of the conventional method for all contents.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123099597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HYBRID TCP/UDP video transport for H.264/AVC content delivery in burst loss networks 突发丢失网络中H.264/AVC内容传输的HYBRID TCP/UDP视频传输
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6011979
Timothy Porter, Xiaohong Peng
{"title":"HYBRID TCP/UDP video transport for H.264/AVC content delivery in burst loss networks","authors":"Timothy Porter, Xiaohong Peng","doi":"10.1109/ICME.2011.6011979","DOIUrl":"https://doi.org/10.1109/ICME.2011.6011979","url":null,"abstract":"In this paper we propose a hybrid TCP/UDP transport, specifically for H.264/AVC encoded video, as a compromise between the delay-prone TCP and the loss-prone UDP. When implementing the hybrid approach, we argue that the playback at the receiver often need not be 100% perfect, provided that a certain level of quality is assured. Reliable TCP is used to transmit and guarantee delivery of the most important packets. This allows use of additional features in the H.264/AVC standard which simultaneously provide an enhanced playback quality, in addition to a reduction in throughput. These benefits are demonstrated through experimental results using a test-bed to emulate the hybrid proposal. We compare the proposed system with other protection methods, such as FEC, and in one case show that for the same bandwidth overhead, FEC is unable to match the performance of the hybrid system in terms of playback quality. Furthermore, we measure the delay associated with our approach, and examine its potential for use as an alternative to the conventional methods of transporting video by either TCP or UDP alone.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"162 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124376606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Hypothesis comparison guided cross validation for unsupervised signer adaptation 假设比较引导的无监督手语自适应交叉验证
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012086
Yu Zhou, Xiaokang Yang, Weiyao Lin, Yi Xu, Long Xu
{"title":"Hypothesis comparison guided cross validation for unsupervised signer adaptation","authors":"Yu Zhou, Xiaokang Yang, Weiyao Lin, Yi Xu, Long Xu","doi":"10.1109/ICME.2011.6012086","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012086","url":null,"abstract":"Signer adaptation is important to sign language recognition systems in that a one-size-fits-all model set can not perform well on all kinds of signers. Supervised signer adaptation must utilize the labeled adaptation data that are collected explicitly. To skip the data collecting process in signer adaptation, we propose an unsupervised adaptation method called hypothesis comparison guided cross validation (HC-CV) algorithm. The algorithm not only addresses the problem of overlap between the data set to be labeled and the data set for adaptation, but also employs an additional hypothesis comparison step to decrease the noise rate of the adaptation data set. Experimental results show that the HC-CV adaptation algorithm is superior to the CV adaptation algorithm and the conventional self-teaching algorithm. Though the algorithm is proposed for signer adaptation, it can also be applied to speaker adaptation and writer adaptation straightforwardly.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123084854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatic transcription of piano music by sparse representation of magnitude spectra 用稀疏谱表示法实现钢琴音乐的自动转录
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012000
Cheng-Te Lee, Yi-Hsuan Yang, Homer H. Chen
{"title":"Automatic transcription of piano music by sparse representation of magnitude spectra","authors":"Cheng-Te Lee, Yi-Hsuan Yang, Homer H. Chen","doi":"10.1109/ICME.2011.6012000","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012000","url":null,"abstract":"Assuming that the waveforms of piano notes are pre-stored and that the magnitude spectrum of a piano signal segment can be represented as a linear combination of the magnitude spectra of the pre-stored piano waveforms, we formulate the automatic transcription of polyphonic piano music as a sparse representation problem. First, the note candidates of the piano signal segment are found by using heuristic rules. Then, the sparse representation problem is solved by l1-regularized minimization, followed by temporal smoothing the frame-level results based on hidden Markov models. Evaluation against three state-of-the-art systems using ten classical music recordings of a real piano is performed to show the performance improvement of the proposed system.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121792327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Efficient 3D region guarding for multimedia data processing 用于多媒体数据处理的高效三维区域保护
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6011953
Wuyi Yu, Maoqing Li, S. Iyengar, Xin Li
{"title":"Efficient 3D region guarding for multimedia data processing","authors":"Wuyi Yu, Maoqing Li, S. Iyengar, Xin Li","doi":"10.1109/ICME.2011.6011953","DOIUrl":"https://doi.org/10.1109/ICME.2011.6011953","url":null,"abstract":"With the advance of scanning devices, 3-d geometric models have been captured and widely used in animation, video, interactive virtual environment design nowadays. Their effective analysis, integration, and retrieval are important research topics in multimedia. This paper studies a geometric modeling problem called 3D region guarding. The 3D region guarding is a well known NP-hard problem; we present an efficient hierarchical integer linear programming (HILP) optimization algorithm to solve it on massive data sets. We show the effectiveness of our algorithm and briefly illustrate its applications in multimedia data processing and computer graphics such as shape analysis and retrieval, and morphing animation.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116604959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Hand tracking based on the combination of 2D and 3D model in gaze-directed video 基于2D和3D模型结合的注视视频手部跟踪
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012236
Li Sun, Guizhong Liu
{"title":"Hand tracking based on the combination of 2D and 3D model in gaze-directed video","authors":"Li Sun, Guizhong Liu","doi":"10.1109/ICME.2011.6012236","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012236","url":null,"abstract":"This paper investigates model based hand tracking in gaze-directed video which contains everyday manipulation activity of human in kitchen environment. The video is recorded by a gaze-directed camera, which can actively directs at the visual attention area from the person who wears the camera. Here we present a method based on the combination of 2D and 3D hand model, which can estimate the position of hand in image accurately and the pose of hand in 3D roughly. The method uses 2D model tracking result to initialize and predict 3D tracking, which saves the number of particles and makes it possible for local configuration adapting. To evaluate our result, we try our algorithm on several pieces of video both from normal camera and gaze-directed camera. The error ratio of the distance between the ground truth and tracking result is used as an objective measurement for evaluating our method. Trajectory of hand movement and results of projected model for every frame show that our method is effective and makes a good foundation for future recognition and analysis.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117209231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A novel scalable Deblocking Filter architecture for H.264/AVC and SVC video codecs 为H.264/AVC和SVC视频编解码器设计了一种新颖的可伸缩的去块滤波器架构
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012075
T. Cervero, A. Otero, S. López, E. D. L. Torre, G. Callicó, R. Sarmiento, T. Riesgo
{"title":"A novel scalable Deblocking Filter architecture for H.264/AVC and SVC video codecs","authors":"T. Cervero, A. Otero, S. López, E. D. L. Torre, G. Callicó, R. Sarmiento, T. Riesgo","doi":"10.1109/ICME.2011.6012075","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012075","url":null,"abstract":"A highly parallel and scalable Deblocking Filter (DF) hardware architecture for H.264/AVC and SVC video codecs is presented in this paper. The proposed architecture mainly consists on a coarse grain systolic array obtained by replicating a unique and homogeneous Functional Unit (FU), in which a whole Deblocking-Filter unit is implemented. The proposal is also based on a novel macroblock-level parallelization strategy of the filtering algorithm which improves the final performance by exploiting specific data dependences. This way communication overhead is reduced and a more intensive parallelism in comparison with the existing state-of-the-art solutions is obtained. Furthermore, the architecture is completely flexible, since the level of parallelism can be changed, according to the application requirements. The design has been implemented in a Virtex-5 FPGA, and it allows filtering 4CIF (704x576 pixels @30fps) video sequences in real-time at frequencies lower than 10.16 Mhz.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120904777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Using eye tracking technology to identify visual and verbal learners 使用眼动追踪技术来识别视觉和语言学习者
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012036
T. Mehigan, M. Barry, Aidan Kehoe, I. Pitt
{"title":"Using eye tracking technology to identify visual and verbal learners","authors":"T. Mehigan, M. Barry, Aidan Kehoe, I. Pitt","doi":"10.1109/ICME.2011.6012036","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012036","url":null,"abstract":"Learner style data is increasingly being incorporated into adaptive eLearning (electronic learning) systems for the development of personalized user models. This practice currently relies heavily on the prior completion of questionnaires by system users. Whilst potentially improving learning outcomes, the completion of questionnaires can be time consuming for users. Recent research indicates that it is possible to detect a user's preference on the Global / Sequential dimension of the FSLSM (Felder-Silverman Learner Style Model) through a user's mouse movement pattern, and other biometric technology including eye tracking and accelerometer technology. In this paper we discuss the potential of eye tracking technology for inference of Visual / Verbal learners. The paper will discuss the results of a study conducted to detect individual user style data based on the Visual / Verbal dimension of the FSLSM.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121305530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Adaptive maximum margin criterion for image classification 图像分类的自适应最大边距准则
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6011920
Jiwen Lu, Yap-Peng Tan
{"title":"Adaptive maximum margin criterion for image classification","authors":"Jiwen Lu, Yap-Peng Tan","doi":"10.1109/ICME.2011.6011920","DOIUrl":"https://doi.org/10.1109/ICME.2011.6011920","url":null,"abstract":"We propose in this paper a novel adaptive maximum margin criterion (AMMC) method for image classification. While a large number of discriminant analysis algorithms have been proposed in recent years, most of them consider an equal importance of each training sample and ignore the different contributions of these samples to learn the discriminative feature subspace for classification. Motivated by the fact that some training samples are more effectual in learning the low-dimensional feature space than other samples, we propose using different weights to characterize the different contributions of the training samples and incorporate such weighting information into the popular maximum margin criterion algorithm to devise the corresponding AMMC for image classification. Moreover, we extend the proposed MMC algorithm to the semi-supervised case, namely, semi-supervised adaptive maximum margin criterion (SAMMC), by making use of both labeled and unlabeled samples to further improve the classification performance. Experimental results are presented to demonstrate the efficacy of the proposed methods.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127218127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
TV as a human interface for Ambient Intelligence environments 电视作为环境智能环境的人机界面
2011 IEEE International Conference on Multimedia and Expo Pub Date : 2011-07-11 DOI: 10.1109/ICME.2011.6012186
Gorka Epelde, Xabier Valencia, J. Abascal, U. Díaz-Orueta, I. Zinnikus, Christian Husodo-Schulz
{"title":"TV as a human interface for Ambient Intelligence environments","authors":"Gorka Epelde, Xabier Valencia, J. Abascal, U. Díaz-Orueta, I. Zinnikus, Christian Husodo-Schulz","doi":"10.1109/ICME.2011.6012186","DOIUrl":"https://doi.org/10.1109/ICME.2011.6012186","url":null,"abstract":"One of the challenges that Ambient Intelligent (AmI) faces is the provision of a usable interaction concept to its users, especially for those with less technical background. In this paper, we describe a new approach to integrate interactive services provided by an AmI environment with the television set, which is one of the most used interaction client in the home environment. An implementation of this approach has been carried out as a multimodal/multipurpose natural human computer interface for elderly people, by creating adapted graphical user interfaces and navigation menus together with multimodal interaction (simplified TV remote control and voice interaction). In addition, this user interface can also be suited to other user groups. We have tested a prototype that adapts the videoconference and the information service with a group of 83 users. The results from the user tests show that the group found the prototype to be both satisfactory and efficient to use.","PeriodicalId":433997,"journal":{"name":"2011 IEEE International Conference on Multimedia and Expo","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127408754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信