2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)最新文献

筛选
英文 中文
Similarity learning for template-based visual tracking 基于模板的视觉跟踪相似性学习
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890723
Xiuzhuang Zhou, Lu Kou, Hui Ding, Xiaoyan Fu, Yuanyuan Shang
{"title":"Similarity learning for template-based visual tracking","authors":"Xiuzhuang Zhou, Lu Kou, Hui Ding, Xiaoyan Fu, Yuanyuan Shang","doi":"10.1109/ICMEW.2014.6890723","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890723","url":null,"abstract":"Most existing discriminative models for visual tracking are often formulated as supervised learning of a binary classification function, whose continuous output is then cast into a specific tracking framework as the confidence of the visual target. We argue that this might be less accurate since the classifier is learned for making binary decision, rather than predicting the similarity score between the candidate image patches and the true target. On the other hand, a generative tracker aims at learning a compact object representation for updating of the visual appearance. This, however, ignores the useful information from background regions surroundding the visual target, and hence might not well separate the visual target from the background distracters. We propose in this work a visual tracking scheme, in which a similarity function is explicitly learned in a generative tracking framework to significantly alleviate the drifting problem suffered by many existing trackers. Experimental results on various challenging human sequences, involving significant appearance changes, severe occlusions, and cluttered backgrounds, demonstrate the effectiveness of our approach compared to the state-of-the-art alternatives.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126367703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Assessing quality of experience for adaptive HTTP video streaming 评估自适应HTTP视频流的体验质量
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890604
Jingteng Xue, Dong-Qing Zhang, H. H. Yu, Chang Wen Chen
{"title":"Assessing quality of experience for adaptive HTTP video streaming","authors":"Jingteng Xue, Dong-Qing Zhang, H. H. Yu, Chang Wen Chen","doi":"10.1109/ICMEW.2014.6890604","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890604","url":null,"abstract":"In this paper, we present a novel subjective quality model for online adaptive movie streaming service. The proposed model considers the Quality of Experience (QoE) of streaming video viewing as a cumulative evaluation process of consecutive segments that compose a story line. Under bandwidth constraint, streaming client may select lower-quality segment, pause playback for re-buffering, or both. The momentary QoE loss at these events are weighted by the content suspension level. Meanwhile such experience penalty remains to influence user's opinion for the rest of the service. If the picture becomes too noisy or the interruptions occur consistently, the user may stop watching. The proposed scheme includes two parts. First, a parametric model estimates the quality loss of a interfered segment based on its network-level packet characteristics. Second, a cumulative function integrates the impact of streaming events. Both steps demand minimum computation and can be updated in real time. A subjective test to train and validate the proposed parametric model is designed and performed. This model is fundamentally different from all existing QoE assessment schemes in that temporally cumulative viewing experience of the users, instead of simple global statistics, is evaluated.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124850968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Modeling a realistic 3D physiological tongue for visual speech synthesis 模拟一个真实的三维生理舌头,用于视觉语音合成
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890595
Chen Jiang, Changwei Luo, Jun Yu, Rui Li, Zengfu Wang
{"title":"Modeling a realistic 3D physiological tongue for visual speech synthesis","authors":"Chen Jiang, Changwei Luo, Jun Yu, Rui Li, Zengfu Wang","doi":"10.1109/ICMEW.2014.6890595","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890595","url":null,"abstract":"We built a 3D anatomically and biomechanically accurate physiological tongue model for use in visual speech synthesis. For the anatomical modeling part, the tongue and its muscles are constructed based on accurate medical data. Due to their complexity, muscles geometry and fiber arrangement are specified by a proposed interactive muscle marking method. For the biomechanical modeling part, a nonlinear, quasi-incompressible, hyperelastic constitutive model is applied for describing the tongue tissues. Particularly, tongue muscles are additionally endowed with an anisotropic constitutive model, which reflects the active and passive mechanical behavior of muscle fibers. The dynamic deformation of tongue is simulated based on finite element method (FEM). Simulation results of tongue movements subjected to certain muscle activations are presented and validated with experimental data. This tongue model can be applied in many areas, like media art, education, entertainment.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121540443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Reducing structure of deep Convolutional Neural Networks for Huawei Accurate and Fast Mobile Video Annotation Challenge 面向华为移动视频准确快速标注挑战的深度卷积神经网络降维结构
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890608
Yunlong Bian, Yuan Dong, Hongliang Bai, Bo Liu, Kai Wang, Yinan Liu
{"title":"Reducing structure of deep Convolutional Neural Networks for Huawei Accurate and Fast Mobile Video Annotation Challenge","authors":"Yunlong Bian, Yuan Dong, Hongliang Bai, Bo Liu, Kai Wang, Yinan Liu","doi":"10.1109/ICMEW.2014.6890608","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890608","url":null,"abstract":"Big structure of deep Convolutional Neural Networks (CNN) has staggeringly impressive improvement in the Imagenet Large Scale Visual Recognition Challenge (ILSVRC) 2012 and 2013. But only tens of classes are required to be trained in the most real applications. After the deep CNNs are trained in the ILSVRC dataset, efficiently transferring the big and deep structure to a new dataset is a tough problem. In this paper, three algorithms are proposed to implement the transfer, namely fine-tunning of the big structure, normalized Google distance and Wordnet lexical semantic similarity. After experiments are conducted in the Huawei accurate and fast Mobile Video Annotation Challenge (MoVAC) dataset, the fine-tuning algorithm has achieved the best performance in the accuracy and training time.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122125541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Non-intrusive quality assessment for enhanced speech signals based on spectro-temporal features 基于频谱时间特征的增强语音信号非侵入性质量评估
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890561
Qiaohong Li, Yuming Fang, Weisi Lin, D. Thalmann
{"title":"Non-intrusive quality assessment for enhanced speech signals based on spectro-temporal features","authors":"Qiaohong Li, Yuming Fang, Weisi Lin, D. Thalmann","doi":"10.1109/ICMEW.2014.6890561","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890561","url":null,"abstract":"We propose to learn a non-intrusive quality assessment metric for enhanced speech signals. High-dimension spectro-temporal features are extracted by the Gabor filter bank for speech signals. To reduce the high-dimension features, we use PCA (Principal Component Analysis) to process these features. After obtaining the feature vector from audio signals, Support Vector Regression (SVR) is used to learn the metric for quality evaluation of enhanced speech signals. Experimental results on NOIZEUS dataset demonstrate that proposed non-intrusive quality assessment metric by using spectro-temporal features can obtain better performance for enhanced speech signals.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123195697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Feature extraction of complex ocean flow field using the helmholtz-hodge decomposition 基于helmholtz-hodge分解的复杂海洋流场特征提取
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890546
Huan Wang, Junhui Deng
{"title":"Feature extraction of complex ocean flow field using the helmholtz-hodge decomposition","authors":"Huan Wang, Junhui Deng","doi":"10.1109/ICMEW.2014.6890546","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890546","url":null,"abstract":"Flow visualization is an important research field in scientific visualization and feature detection is the one of the core problems. This paper presents a novel approach for complex ocean flow visualization and analysis by applying the helmholtz-hodge decomposition theory to the feature extraction problem. We give an efficient implementation on regular grids by solving a large-scale sparse group of linear equations. And to accelerate the computation process, we have used the GMRES parallel library. By making full use of the anti-noise property of the decomposition results, well-designed algorithms are used to identify the features such as critical points and vortices. To illustrate the ability of our techniques, experiments on realistic datasets are conducted. Experimental results demonstrate that our methods are helpful to deeply understand the flow field.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123301665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Comparative analysis of online 2-D and 3-D video sharing 在线2-D和3-D视频共享的对比分析
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890574
Soobeom Jang, Jong-Seok Lee
{"title":"Comparative analysis of online 2-D and 3-D video sharing","authors":"Soobeom Jang, Jong-Seok Lee","doi":"10.1109/ICMEW.2014.6890574","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890574","url":null,"abstract":"In recent years, online video sharing services have been growing rapidly. Nowadays, online video sharing becomes one of the most popular ways to produce and consume media. Moreover, with growing 3-D technologies, the interest to 3-D technologies has increased and it becomes easier to produce, share, and consume 3-D media on Internet. This paper investigates 3-D media production and consumption patterns in comparison to 2-D media to get insight to the online 3-D media ecosystem by analyzing metadata and statistics of 10,597,268 2-D videos and 50,972 3-D videos in YouTube. It is shown that the general production and consumption patterns of 2-D and 3-D videos are similar, but they have differences in uploading patterns, video length, category, and user participations. From these differences, we identify distinguished features of online 3-D video sharing. Our findings will serve as valuable information in technical and commercial aspects of contents sharing services.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"330 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123412373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Ontology driven contextual tagging of multimedia data 本体驱动的多媒体数据上下文标记
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890703
N. Pahal, S. Chaudhury, Brejesh Lall
{"title":"Ontology driven contextual tagging of multimedia data","authors":"N. Pahal, S. Chaudhury, Brejesh Lall","doi":"10.1109/ICMEW.2014.6890703","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890703","url":null,"abstract":"To exhibit multi-modal information and to facilitate people in finding multimedia resources, tagging plays a significant role. Various public events like protests and demonstrations are always consequences of break out of some public outrage resulting from prolonged exploitation and harassment. This outrage can be seen in news footage, blogs, text news and other web data. So, aggregating this variety of data from heterogeneous sources is a prerequisite step for tagging multimedia data with appropriate content. Since content has no meaning without a context, a video should be tagged with its relevant context and content information to assist user in multimedia retrieval. This paper proposes a model for tagging of multimedia data on the basis of contextual meaning. Since context is knowledge based, it has to be guided and learned by ontology which will help fragmented information to be represented in a more meaningful way. Our tagging approach is novel and has practical applicability in the sense that whenever a new video is uploaded on some media sharing site, the context and content information gets attached automatically to a video. Thus, providing relatively complete information associated with the video.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130432478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
[Demo paper] learning to beautify facial image 【演示论文】学习美化面部形象
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890625
Chengze Li, Xiaoyun Yuan, Juyong Zhang, Xiaoxin Lv, Lu Fang, O. Au
{"title":"[Demo paper] learning to beautify facial image","authors":"Chengze Li, Xiaoyun Yuan, Juyong Zhang, Xiaoxin Lv, Lu Fang, O. Au","doi":"10.1109/ICMEW.2014.6890625","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890625","url":null,"abstract":"In this demo, we demonstrate a data-driven facial image beautification system that learns how to beautify portraits from facial image database, and enhances the facial texture of arbitrary portraits automatically by modifying its pigment distribution and correcting its color. Specifically, as human skin can be treated as a turbid medium with multilayered structure, we decompose facial image into melanin and hemoglobin layers. With the extracted attractiveness features, a data-driven qualitative beautify model serves for the guidance of beautification through optimizing hemoglobin and melanin layers. Our beautification operations are conducted in completely automatic and time-efficient way, leading to customized realistic beautified portraits that follows users' preferences.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133503353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Influence of social media on performance of movies 社交媒体对电影表现的影响
2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW) Pub Date : 2014-07-14 DOI: 10.1109/ICMEW.2014.6890664
Shruti, S. Roy, W. Zeng
{"title":"Influence of social media on performance of movies","authors":"Shruti, S. Roy, W. Zeng","doi":"10.1109/ICMEW.2014.6890664","DOIUrl":"https://doi.org/10.1109/ICMEW.2014.6890664","url":null,"abstract":"The performance of movies in terms of revenue depends on many factors such as its production studio, genre, script quality, pre-release promotion etc, - all of which are traditionally used to estimate their potential success at the box office. Recently however, the “Wisdom of Crowd” and social media have been acknowledged as a strong signal in understanding consumer behavior towards media. In this paper, we discuss the impact of socially generated meta-data mined from the social multimedia sites and study the influence of social media on box- office performance and profitability of movies. Our result shows that various social media signals have varying yet significant impact in predicting the performance of movies. Our research also reveals that popularity of actor captured by follower count on Twitter is most relevant to the success of movie at theaters, and Facebook `like' signal has noise which impedes its analytical credibility.","PeriodicalId":178700,"journal":{"name":"2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129421406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信