2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)最新文献

筛选
英文 中文
Hello cleveland! Linked data publication of live music archives 克利夫兰你好!现场音乐档案的关联数据出版
S. Bechhofer, Kevin R. Page, D. D. Roure
{"title":"Hello cleveland! Linked data publication of live music archives","authors":"S. Bechhofer, Kevin R. Page, D. D. Roure","doi":"10.1109/WIAMIS.2013.6616155","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616155","url":null,"abstract":"We describe the publication of a linked data set exposing meta-data from the Internet Archive Live Music Archive. The dataset contains over 17,000,000 triples describing 100,000 performances by 4,000 artists. Links to other existing musical and geographical resources facilitate query of the collection along a number of axes. We describe both the methods used to annotate and layer the metadata - with a focus on considering the patterns used to represent mappings - and the role that views constructed from such a Linked Data set can play to bring together multidisciplinary multimedia analysis techniques.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121109326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Sample specific late fusion for saliency detection 样品特异性晚期融合显著性检测
Jie Sun, Congyan Lang, Songhe Feng
{"title":"Sample specific late fusion for saliency detection","authors":"Jie Sun, Congyan Lang, Songhe Feng","doi":"10.1109/WIAMIS.2013.6616133","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616133","url":null,"abstract":"Typically, the saliency map of an image is usually inferred by only using the information within this image. While efficient, such single-image-based methods may fail to obtain reliable results, because the information within a single image may be insufficient for defining saliency. In this paper, we propose a novel idea of learning with labeled images and adopt a new paradigm called sample specific late fusion (SSLF). To effectively explore the visual neighborhood information, we propose a semi-supervised learning technique for learning robust sample specific fusion parameters for multiply response maps of generic bottom-up saliency detectors. Different from previous methods, the proposed SSLF method integrates both middle-level image representation and unlabeled data information through an effective graph regularization framework. Extensive experiments have clearly validated its superiority over other state-of-the-art methods.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122342881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Densely sampled local visual features on 3D mesh for retrieval 在三维网格上密集采样局部视觉特征进行检索
Yuya Ohishi, Ryutarou Ohbuchi
{"title":"Densely sampled local visual features on 3D mesh for retrieval","authors":"Yuya Ohishi, Ryutarou Ohbuchi","doi":"10.1109/wiamis.2013.6616166","DOIUrl":"https://doi.org/10.1109/wiamis.2013.6616166","url":null,"abstract":"The Local Depth-SIFT (LD-SIFT) algorithm by Darom, et al. [2] captures 3D geometrical features locally at interest points detected on a densely-sampled, manifold mesh representation of the 3D shape. The LD-SIFT has shown good retrieval accuracy for 3D models defined as densely sampled manifold mesh. However, it has two shortcomings. The LD-SIFT requires the input mesh to be densely and evenly sampled. Furthermore, the LD-SIFT can't handle 3D models defined as a set of multiple connected components or a polygon soup. This paper proposes two extensions to the LD-SIFT to alleviate these weaknesses. First extension shuns interest point detection, and employs dense sampling on the mesh. Second extension employs remeshing by dense sample points followed by interest point detection a la LD-SIFT Experiments using three different benchmark databases showed that the proposed algorithms significantly outperform the LD-SIFT in terms of retrieval accuracy.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122347718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
JPEG backward compatible format for 3D content representation 用于3D内容表示的JPEG向后兼容格式
Philippe Hanhart, Pavel Korshunov, M. Rerábek, T. Ebrahimi
{"title":"JPEG backward compatible format for 3D content representation","authors":"Philippe Hanhart, Pavel Korshunov, M. Rerábek, T. Ebrahimi","doi":"10.1109/WIAMIS.2013.6616156","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616156","url":null,"abstract":"Different formats and compression algorithms have been proposed for 3D video content, but 3D images are still mostly represented as a stereo pair only. However, for enhanced 3D rendering capabilities, such as depth perception adjustment or display size adaptation, additional depth data is necessary. To facilitate the standardization process of a common 3D format, backward compatibility with legacy technologies is nec-essary. In this paper, we propose to extend the JPEG file for-mat, as the most popular image format, in a backward compatible manner to represent a stereo pair and additional depth data. We propose an architecture to achieve such backward compatibility with JPEG. The coding efficiency of a simple implementation of the proposed architecture is compared to the state of the art stereoscopic 3D image compression and storage formats.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127237742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A concise survey for 3D reconstruction of building façades 建筑立面三维重建简论
Patrycia Klavdianos, Qianni Zhang, E. Izquierdo
{"title":"A concise survey for 3D reconstruction of building façades","authors":"Patrycia Klavdianos, Qianni Zhang, E. Izquierdo","doi":"10.1109/WIAMIS.2013.6616170","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616170","url":null,"abstract":"3D facade modeling consists of representing architectural elements of a building in detail so that both geometry and appearance are generated in a photo-realistic 3D scene. In this article, we review four main techniques used to solve this problem: photogrammetry, sparse and dense reconstruction based on SfM (Structure from Motion) and procedural modeling. We provide a comparison of several methods in these categories by considering their main advantages and limitations.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116941893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-time guitar string detection for music education software 实时吉他弦检测音乐教育软件
C. Dittmar, Andreas Mannchen, J. Abeßer
{"title":"Real-time guitar string detection for music education software","authors":"C. Dittmar, Andreas Mannchen, J. Abeßer","doi":"10.1109/WIAMIS.2013.6616120","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616120","url":null,"abstract":"In this paper, we present a novel approach to real-time detection of the string number and fretboard position from polyphonic guitar recordings. Our goal is to assess, if a music student is correctly performing guitar exercises presented via music education software or a remote guitar teacher. We combine a state-of-the art approach for multi-pitch detection with a subsequent audio feature extraction and classification stage. Performance of the proposed system is evaluated with manually annotated chords recorded using different guitars.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128945826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Describing audio production workflows on the semantic web 描述语义网络上的音频制作工作流
György Fazekas, M. Sandler
{"title":"Describing audio production workflows on the semantic web","authors":"György Fazekas, M. Sandler","doi":"10.1109/WIAMIS.2013.6616135","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616135","url":null,"abstract":"Modern environments for creating, editing or managing multimedia content involve increasingly complex tools and components. These tools are typically used in multi-aspect workflows exhibiting creative, procedural and computational properties, while most components deal with the underlying electrical or digital signal-based representation of content. Collecting and sharing information about these workflows on the Semantic Web can be beneficial for content management or educational purposes. In this paper, we describe an ontological model for the representation of workflows in audio production, and show how this model facilitates capturing and sharing information about the production process. We then examine how this model can be used in a larger framework for representing domain knowledge about production and outline why this information is beneficial.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124779288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Toward the introduction of auditory information in dynamic visual attention models 动态视觉注意模型中听觉信息的引入
A. Coutrot, N. Guyader
{"title":"Toward the introduction of auditory information in dynamic visual attention models","authors":"A. Coutrot, N. Guyader","doi":"10.1109/WIAMIS.2013.6616164","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616164","url":null,"abstract":"Classical visual attention models only use visual features to predict where observers should look at. However, in daily life, visual information is never perceived without its corresponding audio signal. In a previous study, we found that sound modifies visual exploration by comparing the eye movements recorded when viewing videos with or without their original soundtrack. The aim of the presented research is to further understand how sound influences eye movements by controlling visual and audio contents of the videos, as well as the congruency between them. We describe an experiment with a novel approach in which observers watched videos belonging to four visual categories presenting different visual saliency distributions: landscapes, one moving object, several moving objects and faces. Videos were seen with their original soundtrack or with the soundtrack from another video belonging to the same visual category. Using different metrics to analyze the recorded eye movements, we found that sound has an influence only on videos containing faces and several moving objects. The original soundtrack decreases the variability between the eye positions of observers. Finally, we propose some cues to integrate sound information in classical visual attention models.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125619033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Group detection in still images by F-formation modeling: A comparative study F-formation建模在静止图像中的群体检测:比较研究
F. Setti, H. Hung, M. Cristani
{"title":"Group detection in still images by F-formation modeling: A comparative study","authors":"F. Setti, H. Hung, M. Cristani","doi":"10.1109/WIAMIS.2013.6616147","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616147","url":null,"abstract":"Automatically detecting groups of conversing people has become a hot challenge, although a formal, widely-accepted definition of them is lacking. This gap can be filled by considering the social psychological notion of an F-formation as a loose geometric arrangement. In the literature, two main approaches followed this line, exploiting Hough voting [1] from one side and Graph Theory [2] on the other. This paper offers a thorough comparison of these two methods, highlighting the strengths and weaknesses of both in different real life scenarios. Our experiments demonstrate a deeper understanding of the problem by identifying the circumstances in which to adopt a particular method. Finally our study outlines what aspects of the problem are important to address for future improvements to this task.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122243038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Vision-based maritime surveillance system using fused visual attention maps and online adaptable tracker 融合视觉注意图和在线自适应跟踪器的视觉海上监视系统
Konstantinos Makantasis, A. Doulamis, N. Doulamis
{"title":"Vision-based maritime surveillance system using fused visual attention maps and online adaptable tracker","authors":"Konstantinos Makantasis, A. Doulamis, N. Doulamis","doi":"10.1109/WIAMIS.2013.6616150","DOIUrl":"https://doi.org/10.1109/WIAMIS.2013.6616150","url":null,"abstract":"This paper presents a vision-based system for maritime surveillance using moving PTZ cameras. This system is intended to be used as an early warning system by local authorities. It fuses a visual attention method that exploits low-level image features appropriately selected for maritime environment, with an online adaptable neural network tracker, without making any assumptions about environmental or visual conditions. Systems performance was evaluated with videos from cameras placed at Limassol port and Venetian port of Chania and concerns robustness compared to dynamically changing visual conditions and different kinds of vessels, all in real time.","PeriodicalId":408077,"journal":{"name":"2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)","volume":"100 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115140146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信