ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream最新文献

筛选
英文 中文
A non-parametric unsupervised approach for content based image retrieval and clustering 基于内容的图像检索和聚类的非参数无监督方法
Konstantinos Makantasis, A. Doulamis, N. Doulamis
{"title":"A non-parametric unsupervised approach for content based image retrieval and clustering","authors":"Konstantinos Makantasis, A. Doulamis, N. Doulamis","doi":"10.1145/2510650.2510656","DOIUrl":"https://doi.org/10.1145/2510650.2510656","url":null,"abstract":"Nowadays, there are available extremely large collections of images located on distributed and heterogeneous platforms over the web. The proliferation of billions of shared photos has outpaced the current technology for browsing such collections, but at the same time it spurred the emergence of new image retrieval techniques based not only on photos' visual information, but on geo-location tags and camera exif data. Although, additional image information may be proven very useful for preliminary image retrieval, the final retrieved result is necessary to be refined by exploiting visual information.\u0000 In this paper we present a process for refining image retrieval results by exploiting and fusing two unsupervised clustering techniques: DBSCAN and spectral clustering. DBSCAN algorithm is used to remove outliers from the initially retrieved image set, and spectral clustering finalizes retrieval process by clustering together visually similar images. However, DBSCAN and spectral clustering require manual tunning of their parameters, which usually requires a priori knowledge of the dataset. To overcome this problem we developed a tuning mechanism that automatically tunes the parameters of both algorithms. For the evaluation of the proposed approach we used thousands of images from Flickr downloaded using text queries for well known cultural heritage monuments.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129925117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Behavior recognition from video based on human constrained descriptor and adaptable neural networks 基于人类约束描述符和自适应神经网络的视频行为识别
A. Voulodimos, N. Doulamis, S. Tsafarakis
{"title":"Behavior recognition from video based on human constrained descriptor and adaptable neural networks","authors":"A. Voulodimos, N. Doulamis, S. Tsafarakis","doi":"10.1145/2510650.2510659","DOIUrl":"https://doi.org/10.1145/2510650.2510659","url":null,"abstract":"In this paper we introduce a new descriptor, the Human Constrained Pixel Change History (HC-PCH), which is based on Pixel Change History (PCH) but focuses on the human body movements over time. We propose a modification of the conventional PCH which entails the calculation of two probabilistic maps, based on human face and body detection respectively. The features extracted from this descriptor are used as input to an HMM-based behavior recognition framework. We also introduce a rectification framework of behavior recognition and classification by incorporating an expert user's feedback into the learning process through two proposed schemes: a plain non-linear one and an adaptable one, which requires fewer training samples and is more effective in decreasing misclassification error. The methods presented are validated on a real-world computer vision dataset comprising challenging video sequences from an industrial environment.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126604235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
On improving the robustness of variational optical flow against illumination changes 提高变分光流对光照变化的鲁棒性
M. A. Mohamed, Hatem A. Rashwan, B. Mertsching, M. García, D. Puig
{"title":"On improving the robustness of variational optical flow against illumination changes","authors":"M. A. Mohamed, Hatem A. Rashwan, B. Mertsching, M. García, D. Puig","doi":"10.1145/2510650.2510660","DOIUrl":"https://doi.org/10.1145/2510650.2510660","url":null,"abstract":"The brightness constancy assumption is the base of estimating the flow fields in most differential optical flow approaches. However, the brightness constancy constraint easily violates with any variation in the lighting conditions in the scene. Thus, this work proposes a robust data term against illumination changes based on a rich descriptor. This descriptor extracts the textures features for each image in the two consecutive images using local edge responses. In addition, a weighted non-local term depending on the intensity similarity, the spatial distance and the occlusion state of pixels is integrated within the adapted duality total variational optical flow algorithm in order to obtain accurate flow fields. The proposed model yields state-of-the-art results on the the KITTI optical flow database and benchmark.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122495570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Domain transfer for person re-identification 域名转移用于人员重新识别
Ryan Layne, Timothy M. Hospedales, S. Gong
{"title":"Domain transfer for person re-identification","authors":"Ryan Layne, Timothy M. Hospedales, S. Gong","doi":"10.1145/2510650.2510658","DOIUrl":"https://doi.org/10.1145/2510650.2510658","url":null,"abstract":"Automatic person re-identification in is a crucial capability underpinning many applications in public space video surveillance. It is challenging due to intra-class variation in person appearance when observed in different views, together with limited inter-class variability. Various recent approaches have made great progress in re-identification performance using discriminative learning techniques. However, these approaches are fundamentally limited by the requirement of extensive annotated training data for every pair of views. For practical re-identification, this is an unreasonable assumption, as annotating extensive volumes of data for every pair of cameras to be re-identified may be impossible or prohibitively expensive.\u0000 In this paper we move toward relaxing this strong assumption by investigating flexible multi-source transfer of re-identification models across camera pairs. Specifically, we show how to leverage prior re-identification models learned for a set of source view pairs (domains), and flexibly combine these to obtain good re-identification performance in a target view pair (domain) with greatly reduced training data requirements in the target domain.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134100219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Cross-domain traffic scene understanding by motion model transfer 基于运动模型迁移的跨域交通场景理解
Xun Xu, S. Gong, Timothy M. Hospedales
{"title":"Cross-domain traffic scene understanding by motion model transfer","authors":"Xun Xu, S. Gong, Timothy M. Hospedales","doi":"10.1145/2510650.2510657","DOIUrl":"https://doi.org/10.1145/2510650.2510657","url":null,"abstract":"This paper proposes a novel framework for cross-domain traffic scene understanding. Existing learning-based outdoor wide-area scene interpretation models suffer from requiring long term data collection in order to acquire statistically sufficient model training samples for every new scene. This makes installation costly, prevents models from being easily relocated, and from being used in UAVs with continuously changing scenes. In contrast, our method adopts a geometrical matching approach to relate motion models learned from a database of source scenes (source domains) with a handful sparsely observed data in a new target scene (target domain). This framework is capable of online ''sparse-shot'' anomaly detection and motion event classification in the unseen target domain, without the need for extensive data collection, labelling and offline model training for each new target domain. That is, trained models in different source domains can be deployed to a new target domain with only a few unlabelled observations and without any training in the new target domain. Crucially, to provide cross-domain interpretation without risk of dramatic negative transfer, we introduce and formulate a scene association criterion to quantify transferability of motion models from one scene to another. Extensive experiments show the effectiveness of the proposed framework for cross-domain motion event classification, anomaly detection and scene association.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134579907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Warping trajectories for video synchronization 视频同步的扭曲轨迹
S. Shankar, Joan Lasenby, A. Kokaram
{"title":"Warping trajectories for video synchronization","authors":"S. Shankar, Joan Lasenby, A. Kokaram","doi":"10.1145/2510650.2510654","DOIUrl":"https://doi.org/10.1145/2510650.2510654","url":null,"abstract":"Temporal synchronization of multiple video recordings of the same dynamic event is a critical task in many computer vision applications e.g. novel view synthesis and 3D reconstruction. Typically this information is implied, since recordings are made using the same timebase, or time-stamp information is embedded in the video streams. Recordings using consumer grade equipment do not contain this information; hence, there is a need to temporally synchronize signals using the visual information itself. Previous work in this area has either assumed good quality data with relatively simple dynamic content or the availability of precise camera geometry. In this paper, we propose a technique which exploits feature trajectories across views in a novel way, and specifically targets the kind of complex content found in consumer generated sports recordings, without assuming precise knowledge of fundamental matrices or homographies. Our method automatically selects the moving feature points in the two unsynchronized videos whose 2D trajectories can be best related, thereby helping to infer the synchronization index. We evaluate performance using a number of real recordings and show that synchronization can be achieved to within 1 sec, which is better than previous approaches.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"115 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130028267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Abnormal crowd behavior detection and localization using maximum sub-sequence search 基于最大子序列搜索的人群异常行为检测与定位
Kai-wen Cheng, Yie-Tarng Chen, Wen-Hsien Fang
{"title":"Abnormal crowd behavior detection and localization using maximum sub-sequence search","authors":"Kai-wen Cheng, Yie-Tarng Chen, Wen-Hsien Fang","doi":"10.1145/2510650.2510655","DOIUrl":"https://doi.org/10.1145/2510650.2510655","url":null,"abstract":"This paper presents a novel framework for anomaly event detection and localization in crowded scenes. We propose an anomaly detector that extends the Bayes classifier from multi-class to one-class classification to characterize normal events. We also propose a localization scheme for anomaly localization as a maximum subsequence problem in a video sequence. The maximum subsequence algorithm locates an anomaly event by discovering the optimal collection of successive patches with spatial proximity over time without prior knowledge of the size, start and end of the anomaly event. Our localization scheme can locate multiple occurrences of abnormal events in spite of noise. Experimental results on the well-established UCSD dataset show that the proposed framework significantly outperforms state-of-the-art methods up to 53.55% localization rate. This study concludes that the localization framework plays an important role in abnormal event detection.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"17 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120853365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Background modeling methods for visual detection of maritime targets 海上目标视觉检测的背景建模方法
Paris Kaimakis, N. Tsapatsoulis
{"title":"Background modeling methods for visual detection of maritime targets","authors":"Paris Kaimakis, N. Tsapatsoulis","doi":"10.1145/2510650.2510652","DOIUrl":"https://doi.org/10.1145/2510650.2510652","url":null,"abstract":"We propose a system for real-time detection of maritime targets based on monocular video data. In the absence of a priori knowledge about their appearance, targets are detected implicitly via the statistical modeling of the scene's nonstationary background. A probabilistic treatment regarding target compactness is also presented. The proposed system currently acts as a stand-alone maritime surveillance application, and may also be used as an early detection stage within a larger maritime target tracking framework.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126230793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Hand gesture recognition with depth data 手势识别与深度数据
Fabio Dominio, Mauro Donadeo, Giulio Marin, P. Zanuttigh, G. Cortelazzo
{"title":"Hand gesture recognition with depth data","authors":"Fabio Dominio, Mauro Donadeo, Giulio Marin, P. Zanuttigh, G. Cortelazzo","doi":"10.1145/2510650.2510651","DOIUrl":"https://doi.org/10.1145/2510650.2510651","url":null,"abstract":"Depth data acquired by current low-cost real-time depth cameras provide a very informative description of the hand pose, that can be effectively exploited for gesture recognition purposes. This paper introduces a novel hand gesture recognition scheme based on depth data. The hand is firstly extracted from the acquired depth maps with the aid also of color information from the associated views. Then the hand is segmented into palm and finger regions. Next, two different set of feature descriptors are extracted, one based on the distances of the fingertips from the hand center and the other on the curvature of the hand contour. Finally, a multi-class SVM classifier is employed to recognize the performed gestures. The proposed scheme runs in real-time and is able to achieve a very high accuracy on depth data acquired with the Kinect.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126110147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
Nobody likes Mondays: foreground detection and behavioral patterns analysis in complex urban scenes 没有人喜欢星期一:在复杂的城市场景中进行前景检测和行为模式分析
Gloria Zen, John Krumm, N. Sebe, E. Horvitz, Ashish Kapoor
{"title":"Nobody likes Mondays: foreground detection and behavioral patterns analysis in complex urban scenes","authors":"Gloria Zen, John Krumm, N. Sebe, E. Horvitz, Ashish Kapoor","doi":"10.1145/2510650.2510653","DOIUrl":"https://doi.org/10.1145/2510650.2510653","url":null,"abstract":"Streams of images from large numbers of surveillance webcams are available via the web. The continuous monitoring of activities at different locations provides a great opportunity for research on the use of vision systems for detecting actors, objects, and events, and for understanding patterns of activity and anomaly in real-world settings. In this work we show how images available on the web from surveillance webcams can be used as sensors in urban scenarios for monitoring and interpreting states of interest such as traffic intensity. We highlight the power of the cyclical aspect of the lives of people and of cities. We extract from long-term streams of images typical patterns of behavior and anomalous events and situations, based on considerations of day of the week and time of day. The analysis of typia and atypia required a robust method for background subtraction. For this purpose, we present a method based on sparse coding which outperforms state-of-the-art works on complex and crowded scenes.","PeriodicalId":360789,"journal":{"name":"ACM/IEEE international workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Stream","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126361772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信