2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)最新文献

筛选
英文 中文
DASM: An open source active shape model for automatic registration of objects DASM:用于对象自动配准的开源活动形状模型
David Macurak, Amrutha Sethuram, K. Ricanek, B. Barbour
{"title":"DASM: An open source active shape model for automatic registration of objects","authors":"David Macurak, Amrutha Sethuram, K. Ricanek, B. Barbour","doi":"10.1109/NCVPRIPG.2013.6776244","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776244","url":null,"abstract":"The main contribution of this paper is to introduce DASM - Dynamic Active Shape Models, an open source software for the automatic detection of fiducial points on objects for subsequent registration, to the research community. DASM leverages the tremendous work of STASM, a well known software library for automatic detection of points on faces. In this work we compare DASM to other well-known techniques for automatic face registration: Active Appearance Models (AAM) and Constrained Local Models (CLM). Further we show that DASM outperforms these techniques on a per registration-point error, average object error, and on cumulative error distribution. As a follow on, we show that DASM outperforms STASM v3.1 on model training and registration by leveraging open source libraries for computer vision (OpenCV v2.4) and threading/parallelism (OpenMP). The improvements in speed and performance of DASM allows for extremely dense registration, 252 points on the face, in video applications.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130675652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysing gait sequences using Latent Dirichlet Allocation for certain human actions 基于潜在狄利克雷分配的步态序列分析
A. DeepakN., R. Hariharan, U. Sinha
{"title":"Analysing gait sequences using Latent Dirichlet Allocation for certain human actions","authors":"A. DeepakN., R. Hariharan, U. Sinha","doi":"10.1109/NCVPRIPG.2013.6776173","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776173","url":null,"abstract":"Conventional human action recognition algorithm and method generate coarse clusters of input videos approximately 2-4 clusters with less information regarding the cluster generation. This problem is solved by proposing Latent Dirichlet Allocation algorithm that transforms the extracted gait sequences in gait domain into documents-words in text domain. These words are then used to group the input documents into finer clusters approximately 8-9 clusters. In this approach, we have made an attempt to use gait analysis in recognizing human actions, where the gait analysis requires to have some motion in lower parts of the human body like leg. As the videos of Weizmann dataset have some actions that exhibits these movements, we are able use these motion parameters to recognize certain human actions. Experiments on Weizmann dataset suggest that the proposed Latent Dirichlet Allocation algorithm is an efficient method for recognizing human actions from the video streams.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131159878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Recognition and identification of target images using feature based retrieval in UAV missions 无人机任务中基于特征检索的目标图像识别与识别
Shweta Singh, D. V. Rao
{"title":"Recognition and identification of target images using feature based retrieval in UAV missions","authors":"Shweta Singh, D. V. Rao","doi":"10.1109/NCVPRIPG.2013.6776165","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776165","url":null,"abstract":"With the introduction of unmanned air vehicles as force multipliers in the defense services worldwide, automatic recognition and identification of ground based targets has become an important area of research in the defense community. Due to inherent instabilities in smaller unmanned platforms, image blurredness and distortion need to be addressed for the successful recognition of the target. In this paper, an image enhancement technique that can improve images' quality acquired by an unmanned system is proposed. An image de-blurring technique based on blind de-convolution algorithm which adaptively enhances the edges of characters and wipes off blurredness effectively is proposed. A content-based image retrieval technique based on features extraction to generate an image description and a compact feature vector that represents the visual information, color, texture and shape is used with a minimum distance algorithm to effectively retrieve the plausible target images from a library of images stored in a target folder. This methodology was implemented for planning and gaming the UAV/UCAV missions in the Air Warfare Simulation System.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128882269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A hybrid method for object identification and event detection in video 视频中目标识别与事件检测的混合方法
P. KrishnaKumar, L. Parameswaran
{"title":"A hybrid method for object identification and event detection in video","authors":"P. KrishnaKumar, L. Parameswaran","doi":"10.1109/NCVPRIPG.2013.6776223","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776223","url":null,"abstract":"Video event detection (VED) is a challenging task especially with a large variety of objects in the environment. Even though there exist numerous algorithms for event detection, most of them are unsuitable for a typical consumer purpose. A hybrid method for detecting and identifying the moving objects by their color and spatial information is presented in this paper. In tracking multiple moving objects, the system makes use of motion of changed regions. In this approach, first, the object detector will look for the existence of objects that have already been registered. Then the control is passed on to an event detector which will wait for an event to happen which can be object placement or object removal. The object detector becomes active only if any event is detected. Simple training procedure using a single color camera in HSV color space makes it a consumer application. The proposed model has proved to be robust in various indoor environments and different types of background scenes. The experimental results prove the feasibility of the proposed method.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128443381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
M-ary reversible contrast mapping in reversible watermarking with optimal distortion control 具有最优失真控制的可逆水印中的任意可逆对比度映射
S. Maity, H. Maity
{"title":"M-ary reversible contrast mapping in reversible watermarking with optimal distortion control","authors":"S. Maity, H. Maity","doi":"10.1109/NCVPRIPG.2013.6776269","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776269","url":null,"abstract":"A generalized form of reversible contrast mapping (RCM), analogous to M-ary modulation in communication, is developed here for reversible watermarking in digital images. Then an optimized distortion control framework in M-ary scheme is considered to improve data hiding capacity while meeting the embedding distortion constraint. Simulation results show that the combination of different M-ary approaches, using the different points representing the different RCM transformation functions, outperforms the embedding rate-visual quality-security of the hidden information compared to the existing RCM, difference expansion (DE) and prediction error expansion (PEE) methods during over embedding. Numerical results show that an average of 20% improvement in visual quality, 35% improvement in security of the hidden data at 1 bpp embedding rate is achieved for the proposed method compared to the existing PEE works. All these effectiveness are demonstrated with a number of simulation results.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126785954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Real-time approximate and exact CSG of implicit surfaces on the GPU 在GPU上实现隐式曲面的实时逼近和精确CSG
Jag Mohan Singh
{"title":"Real-time approximate and exact CSG of implicit surfaces on the GPU","authors":"Jag Mohan Singh","doi":"10.1109/NCVPRIPG.2013.6776199","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776199","url":null,"abstract":"We present a simple and powerful scheme to allow CSG of implicit surfaces on the GPU. We decompose the boolean expression of surfaces into sum-of-products form. Our algorithm presented in this paper then renders each product term, sum of products can be automatically by enabling depth test. Our Approximate CSG uses adaptive marching points algorithm for finding ray-surface intersection. Once we find an interval where root exists after root-isolation, this is used for presence of intersection. We perform root-refinement only for the uncomplemented terms in the product. Exact CSG is done by using the discriminant of the ray-surface intersection for the presence of the root. Now we can simply evaluate the product expression by checking all uncomplemented terms should be true and all complemented terms should be false. If our condition is met, we find the maximum of all the roots among uncomplemented terms to be the solution. Our algorithm is linear in the number of terms O(n). We achieve real-time rates for 4-5 terms in the product for approximate CSG. We achieve more than real-time rates for Exact CSG. Our primitives are implicit surfaces so we can achieve fairly complex results with less terms.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122552265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mean-shift based object detection and clustering from high resolution remote sensing imagery 基于均值偏移的高分辨率遥感影像目标检测与聚类
T. SushmaLeela, R. Chandrakanth, J. Saibaba, G. Varadan, S. Mohan
{"title":"Mean-shift based object detection and clustering from high resolution remote sensing imagery","authors":"T. SushmaLeela, R. Chandrakanth, J. Saibaba, G. Varadan, S. Mohan","doi":"10.1109/NCVPRIPG.2013.6776271","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776271","url":null,"abstract":"Object detection from remote sensing images has inherent difficulties due to cluttered backgrounds and noisy regions from the urban area in high resolution images. Detection of objects with regular geometry, such as circles from an image uses strict feature based detection. Using region based segmentation techniques such as K-Means has the inherent disadvantage of knowing the number of classes apriori. Contour based techniques such as Active contour models, sometimes used in remote sensing also has the problem of knowing the approximate location of the region and also the noise will hinder its performance. A template based approach is not scale and rotation invariant with different resolutions and using multiple templates is not a feasible solution. This paper proposes a methodology for object detection based on mean shift segmentation and non-parametric clustering. Mean shift is a non-parametric segmentation technique, which in its inherent nature is able to segment regions according to the desirable properties like spatial and spectral radiance of the object. A prior knowledge about the shape of the object is used to extract the desire object. A hierarchical clustering method is adopted to cluster the objects having similar shape and spatial features. The proposed methodology is applied on high resolution EO images to extract circular objects. The methodology found to be better and robust even in the cluttered and noisy background. The results are also evaluated using different evaluation measures.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127769309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
ISIgraphy: A tool for online handwriting sample database generation ISIgraphy:一个在线手写样本数据库生成工具
Arindam Das, U. Bhattacharya
{"title":"ISIgraphy: A tool for online handwriting sample database generation","authors":"Arindam Das, U. Bhattacharya","doi":"10.1109/NCVPRIPG.2013.6776181","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776181","url":null,"abstract":"Online handwriting recognition research has recently received significant thrust. Specifically for Indian scripts, handwriting recognition has not been focused much till in the near past. However, due to generous Government funding through the group on Technology Development for Indian Languages (TDIL) of the Ministry of Communication & Information Technology (MC&IT), Govt. of India, research in this area has received due attention and several groups are now engaged in research and development works for online handwriting recognition in different Indian scripts. An extensive bottleneck of the desired progress in this area is the difficulty of collection of large sample databases of online handwriting in various scripts. Towards the same, recently a user-friendly tool on Android platform has been developed to collect data on handheld devices. This tool is called ISIgraphy and has been uploaded in the Google Play for free download. This application is designed well enough to store handwritten data samples in large scales in user-given file names for distinct users. Its use is script independent, meaning that it can collect and store handwriting samples written in any language, not necessarily an Indian script. It has an additional module for retrieval and display of stored data. Moreover, it can directly send the collected data to others via electronic mail.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128692314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Extraction of line-word-character segments directly from run-length compressed printed text-documents 直接从运行长度压缩的打印文本-文档中提取行-字-字符段
M. Javed, P. Nagabhushan, B. B. Chaudhuri
{"title":"Extraction of line-word-character segments directly from run-length compressed printed text-documents","authors":"M. Javed, P. Nagabhushan, B. B. Chaudhuri","doi":"10.1109/NCVPRIPG.2013.6776195","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776195","url":null,"abstract":"Segmentation of a text-document into lines, words and characters, which is considered to be the crucial preprocessing stage in Optical Character Recognition (OCR) is traditionally carried out on uncompressed documents, although most of the documents in real life are available in compressed form, for the reasons such as transmission and storage efficiency. However, this implies that the compressed image should be decompressed, which indents additional computing resources. This limitation has motivated us to take up research in document image analysis using compressed documents. In this paper, we think in a new way to carry out segmentation at line, word and character level in run-length compressed printed-text-documents. We extract the horizontal projection profile curve from the compressed file and using the local minima points perform line segmentation. However, tracing vertical information which leads to tracking words-characters in a run-length compressed file is not very straight forward. Therefore, we propose a novel technique for carrying out simultaneous word and character segmentation by popping out column runs from each row in an intelligent sequence. The proposed algorithms have been validated with 1101 text-lines, 1409 words and 7582 characters from a data-set of 35 noise and skew free compressed documents of Bengali, Kannada and English Scripts.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117340286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
An OCR system for the Meetei Mayek script Meetei Mayek脚本的OCR系统
Subhankar Ghosh, U. Barman, P. Bora, Tourangbam Harishore Singh, B. Chaudhuri
{"title":"An OCR system for the Meetei Mayek script","authors":"Subhankar Ghosh, U. Barman, P. Bora, Tourangbam Harishore Singh, B. Chaudhuri","doi":"10.1109/NCVPRIPG.2013.6776228","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776228","url":null,"abstract":"This paper presents an implementation of an OCR system for the Meetei Mayek script. The script has been newly reintroduced and there is a growing set of documents currently available in this script. Our system accepts an image of the textual portion of a page and outputs the text in the Unicode format. It incorporates preprocessing, segmentation and classification stages. However, no post-processing is done to the output. The system achieves an accuracy of about 96% on a moderate database.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"167 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114751563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信