2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)最新文献

筛选
英文 中文
Pollution Detection on Rail Surface for Adhesion Evaluation Using Multispectral Images 基于多光谱图像的轨道表面污染检测及其附着力评价
C. Nicodeme, B. Stanciulescu
{"title":"Pollution Detection on Rail Surface for Adhesion Evaluation Using Multispectral Images","authors":"C. Nicodeme, B. Stanciulescu","doi":"10.1109/DICTA.2017.8227477","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227477","url":null,"abstract":"There is a continuous rise of the number of trains' passengers to transport, therefore a densification of traffic is studied. For safety reasons, it becomes necessary to know the wheel-rail contact condition, in traffic, as it affects driving variables as adherence. The latest determines passenger's safety and the proper functioning of train equipment but it is often deteriorated due to recurrent pollution of the rail. From the analysis of the rolling surface, one can know if the rail is polluted or not and extract the kind of pollution and its influence on the train driving. The aim of this work is to contribute to guarantee and improve passengers' safety. In this paper, we propose a method for pollution detection and clustering, as a first step to the rail surface characterization. Methods using multispectral image analysis, specific data calibration, and hierarchical clustering based on Non Negative Matrix Factorization (NMF) are used.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127070034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Irish Sign Language Recognition Using Principal Component Analysis and Convolutional Neural Networks 使用主成分分析和卷积神经网络的爱尔兰手语识别
Marlon Oliveira, Houssem Chatbri, S. Little, Ylva Ferstl, N. O’Connor, Alistair Sutherland
{"title":"Irish Sign Language Recognition Using Principal Component Analysis and Convolutional Neural Networks","authors":"Marlon Oliveira, Houssem Chatbri, S. Little, Ylva Ferstl, N. O’Connor, Alistair Sutherland","doi":"10.1109/DICTA.2017.8227451","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227451","url":null,"abstract":"Hand-shape recognition is an important problem in computer vision with significant societal impact. In this work, we introduce a new image dataset for Irish Sign Language (ISL) recognition and we compare between two recognition approaches. The dataset was collected by filming human subjects performing ISL hand-shapes and movements. Then, we extracted frames from the videos. This produced a total of 52,688 images for the 23 common hand- shapes from ISL. Afterwards, we filter the redundant images with an iterative image selection process that selects the images which keep the dataset diverse. For classification, we use Principal Component Analysis (PCA) with with K- Nearest Neighbours (k-NN) and Convolutional Neural Networks (CNN). We obtain a recognition accuracy of 0.95 for our PCA model and 0.99 for our CNN model. We show that image blurring improves PCA results to 0.98. In addition, we compare times for classification.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128214418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Tomographic Reconstruction Using Global Statistical Priors 基于全局统计先验的层析重建
Preeti Gopal, Ritwick Chaudhry, S. Chandran, I. Svalbe, Ajit V. Rajwade
{"title":"Tomographic Reconstruction Using Global Statistical Priors","authors":"Preeti Gopal, Ritwick Chaudhry, S. Chandran, I. Svalbe, Ajit V. Rajwade","doi":"10.1109/DICTA.2017.8227496","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227496","url":null,"abstract":"Recent research in tomographic reconstruction is motivated by the need to efficiently recover detailed anatomy from limited measurements. One of the ways to compensate for the increasingly sparse sets of measurements is to exploit the information from emph{templates}, i.e., prior data available in the form of already reconstructed, structurally similar images. Towards this, previous work has exploited using a set of global and patch based dictionary priors. In this paper, we propose a global prior to improve both the speed and quality of tomographic reconstruction within a Compressive Sensing framework. We choose a set of potential representative 2D images referred to as templates, to build an eigenspace; this is subsequently used to guide the iterative reconstruction of a similar slice from sparse acquisition data. Our experiments across a diverse range of datasets show that reconstruction using an appropriate global prior, apart from being faster, gives a much lower reconstruction error when compared to the state of the art.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128964169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Learning Deep Spatial Lung Features by 3D Convolutional Neural Network for Early Cancer Detection 基于三维卷积神经网络的肺深部空间特征学习用于早期癌症检测
Taolin Jin, Hui Cui, Shan Zeng, Xiuying Wang
{"title":"Learning Deep Spatial Lung Features by 3D Convolutional Neural Network for Early Cancer Detection","authors":"Taolin Jin, Hui Cui, Shan Zeng, Xiuying Wang","doi":"10.1109/DICTA.2017.8227454","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227454","url":null,"abstract":"Accurate early lung cancer detection is essential towards precision oncology and would effectively improve the patients' survival rate. In this work, we explore the lung cancer early detection capacity by learning from deep spatial lung features. A 3D CNN network architecture is constructed with segmented CT lung volumes as training and testing samples. The new model extracts and projects 3D features to the following hidden layers, which preserves the temporal relations between neighboring CT slices. The well-built 3D CNN model consists of 11 layers which generates 12,544 neurons and 16 million parameters classifying whether the patient is diagnosed as cancer or not. ReLU nonlinearity and Sigmoid function are used as activation and classification methods. The model achieves a prediction accuracy of 87.5% where only the biomedical images themselves are used as the input dataset. The model's lowest error rate reaches 12.5% that improves the traditional AlexNet architecture by 2.8%.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133858304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Image Retrieval with Attribute-Associated Auxiliary References 基于属性关联辅助引用的图像检索
Lin Nie, Keze Wang, Wenxiong Kang, Yuefang Gao
{"title":"Image Retrieval with Attribute-Associated Auxiliary References","authors":"Lin Nie, Keze Wang, Wenxiong Kang, Yuefang Gao","doi":"10.1109/DICTA.2017.8227444","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227444","url":null,"abstract":"This paper develops a general framework of image retrieval, named A3, by introducing an auxiliary set of samples (object references), each of which is annotated with semantic attributes (tags). Given a query image (without tags), we first map it into the references by a non-convex sparse coding formulation, which jointly optimizes appearance reconstruction of the query and semantics consistency among references. A novel algorithm, namely semantic coherent coding (SCC), is presented to solve this optimization. As a result, a subset of references is selected to augment the query for retrieval. Then we can rank images to be retrieved by intuitively matching them together with the query and the constructed subset. The advantages of our approach are validated on the public datasets (Caltech-256 and ImageNet), and summarized as follows: (1) The auxiliary set of references gives rise to an alternating way for query augmentation while utilizing semantics. (2) The semantics of references can also be propagated into the query and database images.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123766092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Recognition of Human Emotions Induced by Visual Contents of Digital Images Based on Color Histogram 基于颜色直方图的数字图像视觉内容情感自动识别
Seyed Abdolreza Mohseni, H. Wu, J. Thom
{"title":"Automatic Recognition of Human Emotions Induced by Visual Contents of Digital Images Based on Color Histogram","authors":"Seyed Abdolreza Mohseni, H. Wu, J. Thom","doi":"10.1109/DICTA.2017.8227410","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227410","url":null,"abstract":"Using color histograms in automatic emotion recognition systems faces different issues. One of the important challenges is to determine the appropriate number of bins in the color histogram to achieve the highest recognition performance possible with minimal computations. This research focuses on emotion recognition induced by visual contents of images, or REVC for short, using ARTphoto dataset. Twenty-two different classifiers are used with color histograms in both RGB (red, green, blue) and HSV (hue, saturation, value) color spaces across different numbers of bins, and overall performance of each bin size is compared with that of other bin sizes. The research findings show that the performance of REVC system does not improve in terms of overall sensitivity rate, when the number of bins in color histogram is increased. Moreover, this paper identifies the advantage of using HSV color space over RGB in using color histogram for REVC systems. Furthermore, findings recognize the optimum number of bins in both RGB and HSV color spaces, and ANOVA (analysis of variance) is used to analyze experimental data, which identifies the optimum color histogram bin size used for HSV color space is significantly better than that used for RGB color space in REVC systems.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"304 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121281537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Bilinear CNN Models for Food Recognition 用于食物识别的双线性CNN模型
Hesen Chen, Jingyu Wang, Q. Qi, Yujian Li, Haifeng Sun
{"title":"Bilinear CNN Models for Food Recognition","authors":"Hesen Chen, Jingyu Wang, Q. Qi, Yujian Li, Haifeng Sun","doi":"10.1109/DICTA.2017.8227411","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227411","url":null,"abstract":"Due to the diversity of food types and the slight differences between different dishes, the genre of food images becomes a new challenge in the field of computer vision. To tackle this problem, recent efforts are focusing on designing hand-crafted features or extracting features automatically by using deep convolutional neural network. Although these methods have reported a series of success, their general architectures fail to capture the fine-grained features of similar dishes adequately. Inspired by the bilinear CNN models in the field of fine-grained classification, we have exploited such a similar structure in which two deep convolution networks are used as feature extractors and the outputs of them fused to obtain fine-grained features. These features are used to train the food classifier. We have conducted experiments on three publicly benchmark food datasets to evaluate the proposed architecture. The experiments exhibit that our method is comparable to the existing approaches.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127029539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Constrained Stochastic Gradient Descent: The Good Practice 约束随机梯度下降:良好实践
S. Roy, Mehrtash Harandi
{"title":"Constrained Stochastic Gradient Descent: The Good Practice","authors":"S. Roy, Mehrtash Harandi","doi":"10.1109/DICTA.2017.8227420","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227420","url":null,"abstract":"Stochastic Gradient Descent (SGD) is the method of choice for large scale problems, most notably in deep learning. Recent studies target improving convergence and speed of the SGD algorithm. In this paper, we equip the SGD algorithm and its advanced versions with an intriguing feature, namely handling constrained problems. Constraints such as orthogonality are pervasive in learning theory. Nevertheless and to some extent surprising, constrained SGD algorithms are rarely studied. Our proposal makes use of Riemannian geometry and accelerated optimization techniques to deliver efficient and constrained-aware SGD methods.We will assess and contrast our proposed approaches in a wide range of problems including incremental dimensionality reduction, karcher mean and deep metric learning.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132908642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Multi-Object Model-Free Tracking with Joint Appearance and Motion Inference 基于关节外观和运动推理的多目标无模型跟踪
Chongyu Liu, Rui Yao, S. H. Rezatofighi, I. Reid, Javen Qinfeng Shi
{"title":"Multi-Object Model-Free Tracking with Joint Appearance and Motion Inference","authors":"Chongyu Liu, Rui Yao, S. H. Rezatofighi, I. Reid, Javen Qinfeng Shi","doi":"10.1109/DICTA.2017.8227468","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227468","url":null,"abstract":"Multi-object model-free tracking is challenging because the tracker is not aware of the objects' type (not allowed to use object detectors), and needs to distinguish one object from background as well as other similar objects. Most existing methods keep updating their appearance model individually for each target, and their performance is hampered by sudden appearance change and/or occlusion. We propose to use both appearance model and motion model to overcome this issue. We introduce an indicator variable to predict sudden appearance change and occlusion. When they happen, our model stops updating the appearance model to avoid parameter update based on background or incorrect object, and rely more on motion model to track. Moreover, we consider the correlation among all targets, and seek the joint optimal locations for all target simultaneously. We formulate the problem of finding the most likely locations jointly as a graphical model inference problem, and learn the joint parameters for both appearance model and motion model in an online fashion in the framework of LaRank. Experiment results show that our method outperforms the state-of-the-art.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129309853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Method to Create Stable Lighting and Remove Specular Reflections for Vision Systems 一种为视觉系统创造稳定照明和消除镜面反射的方法
Gilbert Eaton, Andrew Busch, Rudi Bartels, Yongsheng Gao
{"title":"A Method to Create Stable Lighting and Remove Specular Reflections for Vision Systems","authors":"Gilbert Eaton, Andrew Busch, Rudi Bartels, Yongsheng Gao","doi":"10.1109/DICTA.2017.8227392","DOIUrl":"https://doi.org/10.1109/DICTA.2017.8227392","url":null,"abstract":"A lighting system and method has been developed which has shown in testing to allow quality images to be obtained that are free from two particular problems, specular reflections on the subject, and light intensity variation. These problems both diminish the ability to compare objects for attributes such as colour variation, edges, contours, and many other features. The system developed eliminates specular reflection by using the cross-polarisation configuration, and reduced flickering due to fluctuations in the power supply to negligible levels by constructing a high-power DC source capable of providing sufficient 12 Volt power. These two improvements create an environment suitable for taking high-quality, noise free images at high shutter speeds for the purpose of assessing the quality of strawberries moving on a real-time production line.","PeriodicalId":194175,"journal":{"name":"2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132898951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信