2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)最新文献

筛选
英文 中文
Deep Learning for Pollen Sac Detection and Measurement on Honeybee Monitoring Video 基于深度学习的蜜蜂监控视频花粉囊检测与测量
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961011
Cheng Yang, J. Collins
{"title":"Deep Learning for Pollen Sac Detection and Measurement on Honeybee Monitoring Video","authors":"Cheng Yang, J. Collins","doi":"10.1109/IVCNZ48456.2019.8961011","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961011","url":null,"abstract":"This paper introduces a new model which applies deep learning techniques to pollen sac detection and measurement on honeybee monitoring video. The outcome of this model is a measurement of the number of pollen sacs being brought to the beehive, so that beekeepers will not need to open beehives frequently to check food storage. The pollen sacs are detected on individual bee images which are collected using a bee detection model on the entire video frame. The pollen detection model is built using a deep convolutional neural network. The architecture is Faster RCNN with VGG-16 core network. The network is trained to detect pollen sacs so the individual bee images are identified as either pollen or nonpollen bee images. This pollen sac detection model is then combined with a bee tracking model, so that each flying bee tracked on successive video frames is identified as carrying pollen or not. Finally, the number of pollen-carrying bees can be counted. The experimental results show that the measurement error of this model is 7%. The deep learning model improves the results from the conventional image processing method, which produced 33% measurement error.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"95 37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129222814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Fingertips Detection in Egocentric Video Frames using Deep Neural Networks 基于深度神经网络的自中心视频帧指尖检测
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8960957
Purnendu Mishra, K. Sarawadekar
{"title":"Fingertips Detection in Egocentric Video Frames using Deep Neural Networks","authors":"Purnendu Mishra, K. Sarawadekar","doi":"10.1109/IVCNZ48456.2019.8960957","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8960957","url":null,"abstract":"In recent years, there has been much advancement in Augmented Reality technologies. Also, there has been a rise in the usage of wearable cameras. These technologies allow us to interact with the virtual world and the real world simultaneously. Hand gestures or finger gestures can be used to provide input instructions replacing conventional tools like a keyboard or a mouse. This paper introduces an improvement over the YOLSE (You Only Look what You Should See) model towards multiple fingertip position estimation. We propose a regression-based technique to locate fingertip(s) in a multi-gesture condition. First, the hand gesture is segmented from the scene using a deep neural network (DNN) based object detection model. Next, fingertip(s) positions are estimated using MobileNetv2 architecture. It is difficult to use direct regression when the varying number of visible fingertips are present in different egocentric hand gestures. We used the multi-label classification concept to identify all the visible extended fingers in the image. Average errors on RGB image with a resolution of 640 × 480 is 6.1527 pixels. The processing time of 9.072 ms is achieved on Nvidia GeForce GTX 1080 GPU.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125365317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Generalization Approach for CNN-based Object Detection in Unconstrained Outdoor Environments 无约束室外环境下基于cnn的目标检测泛化方法
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8960992
Hedi Hedayati, B. McGuinness, M. Cree, J. Perrone
{"title":"Generalization Approach for CNN-based Object Detection in Unconstrained Outdoor Environments","authors":"Hedi Hedayati, B. McGuinness, M. Cree, J. Perrone","doi":"10.1109/IVCNZ48456.2019.8960992","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8960992","url":null,"abstract":"Recent advances in object detection using convolutional neural networks provide new opportunities to address various problems faced in agriculture. However, training a reliable object detection model which can be generalized to different environments requires large amounts of data which is difficult to obtain. This is particularly true when considering detection of wilding conifers, which are scattered across a diverse, unconstrained environment. Here we argue that a reliable dataset can be developed for wilding detection by synthetically enlarging the training dataset. A reliable detection model was trained using only 100 real images of wilding conifers.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"262 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122765743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Coniferous Trees Needles-Based Taxonomy Classification 针叶树的分类与分类
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961023
M. Haindl, Pavel Zid
{"title":"Coniferous Trees Needles-Based Taxonomy Classification","authors":"M. Haindl, Pavel Zid","doi":"10.1109/IVCNZ48456.2019.8961023","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961023","url":null,"abstract":"This paper introduces multispectral rotationally in-variant textural features of the Markovian type applied for the effective coniferous tree needles categorization. Presented texture features are inferred from the descriptive multispectral spiral wide-sense Markov model. Unlike the alternative texture recognition methods based on various gray-scale discriminative textural descriptions, we take advantage of the needles texture representation, which is fully descriptive multispectral and rotationally invariant.The presented method achieves high accuracy for needles recognition. Thus it can be used for reliable coniferous tree taxon classification. Our classifier is tested on the open source needles database Aff, which contains 716 high-resolution images from 11 diverse coniferous tree species.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"438 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131406757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pedestrian Proximity Detection using RGB-D Data 基于RGB-D数据的行人接近检测
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961013
Adam Tupper, R. Green
{"title":"Pedestrian Proximity Detection using RGB-D Data","authors":"Adam Tupper, R. Green","doi":"10.1109/IVCNZ48456.2019.8961013","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961013","url":null,"abstract":"This paper presents a novel method for pedestrian detection and distance estimation using RGB-D data. We use Mask R-CNN for instance-level pedestrian segmentation, and the Semiglobal Matching algorithm for computing depth information from a pair of infrared images captured by an Intel RealSense D435 stereo vision depth camera. The resulting depth map is post-processed using both spatial and temporal edge-preserving filters and spatial hole-filling to mitigate erroneous or missing depth values. The distance to each pedestrian is estimated using the median depth value of the pixels in the depth map covered by the predicted mask. Unlike previous work, our method is evaluated on, and performs well across, a wide spectrum of outdoor lighting conditions. Our proposed technique is able to detect and estimate the distance of pedestrians within 5m with an average accuracy of 87.7%.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124099295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Recognizing Text with a CNN 用CNN识别文本
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961031
Kulsoom Mansoor, C. Olson
{"title":"Recognizing Text with a CNN","authors":"Kulsoom Mansoor, C. Olson","doi":"10.1109/IVCNZ48456.2019.8961031","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961031","url":null,"abstract":"We seek to detect text in images using multiple techniques and recognize characters using a Convolutional Neural Network (CNN). Individual characters are combined to form words, which can then be used in a variety of applications, such as automated translation. Text recognition is difficult when different types of text formats and conditions are involved, such as fonts, orientation, color, complex backgrounds, and low-quality images. Our contribution is a novel combination of techniques to perform text detection and a CNN model to classify text characters. Experiments show that both the detection algorithm and machine learning model generally succeed with clear text. The system has more difficulty detecting text from complex and low-resolution images, as well as parsing words whose characters are connected together, since this causes segmentation issues.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126353551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Adaptation of Bidirectional Kalman Filter to Multi-Frequency Time-of-Flight Range Imaging 双向卡尔曼滤波在多频飞行时间距离成像中的自适应
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8960962
A. Alqassab, L. Streeter, M. Cree, Carl A. Lickfold, V. Farrow, S. Lim
{"title":"Adaptation of Bidirectional Kalman Filter to Multi-Frequency Time-of-Flight Range Imaging","authors":"A. Alqassab, L. Streeter, M. Cree, Carl A. Lickfold, V. Farrow, S. Lim","doi":"10.1109/IVCNZ48456.2019.8960962","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8960962","url":null,"abstract":"Time-of-flight cameras obtain range measurements by capturing multiple raw frames. Motion in a scene during capture leads to inaccurate range measurements. The Bidirectional Kalman filter is a method known to reduce error due to transverse motion in cameras operating with a single modulation frequency. In this paper we adapt the Bidirectional Kalman to multi-frequency operated cameras by having the prediction component of the Kalman take into account the change in amplitude, and phase shift due to frequency. In the quantitative experiment, the proposed method produces less error than the classical discrete Fourier transform approach in 70% of the instances. The qualitative experiment shows a significant reduction in blur due to motion.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132687461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Minutiae Triangle Graphs: A New Fingerprint Representation with Invariance Properties 细部三角形图:一种新的具有不变性的指纹表示
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8960988
Akmal-Jahan Mohamed-Abdul-Cader, Watcharapong Chaidee, Jasmine Banks, V. Chandran
{"title":"Minutiae Triangle Graphs: A New Fingerprint Representation with Invariance Properties","authors":"Akmal-Jahan Mohamed-Abdul-Cader, Watcharapong Chaidee, Jasmine Banks, V. Chandran","doi":"10.1109/IVCNZ48456.2019.8960988","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8960988","url":null,"abstract":"A new algorithm for matching finger or palm prints is presented for use where the full hand is considered as a biometric and only parts may be available in images for comparison. The algorithm uses an extended version of the minutiae-based approach treating the pattern as a graph of minutiae-like points. The procedure to identify minutiae-like points uses Gabor filtering, edge detection and thinning and following line patterns. A set of such points is subjected to Delaunay triangulation yielding a starting set of base-triangles for matching. There can be multiple matches of such triangles between the template and test - as similar triangles with a tolerance in the angles. Graphs are then grown to 5 and more nodes as long as a match can be found, until the maximum size matching graph is obtained. If the test matches a significant part of the template, the maximum order of graph matched will be high. The matching process is robust to transformations such as rotation, translation and scale changes. It can be applied to any part of the hand provided minutiae-like points are identifiable prior to the matching steps. The algorithm is tested using 158 fingerprint images from FVC 2002 DB1. 100 genuine and 5048 impostor scores are generated from 46 templates and 112 testing images. It had an EER of about 6%. It proves the principle behind the methodology and demonstrates that the method can be effective with degraded fingerprint images and is robust to similarity transformations present in the data. It can be applied for forensic fingerprint matching from the palm or parts other than the fingertips. By using multiple parts and multiple templates, the accuracy of the method will be improved with fusion in future versions of the algorithm.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"123 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132090189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Use of Moiré Patterns in Camera Position Estimation 在相机位置估计中的应用
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961010
Samuel Banks, R. Green, Jun Junghyun
{"title":"Use of Moiré Patterns in Camera Position Estimation","authors":"Samuel Banks, R. Green, Jun Junghyun","doi":"10.1109/IVCNZ48456.2019.8961010","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961010","url":null,"abstract":"Previous work has gone towards using moire patterns formed with lenticular lenses to perform pose estimation for short ranges. Current work has investigated existing theory of moire patterns, most notably the Fourier and first harmonic approximation models. These models have been developed in this paper to be able to model moire patterns generated from two patterns in 3D space that are separated. The accuracy of these models has been subjected to real-world tests for camera distance and lateral translation estimation. Preliminary tests from varying camera lateral translation appeared to be accurate for the close-range testing, with about 10 mm accuracy from 160 mm. Results from varying camera distance showed promise, however, they weren’t nearly as accurate as anticipated at 0–130 mm accuracy from ranges 100– 2000 mm.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117010777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image Correction with Curvature and Geometric Wavefront Sensors in Simulation and On-sky 曲率和几何波前传感器在仿真和天空中的图像校正
2019 International Conference on Image and Vision Computing New Zealand (IVCNZ) Pub Date : 2019-12-01 DOI: 10.1109/IVCNZ48456.2019.8961022
Sierra Hickman, S. Weddell, R. Clare
{"title":"Image Correction with Curvature and Geometric Wavefront Sensors in Simulation and On-sky","authors":"Sierra Hickman, S. Weddell, R. Clare","doi":"10.1109/IVCNZ48456.2019.8961022","DOIUrl":"https://doi.org/10.1109/IVCNZ48456.2019.8961022","url":null,"abstract":"Images of astronomical objects are distorted by the turbulence in Earth’s atmosphere. Deconvolution from Wavefront Sensing (DWFS) is a computer post-processing technique used by astronomers to reduce the effects of the atmosphere from images collected by ground-based telescopes. This paper investigates the relative performance of DWFS from two Wavefront Sensors (WFS), the curvature and geometric, used to estimate the aberrations introduced to the optical path by the turbulence. DWFS is performed using both the Wiener filter and the Lucy-Richardson algorithm. Our results show the geometric WFS has superior performance over the curvature WFS in both simulation and on-sky.","PeriodicalId":217359,"journal":{"name":"2019 International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115164610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信