International Conference on Quality Control by Artificial Vision最新文献

筛选
英文 中文
R-CNN based automated visual inspection system for engine parts quality assessment 基于R-CNN的发动机部件质量评估自动目视检测系统
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2586575
A. Léger, G. le Goic, E. Fauvet, D. Fofi, Rémi Kornalewski
{"title":"R-CNN based automated visual inspection system for engine parts quality assessment","authors":"A. Léger, G. le Goic, E. Fauvet, D. Fofi, Rémi Kornalewski","doi":"10.1117/12.2586575","DOIUrl":"https://doi.org/10.1117/12.2586575","url":null,"abstract":"In this paper, we attempt to answer to a quality control problem in the context of an industrial serial production of lower plates (wheel suspensions) for the automotive industry. These frame parts are produced by a 2000-ton stamping machine that can reach 1800 parts per hour. The quality of these parts is assessed by a visual quality control operation. This operation is time-consuming. Moreover, many factors can affect its performance, as the attention of the operators in charge, or a too rapid inspection completion time, and non-detection defects lead to high supplementary costs. To answer this issue and automate this process operation, a system based on a vision system coupled to a pre-trained Convolutional Neural Networks (Mask R-CNN)1 has been designed and implemented. In addition, an artificial enlargement of the reference image base is proposed to improve the robustness of the identification, and reduce the sensitivity of the results to potential imaging artefacts due to non-controlled environments factors such as overexposure, blur, shadows or oil fog.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134300439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Automatic recognition of parasitic bee species using wing vein shape 利用翅脉形状自动识别寄生蜂种类
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589061
Yusuke Misaki, K. Terada
{"title":"Automatic recognition of parasitic bee species using wing vein shape","authors":"Yusuke Misaki, K. Terada","doi":"10.1117/12.2589061","DOIUrl":"https://doi.org/10.1117/12.2589061","url":null,"abstract":"Accurate species identification of parasitic bees is needed for studies of bio-pesticide use and habitat investigation. Therefore, we use differences in the shape of the wing veins between parasitic bee species. In this paper, we build a system that can automatically recognize species based on the extracted features from wing vein images.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"02 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127194689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An intelligent traffic control systems using on-road cameras 一种采用道路摄像头的智能交通控制系统
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589192
Md Mehedi Hassan, S. Karungaru, K. Terada
{"title":"An intelligent traffic control systems using on-road cameras","authors":"Md Mehedi Hassan, S. Karungaru, K. Terada","doi":"10.1117/12.2589192","DOIUrl":"https://doi.org/10.1117/12.2589192","url":null,"abstract":"The main reasons behind the traffic jam and accidents are illegal/double parking, over-speeding, violating signal lights, construction, wrong-way driving, reckless driving, unsafe lane changing, etc. To determines the problems and the solutions, this study proposes a two-step approach. One is data collection and the other is Optimization. In the data collection part, traffic information is obtained from various traffic information units through cameras installed in traffic signals and roads. By analyzing the collected data, the systems can take the next step in the optimization part. In the collected data, it is required to detect vehicles, pedestrians, and lanes. Yolov3 method was used for vehicles and pedestrians’ detection. For lane detection, the Hough transform was used. The main goal of the research is not detecting objects but to determine the intelligent systems which can combine all the collected data and give the optimum solutions to control the traffic signals depending on the situations. The study found that sometimes the vehicles are unnecessarily waiting for the signals. If the unnecessary time could be saved through signals then it would reduce time consumption, oil consumption, and mental impatience. The result would give us opportunities to reduce accidents, pollution, money, and time. Moreover, the systems can measure the speed which helps find out rule violating vehicles. This paper also proposed a method for shortest path calculation. In addition, automatic penalty execution can be carried out through the collected data. For that number plate recognition is included in future works.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132066521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality assessment of dynamic virtual relighting from RTI data: application to the inspection of engineering surfaces 基于RTI数据的动态虚拟重照质量评估:在工程表面检测中的应用
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589178
A. Zendagui, G. L. Le Goïc, H. Chatoux, Jean-Baptiste Thomas, Y. Castro, M. Nurit, A. Mansouri
{"title":"Quality assessment of dynamic virtual relighting from RTI data: application to the inspection of engineering surfaces","authors":"A. Zendagui, G. L. Le Goïc, H. Chatoux, Jean-Baptiste Thomas, Y. Castro, M. Nurit, A. Mansouri","doi":"10.1117/12.2589178","DOIUrl":"https://doi.org/10.1117/12.2589178","url":null,"abstract":"This paper aims to evaluate the visual quality of the dynamic relighting of manufactured surfaces from Reflectance Transformation Imaging acquisitions. The first part of the study aimed to define the optimum parameters of acquisition using the RTI system: Exposure time, Gain, Sampling density. The second part is the psychometric experiment using the Design of Experiments approach. The results of this study help us to determine the influence of the parameters associated with the acquisition of Reflectance Transformation Imaging data, the models associated with relighting, and the dynamic perception of the resulting videos","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121529512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Tiny range image sensors using multiple laser lights for short distance measurement 使用多个激光光源进行短距离测量的微型范围图像传感器
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589196
Tomoaki Fukuda, Yonghoon Ji, K. Umeda
{"title":"Tiny range image sensors using multiple laser lights for short distance measurement","authors":"Tomoaki Fukuda, Yonghoon Ji, K. Umeda","doi":"10.1117/12.2589196","DOIUrl":"https://doi.org/10.1117/12.2589196","url":null,"abstract":"This paper presents very compact range image sensors for short distance measurement, which is suitable for robot hands, etc. Robot manipulation such as grasping is one of the applications that require a range image sensor to obtain threedimensional (3D) information of the target object. For such applications, it is necessary to avoid the occlusion by a robot manipulator or a robot hand while measurement, and it is effective to attach a sensor to the robot hand for the avoidance. For this aim, a range sensor that is small enough and can measure at the short distance is required. Two sensors are constructed in this paper: one uses a multi-slit laser projector and the other uses a multi-spot laser projector. A small laser projector and a small camera is combined and range images are obtained in real time using the principle of active stereo. Appropriate methods to obtain range image are proposed for both sensors, and especially for the one with a multislit laser projector, a method to use both disparity and the intensity of laser light image is presented. The effectiveness of the proposed sensors is verified through short-range object measurement experiments.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114973245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Personal authentication and recognition of aerial input Hiragana using deep neural network 基于深度神经网络的航拍输入平假名的个人认证与识别
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2585333
H. Mimura, Momoyo Ito, S. Ito, M. Fukumi
{"title":"Personal authentication and recognition of aerial input Hiragana using deep neural network","authors":"H. Mimura, Momoyo Ito, S. Ito, M. Fukumi","doi":"10.1117/12.2585333","DOIUrl":"https://doi.org/10.1117/12.2585333","url":null,"abstract":"We use Leap Motion and a deep neural network to perform personal authentication and character recognition of all hiragana characters entered in the air. We use Leap Motion to detect the index finger and store its trajectory as time series data. The input data was pre-processed to unify the data length by linear interpolation. For identification, the accuracy of Long Short Term Memory (LSTM) was compared with Support Vector Machine (SVM). As a result, SVM and LSTM achieved 97.25% and 98.18% F-measure in character recognition, respectively. In personal authentication, SVM has an accuracy of 92.45%, False Acceptance Rate (FAR) was 0.73%, and False Rejection Rate (FRR) was 41.59%. On the other hand, LSTM had an accuracy of 96.13%, FAR of 1.73% and FRR of 14.55%. Overall, the LSTM performed better than the SVM.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123321413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of shot peening machined surface by image processing 喷丸强化加工表面的图像评价
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2585193
Tsuyoshi Shimizu, Yasutake Haramiishi, Y. A. Rahim, Syamir Alihan, Yuji Kobayashi, A. Matsui, Shinji Kotani, H. Watanabe
{"title":"Evaluation of shot peening machined surface by image processing","authors":"Tsuyoshi Shimizu, Yasutake Haramiishi, Y. A. Rahim, Syamir Alihan, Yuji Kobayashi, A. Matsui, Shinji Kotani, H. Watanabe","doi":"10.1117/12.2585193","DOIUrl":"https://doi.org/10.1117/12.2585193","url":null,"abstract":"This paper describes an evaluation method of shot peened surface using image processing. Shot peening is a process that applies compressive residual stress to the product surface, and its evaluation is performed visually by an expert. If visual inspection can be replaced with image processing, the inspection of the entire product will be easier. Therefore, first reference samples that experts evaluate are prepared, next these samples are evaluated by image processing. relationship between expert evaluations and image processing evaluations are compared and the estimation function is defined using gausian distribution. Unknown processed surfaces are evaluated as a classification problem. For image processing, after binarization and labeling, the number of labels and the area ratio of binarization are used.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124175038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training PointNet for human point cloud segmentation with 3D meshes 训练PointNet的人类点云分割与3D网格
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589075
Takuma Ueshima, Katsuya Hotta, Shogo Tokai, Chao Zhang
{"title":"Training PointNet for human point cloud segmentation with 3D meshes","authors":"Takuma Ueshima, Katsuya Hotta, Shogo Tokai, Chao Zhang","doi":"10.1117/12.2589075","DOIUrl":"https://doi.org/10.1117/12.2589075","url":null,"abstract":"PointNet, which enables end-to-end learning for scattered/unordered point data, is a popular neural network architecture. However, in many applications, large amounts of complete point clouds are hardly available for non-rigid objects such as the human body. To generate the training data of PointNet, in this study, we propose to generate human body point clouds of various postures by uniformly sampling point clouds from meshes with respect to multiple human mesh model datasets. Experiments show that the model trained with the point clouds generated from mesh data is effective in the task of human body segmentation.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131460696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Scan quality estimation for industrial computed tomography using convolutional neural networks 基于卷积神经网络的工业计算机断层扫描质量估计
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2588990
Manuel Kaufmann, V. Volland, Yifei Chen, I. Effenberger, C. Veyhl
{"title":"Scan quality estimation for industrial computed tomography using convolutional neural networks","authors":"Manuel Kaufmann, V. Volland, Yifei Chen, I. Effenberger, C. Veyhl","doi":"10.1117/12.2588990","DOIUrl":"https://doi.org/10.1117/12.2588990","url":null,"abstract":"Artefacts in industrial Computed Tomography (CT) compromise the image quality of a CT scan and deteriorate evaluations such as inspections for material defects or dimensional measurements. Due to a large variety of scanning objects made of different materials and of various part sizes, artefacts appear in various manifestations in the reconstructed image. Existing analytical approaches allow quantifying the CT scan quality, but still a lack of generalizability exists. Thus, assessing the scan quality is complex and error-prone, as an inappropriate set of analytical quality metrics might be considered for a certain scan setup. In our work, a scan quality estimation based on a Convolutional Neural Network (CNN) is proposed. In order to train the network, projection images of various scans are used. The reconstructed scans are labeled in a pairwise comparison by an experienced user regarding their image quality. A scalar quality value is assigned to every projection image to assess the quality. The network is deployed to perform regression for the quality value. The network is trained on multiple objects that cover the range of objects which can be sufficiently acquired with the used CT scanner. In order to enrich the features from scans of different qualities, each object is captured with various scanning parameters. Our work showed a test accuracy of approximately 80 % on prior unseen data and of up to 95 % on trained objects. In order to comprehend the black box approach incorporated by the trained CNN, visualizations of feature maps are analyzed, as regions in the projection images relevant for the quality estimation are highlighted.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131142191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Supporting sports instruction with comparative display of forms 以形式对比展示辅助体育教学
International Conference on Quality Control by Artificial Vision Pub Date : 2021-07-16 DOI: 10.1117/12.2589101
Sota Akiyama, Nobuyuki Umezu
{"title":"Supporting sports instruction with comparative display of forms","authors":"Sota Akiyama, Nobuyuki Umezu","doi":"10.1117/12.2589101","DOIUrl":"https://doi.org/10.1117/12.2589101","url":null,"abstract":"In this research, we propose a method to support sports instruction by displaying visualized comparison between a model motion and that of the player. Human pose estimation with the OpenPose framework based on deep learning technologies is first applied to both model and user motions. The motion difference between the model and the user is visualized with their trajectories superimposed on those two motions. To help users improve their motions and forms, our system displays where the specific body part, for example, the elbow of the throwing arm, should be located at a specific time, computed from the location and timing in the model motion. Our system also presents changes of velocity and position of body parts over time with line graphs for supposing users better understand that motion. Our current implementation is designed for pitching motions, and several other motions is included in our future work. We conducted user experiments where four participants used our system to improve their pitching forms. After recording pitching motion of these four participants, they were divided into two groups. First two participants were given the visualizations of our system while other two were presented only a pair of motions from a model and their own. The participants then performed their improved pitching motions again in front of our system to evaluate which of these two groups showed more refinements. The experimental results have shown that their forms are improved in both groups, and the usability of the proposed system has not successfully tested. Future work includes conducting larger-scale experiments, automatic selection of model motions for each user, as well as extending our method to other motions in different sports.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115260929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信