A. Léger, G. le Goic, E. Fauvet, D. Fofi, Rémi Kornalewski
{"title":"R-CNN based automated visual inspection system for engine parts quality assessment","authors":"A. Léger, G. le Goic, E. Fauvet, D. Fofi, Rémi Kornalewski","doi":"10.1117/12.2586575","DOIUrl":"https://doi.org/10.1117/12.2586575","url":null,"abstract":"In this paper, we attempt to answer to a quality control problem in the context of an industrial serial production of lower plates (wheel suspensions) for the automotive industry. These frame parts are produced by a 2000-ton stamping machine that can reach 1800 parts per hour. The quality of these parts is assessed by a visual quality control operation. This operation is time-consuming. Moreover, many factors can affect its performance, as the attention of the operators in charge, or a too rapid inspection completion time, and non-detection defects lead to high supplementary costs. To answer this issue and automate this process operation, a system based on a vision system coupled to a pre-trained Convolutional Neural Networks (Mask R-CNN)1 has been designed and implemented. In addition, an artificial enlargement of the reference image base is proposed to improve the robustness of the identification, and reduce the sensitivity of the results to potential imaging artefacts due to non-controlled environments factors such as overexposure, blur, shadows or oil fog.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134300439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic recognition of parasitic bee species using wing vein shape","authors":"Yusuke Misaki, K. Terada","doi":"10.1117/12.2589061","DOIUrl":"https://doi.org/10.1117/12.2589061","url":null,"abstract":"Accurate species identification of parasitic bees is needed for studies of bio-pesticide use and habitat investigation. Therefore, we use differences in the shape of the wing veins between parasitic bee species. In this paper, we build a system that can automatically recognize species based on the extracted features from wing vein images.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"02 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127194689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An intelligent traffic control systems using on-road cameras","authors":"Md Mehedi Hassan, S. Karungaru, K. Terada","doi":"10.1117/12.2589192","DOIUrl":"https://doi.org/10.1117/12.2589192","url":null,"abstract":"The main reasons behind the traffic jam and accidents are illegal/double parking, over-speeding, violating signal lights, construction, wrong-way driving, reckless driving, unsafe lane changing, etc. To determines the problems and the solutions, this study proposes a two-step approach. One is data collection and the other is Optimization. In the data collection part, traffic information is obtained from various traffic information units through cameras installed in traffic signals and roads. By analyzing the collected data, the systems can take the next step in the optimization part. In the collected data, it is required to detect vehicles, pedestrians, and lanes. Yolov3 method was used for vehicles and pedestrians’ detection. For lane detection, the Hough transform was used. The main goal of the research is not detecting objects but to determine the intelligent systems which can combine all the collected data and give the optimum solutions to control the traffic signals depending on the situations. The study found that sometimes the vehicles are unnecessarily waiting for the signals. If the unnecessary time could be saved through signals then it would reduce time consumption, oil consumption, and mental impatience. The result would give us opportunities to reduce accidents, pollution, money, and time. Moreover, the systems can measure the speed which helps find out rule violating vehicles. This paper also proposed a method for shortest path calculation. In addition, automatic penalty execution can be carried out through the collected data. For that number plate recognition is included in future works.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132066521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Zendagui, G. L. Le Goïc, H. Chatoux, Jean-Baptiste Thomas, Y. Castro, M. Nurit, A. Mansouri
{"title":"Quality assessment of dynamic virtual relighting from RTI data: application to the inspection of engineering surfaces","authors":"A. Zendagui, G. L. Le Goïc, H. Chatoux, Jean-Baptiste Thomas, Y. Castro, M. Nurit, A. Mansouri","doi":"10.1117/12.2589178","DOIUrl":"https://doi.org/10.1117/12.2589178","url":null,"abstract":"This paper aims to evaluate the visual quality of the dynamic relighting of manufactured surfaces from Reflectance Transformation Imaging acquisitions. The first part of the study aimed to define the optimum parameters of acquisition using the RTI system: Exposure time, Gain, Sampling density. The second part is the psychometric experiment using the Design of Experiments approach. The results of this study help us to determine the influence of the parameters associated with the acquisition of Reflectance Transformation Imaging data, the models associated with relighting, and the dynamic perception of the resulting videos","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121529512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Tiny range image sensors using multiple laser lights for short distance measurement","authors":"Tomoaki Fukuda, Yonghoon Ji, K. Umeda","doi":"10.1117/12.2589196","DOIUrl":"https://doi.org/10.1117/12.2589196","url":null,"abstract":"This paper presents very compact range image sensors for short distance measurement, which is suitable for robot hands, etc. Robot manipulation such as grasping is one of the applications that require a range image sensor to obtain threedimensional (3D) information of the target object. For such applications, it is necessary to avoid the occlusion by a robot manipulator or a robot hand while measurement, and it is effective to attach a sensor to the robot hand for the avoidance. For this aim, a range sensor that is small enough and can measure at the short distance is required. Two sensors are constructed in this paper: one uses a multi-slit laser projector and the other uses a multi-spot laser projector. A small laser projector and a small camera is combined and range images are obtained in real time using the principle of active stereo. Appropriate methods to obtain range image are proposed for both sensors, and especially for the one with a multislit laser projector, a method to use both disparity and the intensity of laser light image is presented. The effectiveness of the proposed sensors is verified through short-range object measurement experiments.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114973245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Personal authentication and recognition of aerial input Hiragana using deep neural network","authors":"H. Mimura, Momoyo Ito, S. Ito, M. Fukumi","doi":"10.1117/12.2585333","DOIUrl":"https://doi.org/10.1117/12.2585333","url":null,"abstract":"We use Leap Motion and a deep neural network to perform personal authentication and character recognition of all hiragana characters entered in the air. We use Leap Motion to detect the index finger and store its trajectory as time series data. The input data was pre-processed to unify the data length by linear interpolation. For identification, the accuracy of Long Short Term Memory (LSTM) was compared with Support Vector Machine (SVM). As a result, SVM and LSTM achieved 97.25% and 98.18% F-measure in character recognition, respectively. In personal authentication, SVM has an accuracy of 92.45%, False Acceptance Rate (FAR) was 0.73%, and False Rejection Rate (FRR) was 41.59%. On the other hand, LSTM had an accuracy of 96.13%, FAR of 1.73% and FRR of 14.55%. Overall, the LSTM performed better than the SVM.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123321413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tsuyoshi Shimizu, Yasutake Haramiishi, Y. A. Rahim, Syamir Alihan, Yuji Kobayashi, A. Matsui, Shinji Kotani, H. Watanabe
{"title":"Evaluation of shot peening machined surface by image processing","authors":"Tsuyoshi Shimizu, Yasutake Haramiishi, Y. A. Rahim, Syamir Alihan, Yuji Kobayashi, A. Matsui, Shinji Kotani, H. Watanabe","doi":"10.1117/12.2585193","DOIUrl":"https://doi.org/10.1117/12.2585193","url":null,"abstract":"This paper describes an evaluation method of shot peened surface using image processing. Shot peening is a process that applies compressive residual stress to the product surface, and its evaluation is performed visually by an expert. If visual inspection can be replaced with image processing, the inspection of the entire product will be easier. Therefore, first reference samples that experts evaluate are prepared, next these samples are evaluated by image processing. relationship between expert evaluations and image processing evaluations are compared and the estimation function is defined using gausian distribution. Unknown processed surfaces are evaluated as a classification problem. For image processing, after binarization and labeling, the number of labels and the area ratio of binarization are used.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124175038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Training PointNet for human point cloud segmentation with 3D meshes","authors":"Takuma Ueshima, Katsuya Hotta, Shogo Tokai, Chao Zhang","doi":"10.1117/12.2589075","DOIUrl":"https://doi.org/10.1117/12.2589075","url":null,"abstract":"PointNet, which enables end-to-end learning for scattered/unordered point data, is a popular neural network architecture. However, in many applications, large amounts of complete point clouds are hardly available for non-rigid objects such as the human body. To generate the training data of PointNet, in this study, we propose to generate human body point clouds of various postures by uniformly sampling point clouds from meshes with respect to multiple human mesh model datasets. Experiments show that the model trained with the point clouds generated from mesh data is effective in the task of human body segmentation.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131460696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Manuel Kaufmann, V. Volland, Yifei Chen, I. Effenberger, C. Veyhl
{"title":"Scan quality estimation for industrial computed tomography using convolutional neural networks","authors":"Manuel Kaufmann, V. Volland, Yifei Chen, I. Effenberger, C. Veyhl","doi":"10.1117/12.2588990","DOIUrl":"https://doi.org/10.1117/12.2588990","url":null,"abstract":"Artefacts in industrial Computed Tomography (CT) compromise the image quality of a CT scan and deteriorate evaluations such as inspections for material defects or dimensional measurements. Due to a large variety of scanning objects made of different materials and of various part sizes, artefacts appear in various manifestations in the reconstructed image. Existing analytical approaches allow quantifying the CT scan quality, but still a lack of generalizability exists. Thus, assessing the scan quality is complex and error-prone, as an inappropriate set of analytical quality metrics might be considered for a certain scan setup. In our work, a scan quality estimation based on a Convolutional Neural Network (CNN) is proposed. In order to train the network, projection images of various scans are used. The reconstructed scans are labeled in a pairwise comparison by an experienced user regarding their image quality. A scalar quality value is assigned to every projection image to assess the quality. The network is deployed to perform regression for the quality value. The network is trained on multiple objects that cover the range of objects which can be sufficiently acquired with the used CT scanner. In order to enrich the features from scans of different qualities, each object is captured with various scanning parameters. Our work showed a test accuracy of approximately 80 % on prior unseen data and of up to 95 % on trained objects. In order to comprehend the black box approach incorporated by the trained CNN, visualizations of feature maps are analyzed, as regions in the projection images relevant for the quality estimation are highlighted.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131142191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supporting sports instruction with comparative display of forms","authors":"Sota Akiyama, Nobuyuki Umezu","doi":"10.1117/12.2589101","DOIUrl":"https://doi.org/10.1117/12.2589101","url":null,"abstract":"In this research, we propose a method to support sports instruction by displaying visualized comparison between a model motion and that of the player. Human pose estimation with the OpenPose framework based on deep learning technologies is first applied to both model and user motions. The motion difference between the model and the user is visualized with their trajectories superimposed on those two motions. To help users improve their motions and forms, our system displays where the specific body part, for example, the elbow of the throwing arm, should be located at a specific time, computed from the location and timing in the model motion. Our system also presents changes of velocity and position of body parts over time with line graphs for supposing users better understand that motion. Our current implementation is designed for pitching motions, and several other motions is included in our future work. We conducted user experiments where four participants used our system to improve their pitching forms. After recording pitching motion of these four participants, they were divided into two groups. First two participants were given the visualizations of our system while other two were presented only a pair of motions from a model and their own. The participants then performed their improved pitching motions again in front of our system to evaluate which of these two groups showed more refinements. The experimental results have shown that their forms are improved in both groups, and the usability of the proposed system has not successfully tested. Future work includes conducting larger-scale experiments, automatic selection of model motions for each user, as well as extending our method to other motions in different sports.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115260929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}