{"title":"Webcam-based categorization of task engagement of PC users at work","authors":"T. Ohara, Nobuyuki Umezu","doi":"10.1117/12.2589104","DOIUrl":"https://doi.org/10.1117/12.2589104","url":null,"abstract":"In this paper, we propose a support method for PC users to monitor their task engagement. Our approach is based on the number of keyboard strokes and pixel changes on the screen, and images from an ordinary webcam equipped with a PC. With our system, supervisors or teachers would improve their quality of guidance and instructions given at the right moment when workers or students require some support due to reasons such as slow progress and technical difficulties. In conventional methods, a special device such as an acceleration sensor for each individual is often required to acquire information on one’s working status and body movements, which is difficult to deploy in a real environment due to its cost for sensors. A face detection method based on Deep Neural Network, such as SSD, allows as to implement a cheaper system using an ordinary web camera. We calculate the average difference between two grabbed frames from the user’s screen to estimate the amount of screen changes between a given time interval. The number of key strokes typed by the user is another factor to estimate their task engagement. These factors are used to categorize the work mode of users. We use the K-Means method based on the Euclidean distance to cluster the recorded factors to determine thresholds for task categorization. We conducted experiments seven participants to evaluate the accuracy of our categorization method. Every participant is asked to categorize 15 scenes into four work modes. A scene includes a camera image with the PC user’s face, the screenshot path the moment, and the number of key strokes. The results from these participants are then compared with those of our system that categorized the same scenes with the thresholds on three factors. Approximately only 60% of these results matched each other, where we have enough room to improve our approach. Future work includes the selection of features that are far more effective for categorization, a better estimation of pixel changes on the PC screen, and evaluation experiments with more participants.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134288356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Assya Boughrara, Igor Jovančević, Hamdi Ben Abdallah, B. Dolives, Mathieu Belloc, J. Orteu
{"title":"Inspection of mechanical assemblies based on 3D deep learning approaches","authors":"Assya Boughrara, Igor Jovančević, Hamdi Ben Abdallah, B. Dolives, Mathieu Belloc, J. Orteu","doi":"10.1117/12.2588986","DOIUrl":"https://doi.org/10.1117/12.2588986","url":null,"abstract":"Our research work is being carried out within the framework of the joint research laboratory ”Inspection 4.0” between IMT Mines Albi/ICA and the company DIOTA specialized in the development of numerical tools for Industry 4.0. In this work, we are focused on conformity control of complex aeronautical mechanical assemblies, typically an aircraft engine at the end or in the middle of the assembly process. A 3D scanner carried by a robot arm provides acquisitions of 3D point clouds which are further processed by deep classification networks. Computer Aided Design (CAD) model of the mechanical assembly to be inspected is available, which is an important asset of our approach. Our deep learning models are trained on synthetic and simulated data, generated from the CAD models. Several networks are trained and evaluated and results on real clouds are presented.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124147229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"You don't drink a cupboard: improving egocentric action recognition with co-occurrence of verbs and nouns","authors":"Hiroki Kojima, Naoshi Kaneko, Seiya Ito, K. Sumi","doi":"10.1117/12.2591298","DOIUrl":"https://doi.org/10.1117/12.2591298","url":null,"abstract":"We propose a refinement module to improve action recognition by considering the semantic relevance between verbs and nouns. Existing methods recognize actions as a combination of verb and noun. However, they occasionally produce the semantically implausible combination, such as “drink a cupboard” or “open a carrot”. To tackle this problem, we propose a method that incorporates a word embedding model into an action recognition network. The word embedding model is trained to obtain co-occurrence between verbs and nouns and used to refine the initial class probabilities estimated by the network. Experimental results show that our method improves the estimation accuracy of verbs and nouns on the EPIC-KITCHENS Dataset.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124204671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dingyu Liu, Yusheng Wang, Yonghoon Ji, Hiroshi Tsuchiya, A. Yamashita, H. Asama
{"title":"Development of image simulator for forward-looking sonar using 3D rendering","authors":"Dingyu Liu, Yusheng Wang, Yonghoon Ji, Hiroshi Tsuchiya, A. Yamashita, H. Asama","doi":"10.1117/12.2590004","DOIUrl":"https://doi.org/10.1117/12.2590004","url":null,"abstract":"This paper proposes an efficient imaging sonar simulation method based on 3D modeling. In underwater scenarios, a forward-looking sonar, which is also known as an acoustic camera, outperforms other sensors including popular optical cameras, for it is resistant to turbidity and weak illumination, which are typical in underwater environments, and thus able to provide accurate information of the environments. For those underwater tasks highly automated along with artificial intelligence and computer vision, the development of the acoustic image simulator can provide support by reproducing the environment and generating synthetic acoustic images. It can also facilitate researchers to tackle the scarcity of real underwater data in some theoretical studies. In this paper, we make use of the 3D modeling technique to simulate the underwater scenarios and the flexible automated control of the acoustic camera and objects in the scenarios. The simulation results and the comparison to real acoustic images demonstrate that the proposed simulator can generate accurate synthetic acoustic images efficiently and flexibly.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132158935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Kondo, Y. Yamaguchi, H. Saito, I. Yoshida, M. Numada, H. Koshimizu
{"title":"Verification of denoising performance of edge-preserving noise reduction filter using fast M-estimation method","authors":"Y. Kondo, Y. Yamaguchi, H. Saito, I. Yoshida, M. Numada, H. Koshimizu","doi":"10.1117/12.2592763","DOIUrl":"https://doi.org/10.1117/12.2592763","url":null,"abstract":"Random noise injures both the basic image quality and also the following image processing procedures. The low-pass filter is commonly used as the image denoising. Low-pass filter can reduce noise; however, the edge becomes always blur as the side effect. In order to suppress this side effect, we proposed edge preserving noise reduction filter using Fast Mestimation method. As the Proposed method is applied experimentally to the noisy image, it was clarified that the noise was clearly reduced and the performance of edge preserving was realized at the same time. In this study, a quantitative evaluation of the denoising performance of the proposed method is obtained by varying the amount of noise applied and obtaining the denoising ratio.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132296763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Active learning using weakly supervised signals for quality inspection","authors":"Antoine Cordier, Deepan Das, Pierre Gutierrez","doi":"10.1117/12.2586595","DOIUrl":"https://doi.org/10.1117/12.2586595","url":null,"abstract":"Because manufacturing processes evolve fast and production visual aspect can vary significantly on a daily basis, the ability to rapidly update machine vision based inspection systems is paramount. Unfortunately, supervised learning of convolutional neural networks requires a significant amount of annotated images in order to learn effectively from new data. Acknowledging the abundance of continuously generated images coming from the production line and the cost of their annotation, we demonstrate it is possible to prioritize and accelerate the annotation process. In this work, we develop a methodology for learning actively,1 from rapidly mined, weakly (i.e. partially) annotated data, enabling a fast, direct feedback from the operators on the production line and tackling a big machine vision weakness: false positives. These may arise with covariate shift, which happens inevitably due to changing conditions of the data acquisition setup. In that regard, we show domain-adversarial training2 to be an efficient way to address this issue.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123407940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pierre Gutierrez, Maria Luschkova, Antoine Cordier, Mustafa Shukor, Mona Schappert, Tim Dahmen
{"title":"Synthetic training data generation for deep learning based quality inspection","authors":"Pierre Gutierrez, Maria Luschkova, Antoine Cordier, Mustafa Shukor, Mona Schappert, Tim Dahmen","doi":"10.1117/12.2586824","DOIUrl":"https://doi.org/10.1117/12.2586824","url":null,"abstract":"Deep learning is now the gold standard in computer vision-based quality inspection systems. In order to detect defects, supervised learning is often utilized, but necessitates a large amount of annotated images, which can be costly: collecting, cleaning, and annotating the data is tedious and limits the speed at which a system can be deployed as everything the system must detect needs to be observed first. This can impede the inspection of rare defects, since very few samples can be collected by the manufacturer. In this work, we focus on simulations to solve this issue. We first present a generic simulation pipeline to render images of defective or healthy (non defective) parts. As metallic parts can be highly textured with small defects like holes, we design a texture scanning and generation method. We assess the quality of the generated images by training deep learning networks and by testing them on real data from a manufacturer. We demonstrate that we can achieve encouraging results on real defect detection using purely simulated data. Additionally, we are able to improve global performances by concatenating simulated and real data, showing that simulations can complement real images to boost performances. Lastly, using domain adaptation techniques helps improving slightly our final results.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121692138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Visual inspection for metallic surfaces: CNN driven by features","authors":"Riccardo Fantinel, A. Cenedese","doi":"10.1117/12.2521455","DOIUrl":"https://doi.org/10.1117/12.2521455","url":null,"abstract":"In this paper, an effective and novel automatic learning solution for the quality control of metallic objects surfaces is proposed, which can be seamlessly integrated into the industrial process. Such a system requires a coaxial illuminator to capture the object view with a single camera while lighting it with structured light: in this way, the object surface can be viewed in time as a dynamic scene under different illumination conditions. By relying on a linear model to describe the expected evolution of the light over the object of interest, the Residuals of Linear Evolution of Light (RLEL) algorithm is derived with the specific aim of identifying and characterizing anomalies and defects through the residuals of a least square approach. Then, a novel learning strategy is developed that exploits the model-based RLEL descriptor and thus promotes itself as an alternative strategy to the black box approach of Convolutional Neural Networks (CNNs). By combining both the data-driven and the model-based learning approaches to perform the inspection task, an Hybrid Learning (HL) procedure is defined: in a first phase, the HL exploits an Encoder-Decoder network to incorporate the model-based description while, in a second phase, it uses only the pre-trained encoder to drive the learning process of a 3D-CNN. In doing so, the proposed procedure reaches interesting results that exceed also the performance of state-of-the-art 3D-Inception and 3D-Residual networks.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127460136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automated vision system for crankshaft inspection using deep learning approaches","authors":"K. Tout, Mohamed Bouabdellah, C. Cudel, J. Urban","doi":"10.1117/12.2521751","DOIUrl":"https://doi.org/10.1117/12.2521751","url":null,"abstract":"This paper proposes a fully automated vision system to inspect the whole surface of crankshafts, based on the magnetic particle testing technique. Multiple cameras are needed to ensure the inspection of the whole surface of the crankshaft in real-time. Due to the very textured surface of crankshafts and the variability in defect shapes and types, defect detection methods based on deep learning algorithms, more precisely convolutional neural networks (CNNs), become a more efficient solution than traditional methods. This paper reviews the various approaches of defect detection with CNNs, and presents the advantages and weaknesses of each approach for real-time defect detection on crankshafts. It is important to note that the proposed visual inspection system only replaces the manual inspection of crankshafts conducted by operators at the end of the magnetic particle testing procedure.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125077364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
L. Foucault, N. Verrier, M. Debailleul, B. Simon, O. Haeberlé
{"title":"A simplified approach for tomographic diffractive microscopy","authors":"L. Foucault, N. Verrier, M. Debailleul, B. Simon, O. Haeberlé","doi":"10.1117/12.2521798","DOIUrl":"https://doi.org/10.1117/12.2521798","url":null,"abstract":"Tomographic diffractive microscopy (TDM) is an imaging technique, which allows for recording the refractive index of unlabelled transparent specimens. Based on diffraction theory, it can be implemented in transmission or in reflection. In this paper, a new TDM data acquisition and reconstruction method is proposed. The purpose is to use the mirror effect of a reflecting material to establish a double illumination system. Neglecting backward-diffracted fields, the setup reduces to a double-transmission TDM, which combined with an azimuthal rotation of the illumination, allows for faster and simplified acquisitions. We also point out a new demodulation method based only on Fourier transforms.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"80 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125920974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}