M. Al Assaad, S. Bazeille, T. Laurain, A. Dieterlen, C. Cudel
{"title":"Interest of pseudo-focused images for key-points detection in plenoptic imaging","authors":"M. Al Assaad, S. Bazeille, T. Laurain, A. Dieterlen, C. Cudel","doi":"10.1117/12.2588954","DOIUrl":"https://doi.org/10.1117/12.2588954","url":null,"abstract":"Light-Field (LF) cameras allow the extraction not only of the intensity of light but also of the direction of light rays in the scene, hence it records much more information of the scene than a conventional camera. In this paper, we present a novel method to detect key-points in raw LF images by applying key-points detectors on Pseudo-Focused images (PFIs). The main advantage of this method is that we don’t need to use complex key-points detectors dedicated to light-field images. We illustrate the method in two use cases: the extraction of corners in a checkerboard and the key-points matching in two view raw light-field images. These key-points can be used for different applications e.g. calibration, depth estimation or visual odometry. Our experiments showed that our method preserves the accuracy of detection by re-projecting the pixels in the original raw images.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"11794 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130024170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Meguenani, K. Tout, S. Kohler, S. Bazeille, J. Chambard, C. Cudel
{"title":"Deflectometry based on light-field imaging","authors":"A. Meguenani, K. Tout, S. Kohler, S. Bazeille, J. Chambard, C. Cudel","doi":"10.1117/12.2588988","DOIUrl":"https://doi.org/10.1117/12.2588988","url":null,"abstract":"This work presents how deflectometry can be coupled with a light-field camera to better characterize and quantify the depth of anomalies on specular surfaces. In our previous work,1 we proposed a new scanning scheme for the detection and 3D reconstruction of defects on reflective objects. However, the quality of the reconstruction was strongly dependent on the object-camera distance which was required as an external input parameter. In this paper, we propose a new approach that integrates an estimation of this distance into our system by replacing the standard camera with a light-field camera.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123267102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yutaka Kawashima, Mayuka Higo, T. Tokiwa, Yukihiro Asami, K. Nonaka, Y. Aoki
{"title":"Out-of-distribution detection for fungi images with similar features","authors":"Yutaka Kawashima, Mayuka Higo, T. Tokiwa, Yukihiro Asami, K. Nonaka, Y. Aoki","doi":"10.1117/12.2591725","DOIUrl":"https://doi.org/10.1117/12.2591725","url":null,"abstract":"In order to create a classification model for fungi, it is necessary to have robustness against out-of-distribution data from the viewpoint of practicality. Therefore, in this paper, we perform out-of-distribution detection on a fungi. Unlike the case of conventional out-of-distribution detection, the characteristics of in-distribution data and out-of-distribution data in this paper are very similar. Therefore, the problem in which conventional methods using out-of-distribution data for validation are not effective is mentioned. We also verify whether the accuracy of out-of-distribution detection can be improved using the attention branch network.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129433475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Nurit, G. L. Le Goïc, S. Maniglier, P. Jochum, H. Chatoux, A. Mansouri
{"title":"Improved visual saliency estimation on manufactured surfaces using high-dynamic reflectance transformation imaging","authors":"M. Nurit, G. L. Le Goïc, S. Maniglier, P. Jochum, H. Chatoux, A. Mansouri","doi":"10.1117/12.2589748","DOIUrl":"https://doi.org/10.1117/12.2589748","url":null,"abstract":"Reflectance Transformation Imaging (RTI) is a technique for estimating the surface local angular reflectance and characterizing the visual properties by varying lighting directions and capturing a set of stereo-photometric images. The proposed method, namely HD-RTI, is based on the coupling of RTI and HDR imaging techniques. The HD-RTI automatically optimizes the necessary exposure times for each angle of illumination by using the response of the scene. Our method is applied to industrial surfaces with micro-scratches from which we will estimate saliency information. Results show that coupling HDR and RTI enhance the characterization and therefore the discrimination on the surfaces visual saliency maps. It leads to an increase in robustness for visual quality assessment tasks.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122980298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abdelrahman G. Abubakr, Igor Jovančević, Nour Islam Mokhtari, Hamdi Ben Abdallah, J. Orteu
{"title":"On learning deep domain-invariant features from 2D synthetic images for industrial visual inspection","authors":"Abdelrahman G. Abubakr, Igor Jovančević, Nour Islam Mokhtari, Hamdi Ben Abdallah, J. Orteu","doi":"10.1117/12.2589040","DOIUrl":"https://doi.org/10.1117/12.2589040","url":null,"abstract":"Deep learning resulted in a huge advancement in computer vision. However, deep models require a large amount of manually annotated data, which is not easy to obtain, especially in a context of sensitive industries. Rendering of Computer Aided Design (CAD) models to generate synthetic training data could be an attractive workaround. This paper focuses on using Deep Convolutional Neural Networks (DCNN) for automatic industrial inspection of mechanical assemblies, where training images are limited and hard to collect. The ultimate goal of this work is to obtain a DCNN classification model trained on synthetic renders, and deploy it to verify the presence of target objects in never-seen-before real images collected by RGB cameras. Two approaches are adopted to close the domain gap between synthetic and real images. First, Domain Randomization technique is applied to generate synthetic data for training. Second, a novel approach is proposed to learn better features representations by means of self-supervision: we used an Augmented Auto-Encoder (AAE) and achieved results competitive to our baseline model trained on real images. In addition, this approach outperformed baseline results when the problem was simplified to binary classification for each object individually.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124770077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Identification of insect infiltration time by tissue slice","authors":"Y. Hashiguchi, K. Terada, K. Shinozaki, K. Miyama","doi":"10.1117/12.2588838","DOIUrl":"https://doi.org/10.1117/12.2588838","url":null,"abstract":"Since contamination of food products by insects can cause serious damage to manufacturers, it is necessary to determine the timing of contamination as a countermeasure. However, manual surveys are time consuming and expensive. In this paper, we propose a method for automatic identification of insect infiltration routes using image processing.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132165186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ryousuke Tsubaki, Takumi Toyoda, Kota Yoshida, Akio Nakamura
{"title":"Vision-based classification of mosquito species: data augmentation by background replacement for convolutional neural network-based species classification of smashed mosquitoes","authors":"Ryousuke Tsubaki, Takumi Toyoda, Kota Yoshida, Akio Nakamura","doi":"10.1117/12.2589100","DOIUrl":"https://doi.org/10.1117/12.2589100","url":null,"abstract":"This study proposes a method of data augmentation by background replacement for the species classification of smashed mosquitoes using convolutional neural networks (CNNs). To augment data from a limited number of images of smashed mosquitoes, varieties of foreground mosquito and background are ensured by clipping a foreground mosquito image and pasting it into different backgrounds. For the background images, a white image is prepared as the ideal background, and a hand palm image is assumed as the background for practical use. Images extracted from three publicly available datasets are also prepared, which are considered as the variable backgrounds. A CNN-based deep classification is used with three types of architecture, and the classification accuracy is compared using training images corresponding to different background conditions. The classification accuracy using training images with a variety of backgrounds is better than that with a white or palm background. Moreover, deep classification with a residual network achieves the highest classification accuracy. The results of this work show that the species classification of the smashed mosquitoes can be achieved by using datasets with the proposed data augmentation method.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115821522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Social distance measurement for indoor environments","authors":"E. Nadine-Erdene, S. Karungaru, K. Terada","doi":"10.1117/12.2589105","DOIUrl":"https://doi.org/10.1117/12.2589105","url":null,"abstract":"Social distancing is a suggested solution by many scientists, health care providers and researchers to reduce the spread of COVID-19 in public places. Over a year ago most countries have closed their borders, put people under lockdown, and have been suspending people from work and travel. However, there are still many organizations that need to operate, especially hospitals, services industry, governments, etc. However, people cannot maintain social distancing which includes staying at least 1.5 2 meters from other people because they need to communicate with each other. As a result, this increases the infection of Covid-19. This work proposes a social distancing tracking tool in offices or indoor places. We propose a YOLOv5-based Deep Neural Network (DNN) model to automate the process of monitoring the social distancing via object detection and tracking approaches. We detect office objects of known size and use it to estimate the social distance in real-time with the bounding boxes in indoor environments.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"11794 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130152896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Piano performance support system with a head-mounted display","authors":"Tomoyasu Kaji, K. Terada, S. Karungaru","doi":"10.1117/12.2589062","DOIUrl":"https://doi.org/10.1117/12.2589062","url":null,"abstract":"In this paper, we propose a method to analyze and support the performance status from the images of hands and keys obtained by installing a web camera on a head-mounted display (HMD).","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132013180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recipe recommendation and cooking instruction based on food material recognition","authors":"Saki Asahina, Nobuyuki Umezu","doi":"10.1117/12.2589081","DOIUrl":"https://doi.org/10.1117/12.2589081","url":null,"abstract":"As devices such as depth cameras and projectors become cheaper, various methods have been proposed to make people's work environment smarter with AR technologies. In a kitchen, there would be a wide variety of advantages in an intelligent work environment, because people have to manage so many things such as various ingredients, typology of procedures with many recipes, and complicated work environment by mixing materials and tools. In this research, we aim to support cooking by recommending recipes with an installed projector and camera in the kitchen and projecting an operation interface. Our system recognizes ingredients with a DNN-based method called Mask R-CNN. It also has gesture controls that offer users contactless operations based on a Kinect depth sensor. We present four popular recipes obtained from a famous site named the Rakuten recipe. Displayed information includes dish names, ingredients, cooking procedures, process photographs, and time to cook. A series of user experiments with 10 participants was conducted to evaluate the usability of gesture operation with Kinect of the proposed system. The distance between Kinect and the hands of the participants is 0.8 (m). Each participant is given one trial and uses gestures to select one of four recipes displayed on a iMac screen. We received high evaluations (average 4.2 to 4.5 on a 5-point scale) in the results of the experiment questionnaire. Future work includes integrating more functions into our system, such as estimating ingredient amount based on the areas of recognized materials with Mask R-CNN, and cooking process recognition.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130567428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}