{"title":"Unveiling the secrets of paintings: deep neural networks trained on high-resolution multispectral images for accurate attribution and authentication","authors":"Michael E. Sander, Tom Sander, Maxime Sylvestre","doi":"10.1117/12.3000286","DOIUrl":"https://doi.org/10.1117/12.3000286","url":null,"abstract":"Attribution and authentication of paintings are difficult tasks, often based on human expertise. In this work, we present SpectrumArt: a new dataset of multispectral (13 channels) image patches of paintings acquired at very high resolution (800 pixels per mm2 ). We train deep neural networks on SpectrumArt for attribution (i.e., authorship classification) and authentication (i.e., whether of undisputed origin). For attribution, we obtain an accuracy of 92% on a test set of patches coming from unseen paintings. We also propose two classification metrics for attribution of full paintings based on the prediction for the patches: majority vote and entropy weighted vote. Both metrics lead to an attribution score of 100% on unseen paintings. For authenticity testing, our model agrees with the experts’ conclusions on genuine and fake paintings, and provides new insights into the authenticity of paintings where the expert community is divided by proposing a spectral matching score between the painting and an artist. To validate the important advantage of our data collection method, we show that the use of 13 channels instead of 3 and the high resolution of the data significantly improve the accuracy of our models.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124835614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yukiya Taki, K. Kato, Kazunori Terada, Kensuke Tobitani
{"title":"Visual impression estimation system considering attribute information","authors":"Yukiya Taki, K. Kato, Kazunori Terada, Kensuke Tobitani","doi":"10.1117/12.2691716","DOIUrl":"https://doi.org/10.1117/12.2691716","url":null,"abstract":"Detailed identification of visual impressions of objects by attributes can be leveraged to develop products and improve customer satisfaction. In this study, we propose a method to estimate Kansei (affective) information for each attribute, which is the visual impression received from the image. For each attribute, we created a dataset with Kansei indices. By fine-tuning the created dataset to combine attribute information with the output of ResNet18 which was already trained with ImageNet to output indexes, we confirmed that the correlation coefficients for multiple item ratings were higher than those of a deep learning model without attribute information.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128653813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ichraq Lemghari, Sylvie Le-Hégarat, Emanuel Aldea, Jennifer Vandoni
{"title":"Handling noisy annotations in deep supervised learning","authors":"Ichraq Lemghari, Sylvie Le-Hégarat, Emanuel Aldea, Jennifer Vandoni","doi":"10.1117/12.2692547","DOIUrl":"https://doi.org/10.1117/12.2692547","url":null,"abstract":"Non-destructive testing (NDT) is employed by companies to assess the features of a material, in order to identify some variations or anomalies in its properties without causing any damage to the original object. In this context of industrial visual inspection, the help of new technologies and especially deep supervised learning is nowadays required to reach a very high level of performance. Data labelling, that is essential to reach such performance, may be fastidious and tricky, and only experts can provide the labelling of the material possible defects. Considering classification problems, this paper addresses the issue of handling noisy labels in datasets. We will first present the existing works related to the problem, our general idea of how to handle it, then we will present our proposed method in detail along with the obtained results that reach more than 0.96 and 0.88 of accuracy for noisified MNIST and CIFAR-10 respectively with a 40% noise ratio. Finally, we present some potential perspectives for future works.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"12749 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129932839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Al Assaad, S. Bazeille, Thomas Josso-Laurain, A. Dieterlen, C. Cudel
{"title":"Enhanced key-point detection for plenoptic imaging","authors":"M. Al Assaad, S. Bazeille, Thomas Josso-Laurain, A. Dieterlen, C. Cudel","doi":"10.1117/12.2692367","DOIUrl":"https://doi.org/10.1117/12.2692367","url":null,"abstract":"Standard imaging techniques do not get as much information from a scene as light-field imaging. Light-field (LF) cameras can measure the light intensity reflected by an object and, most importantly, the direction of its light rays. This information can be used in different applications, such as depth estimation, in-plane focusing, creating full-focused images, etc. However, standard key-point detectors often employed in computer vision applications cannot be applied directly to plenoptic images due to the nature of raw LF images. This work presents an approach for key-point detection dedicated to plenoptic images. Our method allows using of conventional key-point detector methods. It forces the detection of this key-point in a set of micro-images of the raw LF image. Obtaining this important number of key-points is essential for applications that require finding additional correspondences in the raw space, such as disparity estimation, indirect visual odometry techniques, and others. The approach is set to the test by modifying the Harris key-point detector.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129956504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Combining unsupervised and supervised deep learning approaches for surface anomaly detection","authors":"Domen Rački, Dejan Tomazevic, D. Skočaj","doi":"10.1117/12.2688559","DOIUrl":"https://doi.org/10.1117/12.2688559","url":null,"abstract":"Anomaly detection in an unsupervised manner has become the go-to approach in applications where data labeling proves problematic. However, these approaches aren’t completely unsupervised, since they rely on the weak knowledge of the dataset distribution into anomalous and anomaly-free subsets and typically require post-training threshold calibration in order to perform anomaly detection. Yet, they do not take advantage of available positive samples during training. In contrast, fully supervised approaches have proven to be more accurate and more efficient, however, they require a sufficient number of anomalous images to be labeled on a per-pixel level, which represents a labour-intensive task. In this paper, we propose a new hybrid approach that utilizes the best of both worlds. We use an unsupervised approach to build a model for generating pseudo labels, followed by a supervised approach in order to robustify anomaly detection. Moreover, we extend this approach with an active learning schema, that results in learning with mixed supervision. We achieve several improvements, i.e., the utilization of available positive image samples, improved anomaly detection performance, and the retention of real-time performance. The proposed approach yields results that are comparable to the fully supervised approach, and at the very least, reduces the number of required labeled anomalous samples.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117052944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"XRANet: an extra-wide, residual and attention-based deep convolutional neural network for semantic segmentation","authors":"Roger Booto Tokime, M. Akhloufi","doi":"10.1117/12.2692337","DOIUrl":"https://doi.org/10.1117/12.2692337","url":null,"abstract":"In this paper, we propose XRANet, a Deep Convolutional Neural Network (DNN) architecture for Semantic Segmentation. The recent advancements in deep learning and convolutional neural networks have greatly improved the accuracy of segmentation tasks. XRANet builds on the widely used U-Net architecture and adds several improvements to increase performance. The eXtra-wide mechanism in the encoder, combined with residual connections and an attention mechanism in both the encoder and decoder, enhances feature extraction and reduces the activation of pixels outside the regions of interest. The proposed architecture was evaluated on various public datasets, and the results were measured using the dice coefficient metric, obtaining promising quantitavive and qualitative results.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131705773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Baptiste Brument, L. Calvet, Robin Bruneau, J. Mélou, Simone Gasparini, Y. Quéau, F. Lauze, Jean-Denis Durou
{"title":"A shape-from-silhouette method for 3D reconstruction of a convex polyhedron","authors":"Baptiste Brument, L. Calvet, Robin Bruneau, J. Mélou, Simone Gasparini, Y. Quéau, F. Lauze, Jean-Denis Durou","doi":"10.1117/12.3000368","DOIUrl":"https://doi.org/10.1117/12.3000368","url":null,"abstract":"We present a pipeline to recover precisely the geometry of a convex polyhedral object from multiple views under circular motion. It is based on the extraction of visible polyhedron vertices from silhouette images and matching across a sequence of images. Compared to standard structure-from-motion pipelines, the method is well suited to the 3D-reconstruction of low-textured and non-Lambertian materials. Experiments on synthetic and real datasets show the efficacy of the proposed framework.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123734785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Laurent Fainsin, J. Mélou, L. Calvet, A. Carlier, Jean-Denis Durou
{"title":"Neural detection of spheres in images for lighting calibration","authors":"Laurent Fainsin, J. Mélou, L. Calvet, A. Carlier, Jean-Denis Durou","doi":"10.1117/12.3000202","DOIUrl":"https://doi.org/10.1117/12.3000202","url":null,"abstract":"Accurate detection of spheres in images holds significant value for photometric 3D vision techniques such as photometric stereo.1 These techniques require precise calibration of lighting, and sphere detection can help in the calibration process. Our proposed approach involves training neural networks to automatically detect spheres of three different material classes: matte, shiny and chrome. We get fast and accurate segmentation of spheres in images, outperforming manual segmentation in terms of speed while maintaining comparable accuracy.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130292509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Clément Joubert, B. Bringier, Julien Garnier, M. Khoudeir, N. Amalric
{"title":"Gloss assessment with deep photometric stereo: application to human skin","authors":"Clément Joubert, B. Bringier, Julien Garnier, M. Khoudeir, N. Amalric","doi":"10.1117/12.2692082","DOIUrl":"https://doi.org/10.1117/12.2692082","url":null,"abstract":"This contribution presents a novel method to extract skin physical parameters as geometry, colour and gloss with photometric stereo. Our method is based on QNN (Quaternion Neural Network) to estimate the surface geometry from images with a fixed viewpoint modifying surface illumination, i.e. photometric stereo. To that end, we assume that surface BRDF (Bidirectional Reflectance Distribution Function) can be separated by a diffuse and specular component. Once the geometry is estimated, colour is estimated from geometry to finally compute gloss. This method results on multiple gloss maps which are used to compute features that characterise surface gloss. Unlike other approaches, our method does not require polarising filters that suffer from a more complex light modelling. We demonstrate the effectiveness of our approach through experiments on rendering, cow leather and ex-vivo skin samples. The proposed method has potential for various real-world applications such as evaluating the appearance of skin care products or assessing skin health.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"12749 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128896412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kazuki Nakashima, Ryo Nakazawa, Hideharu Toda, H. Aomori, T. Otake, I. Matsuda, S. Itoh
{"title":"Hierarchical lossless image coding using CNN predictors considering prediction error distribution","authors":"Kazuki Nakashima, Ryo Nakazawa, Hideharu Toda, H. Aomori, T. Otake, I. Matsuda, S. Itoh","doi":"10.1117/12.2691664","DOIUrl":"https://doi.org/10.1117/12.2691664","url":null,"abstract":"We have researched a hierarchical lossless encoding method using cellular neural networks (CNN) as predictors. In our method, which belongs to the hierarchical lossless coding method, the prediction accuracy is improved by adaptively using different CNN predictors depending on the direction of the image edges. The prediction error obtained by CNN prediction is encoded by adaptive arithmetic coding using multiple probabilistic models based on the context modeling. In previous research,1 a new approach is introduced in which the prediction errors of each predictor are encoded separately by arithmetic coding. Although this method improves the performance of encoding prediction errors, increasing side information became an issue. Therefore, to reduce the side information of the arithmetic coders, we propose a grouping algorithm that groups the prediction errors corresponding to each predictor based on the utilization of the predictors.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"97 11","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120876951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}