Lázaro J. González Soler, C. Rathgeb, Daniel Fischer
{"title":"Semi-synthetic Data Generation for Tattoo Segmentation","authors":"Lázaro J. González Soler, C. Rathgeb, Daniel Fischer","doi":"10.1109/IWBF57495.2023.10157837","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157837","url":null,"abstract":"Tattoos have been successfully employed to assist law enforcement in the identification of criminals and victims. Due to various privacy issues in acquiring images containing tattoos, only a limited number of databases exist. This lack of databases has slowed down the development of new tattoo segmentation and retrieval methods. In our work, we propose a new unsupervised generator that allows generating a large number of semi-synthetic images with tattooed subjects. To successfully generate realistic images, a database including the respective skin segmentation map is also proposed. Using this new generator and the skin database, 5,500 semi-synthetic images were created and evaluated for the tattoo segmentation use case. Experimental results on real data show the usefulness of using semi-synthetic images to train semantic segmentation algorithms: several manually mislabelled real samples were successfully corrected. The tattoo generator code, the skin database and generated images have been made available at https://dasec.h-da.de/hda-sstd/.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"494 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133952021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pamela C. Zurita, Daniel P. Benalcazar, Juan E. Tapia
{"title":"Fitness-for-Duty Classification using Temporal Sequences of Iris Periocular images","authors":"Pamela C. Zurita, Daniel P. Benalcazar, Juan E. Tapia","doi":"10.1109/IWBF57495.2023.10157018","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157018","url":null,"abstract":"Fitness for Duty (FFD) techniques detects whether a subject is Fit to perform their work safely, which means no reduced alertness condition and security, or if they are Unfit, which means alertness condition reduced by sleepiness or consumption of alcohol and drugs. Human iris behaviour provides valuable information to predict FFD since pupil and iris movements are controlled by the central nervous system and are influenced by illumination, fatigue, alcohol, and drugs. This work aims to classify FFD using sequences of 8 iris images and to extract spatial and temporal information using Convolutional Neural Networks (CNN) and Long Short Term Memory Networks (LSTM). Our results achieved a precision of 81.4% and 96.9% for the prediction of Fit and Unfit subjects, respectively. The results also show that it is possible to determine if a subject is under alcohol, drug, and sleepiness conditions. Sleepiness can be identified as the most difficult condition to be determined. This system opens a different insight into iris biometric applications.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"212 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122840719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On the Use of Cross-module Attention Statistics Pooling for Speaker Verification","authors":"J. Alam, A. Fathan","doi":"10.1109/IWBF57495.2023.10157564","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157564","url":null,"abstract":"In deep learning-based speaker verification frameworks, extraction of a speaker embedding vector plays a key role. In this contribution, we propose a hybrid neural network that employs a cross-module attention pooling mechanism for the extraction of speaker discriminant utterance-level embeddings. In particular, the proposed system incorporates a 2D-Convolution Neural Network (CNN)-based feature extraction module in cascade with a frame-level network, which is composed of a fully Time Delay Neural Network (TDNN) network and a TDNN-Long Short Term Memory (TDNN-LSTM) hybrid network in a parallel manner. The proposed system also employs cross-module attention statistics pooling for aggregating the speaker information within an utterance-level context by capturing the complementarity between two parallelly connected modules. We conduct a set of experiments on the Voxceleb corpus for evaluating the performance of the proposed system and the proposed hybrid network is able to provide better results than the conventional approaches trained on the same dataset.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131655716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Muhammad Ahmad Amin, Yongjian Hu, Huimin She, Jicheng Li, Yu Guan, Muhammad Zain Amin
{"title":"Exposing Deepfake Frames through Spectral Analysis of Color Channels in Frequency Domain","authors":"Muhammad Ahmad Amin, Yongjian Hu, Huimin She, Jicheng Li, Yu Guan, Muhammad Zain Amin","doi":"10.1109/IWBF57495.2023.10157211","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157211","url":null,"abstract":"Highly realistic deepfakes are generated by employing generative neural networks, even to the point that it is difficult for humans to tell them apart from the real ones. Nowadays they are one of the causes of misrepresentation or misinformation regarding different subjects. The detection of deepfake content is very important. It can be analyzed in different domains, such as spatial domain and frequency domain, or by employing combinations of them. In this work, we first took inspiration from traditional image forensics and performed a comprehensive frequency spectrum analysis on the deepfake frames and their context color channels to detect spectral anomalies and statistical features. We then use the frequency spectrum statistical features to distinguish between pristine and deepfake content using both unsupervised and supervised learning approaches. Finally, we scrutinize the trained deepfake detection models’ generalization capability from the perspective of suggested statistical features across different deepfake datasets and methods. Our analysis demonstrated the effectiveness of statistical features by identifying real and deepfake content with high accuracy, surpassing the performance of several state-of-the-art methods.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116625591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Johannes Schuiki, Christof Kauba, H. Hofbauer, A. Uhl
{"title":"Cross-Sensor Micro-Texture Material Classification and Smartphone Acquisition do not go well together","authors":"Johannes Schuiki, Christof Kauba, H. Hofbauer, A. Uhl","doi":"10.1109/IWBF57495.2023.10157739","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157739","url":null,"abstract":"Intrinsic, non-invasive product authentication is still an important topic as it does not generate additional costs during the production process. This topic is of specific interest for medical products as non-genuine products can directly effect the patients’ health. This work investigates micro-texture classification as a mean of proving the authenticity of zircon oxide blocks (for dental implants). Samples of three different manufacturers were acquired using four smartphone devices with a clip-on macro lens. In addition, an existing drug packaging material database was utilized. While the intra-sensor microtexture classification worked well, the cross-sensor classification results were less promising. In an attempt to track down the limiting factors, intrinsic sensor features usually used in device identification were investigated as well.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121259063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Diego Pasmino, C. Aravena, Juan E. Tapia, C. Busch
{"title":"Flickr-PAD: New Face High-Resolution Presentation Attack Detection Database","authors":"Diego Pasmino, C. Aravena, Juan E. Tapia, C. Busch","doi":"10.1109/IWBF57495.2023.10157771","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157771","url":null,"abstract":"Nowadays, Presentation Attack Detection is a very active research area. Several databases are constituted in the state-of-the-art using images extracted from videos. One of the main problems identified is that many databases present a low-quality, small image size and do not represent an operational scenario in a real remote biometric system. Currently, these images are captured from smartphones with high-quality and bigger resolutions. In order to increase the diversity of image quality, this work presents a new PAD database based on open-access Flickr images called: “Flickr-PAD”. Our new hand-made database shows high-quality printed and screen scenarios. This will help researchers to compare new approaches to existing algorithms on a wider database. This database will be available for other researchers. A leave-one-out protocol was used to train and evaluate three PAD models based on MobileNet-V3 (small and large) and EfficientNet-B0. The best result was reached with MobileNet-V3 large with BPCER10 of 7.08% and BPCER20 of 11.15%.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"22 9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131594020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lorenzo Papa, Lorenzo Faiella, Luca Corvitto, Luca Maiano, Irene Amerini
{"title":"On the use of Stable Diffusion for creating realistic faces: from generation to detection","authors":"Lorenzo Papa, Lorenzo Faiella, Luca Corvitto, Luca Maiano, Irene Amerini","doi":"10.1109/IWBF57495.2023.10156981","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10156981","url":null,"abstract":"The mass adoption of diffusion models has shown that artificial intelligence (AI) systems can be used to easily generate realistic images. The spread of these technologies paves the way to previously unimaginable creative uses while also raising the possibility of malicious applications. In this work, we propose a critical analysis of the overall pipeline, i.e., from creating realistic human faces with Stable Diffusion v1.5 [1] to recognizing fake ones. We first propose an analysis of the prompts that allow the generation of extremely realistic faces with a human-in-the-loop approach. Our objective is to identify the text prompts that drive the image generation process to obtain realistic photos that resemble everyday portraits captured with any camera. Next, we study how complex it is to recognize these fake contents for both AI-based models and non-expert humans. We conclude that similar to other deepzfake creation techniques, despite some limitations in generalization across different datasets, it is possible to use AI to recognize these contents more accurately than non-expert humans would.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126142517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MMFV: A Multi-Movement Finger-Video Database for Contactless Fingerprint Recognition","authors":"Aakarsh Malhotra, Mayank Vatsa, Richa Singh","doi":"10.1109/IWBF57495.2023.10156919","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10156919","url":null,"abstract":"Biometric authentication during the COVID-19 and post-pandemic times require a touchless authentication mechanism. While existing studies showcase the use of fingerphoto for touchless authentication, a short video of the finger can provide many good-quality frames. This research presents the first publicly available finger-video dataset, titled Multi-Movement Finger-Video (MMFV) Database. The MMFV dataset has 3792 videos from 336 classes, acquired over two sessions, and spans three different movement types (pitch, yaw, and roll). To establish the baseline performance for the proposed MMFV database, we perform recognition using seven popular fingerprint and deep learning-based algorithms for fingerphoto recognition. The recognition is performed using a fixed, randomly selected frame from all the algorithms. Experimental results showcase that Siamese network-based verification provides the most optimal results across different movements, with observed EER as low as 2.70%.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124595711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face Morphing Attack Detection with Denoising Diffusion Probabilistic Models","authors":"Marija Ivanovska, Vitomir Štruc","doi":"10.1109/IWBF57495.2023.10156877","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10156877","url":null,"abstract":"Morphed face images have recently become a growing concern for existing face verification systems, as they are relatively easy to generate and can be used to impersonate someone’s identity for various malicious purposes. Efficient Morphing Attack Detection (MAD) that generalizes well across different morphing techniques is, therefore, of paramount importance. Existing MAD techniques predominantly rely on discriminative models that learn from examples of bona fide and morphed images and, as a result, often exhibit sub-optimal generalization performance when confronted with unknown types of morphing attacks. To address this problem, we propose a novel, diffusion–based MAD method in this paper that learns only from the characteristics of bona fide images. Various forms of morphing attacks are then detected by our model as out-of-distribution samples. We perform rigorous experiments over four different datasets (CASIA-WebFace, FRLL-Morphs, FERET-Morphs and FRGC-Morphs) and compare the proposed solution to both discriminatively-trained and once-class MAD models. The experimental results show that our MAD model achieves highly competitive results on all considered datasets.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116733137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Roberto Leyva, G. Epiphaniou, C. Maple, Victor Sanchez
{"title":"Unsupervised Face Synthesis Based on Human Traits","authors":"Roberto Leyva, G. Epiphaniou, C. Maple, Victor Sanchez","doi":"10.1109/IWBF57495.2023.10157232","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157232","url":null,"abstract":"This paper presents a strategy to synthesize face images based on human traits. Specifically, the strategy allows synthesizing face images with similar age, gender, and ethnicity, after discovering groups of people with similar facial features. Our synthesizer is based on unsupervised learning and is capable to generate realistic faces. Our experiments reveal that grouping the training samples according to their similarity can lead to more realistic face images while having semantic control over the synthesis. The proposed strategy achieves competitive performance compared to the state-of-the-art and outperforms the baseline in terms of the Frechet Inception Distance.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128589610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}