IET BiometricsPub Date : 2022-06-13DOI: 10.1049/bme2.12081
Jinxiao Zhong, Liangnian Jin, Ran Wang
{"title":"Point-convolution-based human skeletal pose estimation on millimetre wave frequency modulated continuous wave multiple-input multiple-output radar","authors":"Jinxiao Zhong, Liangnian Jin, Ran Wang","doi":"10.1049/bme2.12081","DOIUrl":"10.1049/bme2.12081","url":null,"abstract":"<p>Compared with traditional approaches that used vision sensors which can provide a high-resolution representation of targets, millimetre-wave radar is robust to scene lighting and weather conditions, and has more applications. Current methods of human skeletal pose estimation can reconstruct targets, but they lose the spatial information or don't take the density of point cloud into consideration. We propose a skeletal pose estimation method that combines point convolution to extract features from the point cloud. By extracting the local information and density of each point in the point cloud of the target, the spatial location and structure information of the target can be obtained, and the accuracy of the pose estimation is increased. The extraction of point cloud features is based on point-by-point convolution, that is, different weights are applied to different features of each point, which also increases the nonlinear expression ability of the model. Experiments show that the proposed approach is effective. We offer more distinct skeletal joints and a lower mean absolute error, average localisation errors of 6.1 cm in <i>X</i>, 3.5 cm in <i>Y</i> and 3.3 cm in <i>Z</i>, respectively.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 4","pages":"333-342"},"PeriodicalIF":2.0,"publicationDate":"2022-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12081","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91101921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
IET BiometricsPub Date : 2022-06-07DOI: 10.1049/bme2.12084
Jose Maureira, Juan E. Tapia, Claudia Arellano, Christoph Busch
{"title":"Analysis of the synthetic periocular iris images for robust Presentation Attacks Detection algorithms","authors":"Jose Maureira, Juan E. Tapia, Claudia Arellano, Christoph Busch","doi":"10.1049/bme2.12084","DOIUrl":"10.1049/bme2.12084","url":null,"abstract":"<p>The LivDet-2020 competition focuses on Presentation Attacks Detection (PAD) algorithms, has still open problems, mainly unknown attack scenarios. It is crucial to enhance PAD methods. This can be achieved by augmenting the number of Presentation Attack Instruments (PAI) and Bona fide (genuine) images used to train such algorithms. Unfortunately, the capture and creation of PAI and even the capture of Bona fide images are sometimes complex to achieve. The generation of synthetic images with Generative Adversarial Networks (GAN) algorithms may help and has shown significant improvements in recent years. This paper presents a benchmark of GAN methods to achieve a novel synthetic PAI from a small set of periocular near-infrared images. The best PAI was obtained using StyleGAN2, and it was tested using the best PAD algorithm from the LivDet-2020. The synthetic PAI was able to fool such an algorithm. As a result, all images were classified as Bona fide. A MobileNetV2 was trained using the synthetic PAI as a new class to achieve a more robust PAD. The resulting PAD was able to classify 96.7% of synthetic images as attacks. BPCER<sub>10</sub> was 0.24%. Such results demonstrated the need for PAD algorithms to be constantly updated and trained with synthetic images.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 4","pages":"343-354"},"PeriodicalIF":2.0,"publicationDate":"2022-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12084","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82133166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
IET BiometricsPub Date : 2022-06-03DOI: 10.1049/bme2.12083
Andre Brasil Vieira Wyzykowski, Mauricio Pamplona Segundo, Rubisley de Paula Lemes
{"title":"Multiresolution synthetic fingerprint generation","authors":"Andre Brasil Vieira Wyzykowski, Mauricio Pamplona Segundo, Rubisley de Paula Lemes","doi":"10.1049/bme2.12083","DOIUrl":"10.1049/bme2.12083","url":null,"abstract":"<p>Public access to existing high-resolution databases was discontinued. Besides, a hybrid database that contains fingerprints of different sensors with high and medium resolutions does not exist. A novel hybrid approach to synthesise realistic, multiresolution, and multisensor fingerprints to address these issues is presented. The first step was to improve Anguli, a handcrafted fingerprint generator, to create pores, scratches, and dynamic ridge maps. Using CycleGAN, then the maps are converted into realistic fingerprints, adding textures to images. Unlike other neural network-based methods, the authors’ method generates multiple images with different resolutions and styles for the same identity. With the authors’ approach, a synthetic database with 14,800 fingerprints is built. Besides that, fingerprint recognition experiments with pore- and minutiae-based matching techniques and different fingerprint quality analyses are conducted to confirm the similarity between real and synthetic databases. Finally, a human classification analysis is performed, where volunteers could not distinguish between authentic and synthetic fingerprints. These experiments demonstrate that the authors’ approach is suitable for supporting further fingerprint recognition studies in the absence of real databases.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 4","pages":"314-332"},"PeriodicalIF":2.0,"publicationDate":"2022-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12083","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77469700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
IET BiometricsPub Date : 2022-05-27DOI: 10.1049/bme2.12080
Chaoying Tang, Mengen Qian, Ru Jia, Haodong Liu, Biao Wang
{"title":"Forearm multimodal recognition based on IAHP-entropy weight combination","authors":"Chaoying Tang, Mengen Qian, Ru Jia, Haodong Liu, Biao Wang","doi":"10.1049/bme2.12080","DOIUrl":"https://doi.org/10.1049/bme2.12080","url":null,"abstract":"<p>Biometrics are the among most popular authentication methods due to their advantages over traditional methods, such as higher security, better accuracy and more convenience. The recent COVID-19 pandemic has led to the wide use of face masks, which greatly affects the traditional face recognition technology. The pandemic has also increased the focus on hygienic and contactless identity verification methods. The forearm is a new biometric that contains discriminative information. In this paper, we proposed a multimodal recognition method that combines the veins and geometry of a forearm. Five features are extracted from a forearm Near-Infrared (Near-Infrared) image: SURF, local line structures, global graph representations, forearm width feature and forearm boundary feature. These features are matched individually and then fused at the score level based on the Improved Analytic Hierarchy Process-entropy weight combination. Comprehensive experiments were carried out to evaluate the proposed recognition method and the fusion rule. The matching results showed that the proposed method can achieve a satisfactory performance.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"12 1","pages":"52-63"},"PeriodicalIF":2.0,"publicationDate":"2022-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12080","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50146449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Towards pen-holding hand pose recognition: A new benchmark and a coarse-to-fine PHHP recognition network","authors":"Pingping Wu, Lunke Fei, Shuyi Li, Shuping Zhao, Xiaozhao Fang, Shaohua Teng","doi":"10.1049/bme2.12079","DOIUrl":"10.1049/bme2.12079","url":null,"abstract":"<p>Hand pose recognition has been one of the most fundamental tasks in computer vision and pattern recognition, and substantial effort has been devoted to this field. However, owing to lack of public large-scale benchmark dataset, there is little literature to specially study pen-holding hand pose (PHHP) recognition. As an attempt to fill this gap, in this paper, a PHHP image dataset, consisting of 18,000 PHHP samples is established. To the best of the authors’ knowledge, this is the largest vision-based PHHP dataset ever collected. Furthermore, the authors design a coarse-to-fine PHHP recognition network consisting of a coarse multi-feature learning network and a fine pen-grasping-specific feature learning network, where the coarse learning network aims to extensively exploit the multiple discriminative features by sharing a hand-shape-based spatial attention information, and the fine learning network further learns the pen-grasping-specific features by embedding a couple of convolutional block attention modules into three convolution blocks models. Experimental results show that the authors’ proposed method can achieve a very competitive PHHP recognition performance when compared with the baseline recognition models.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 6","pages":"581-587"},"PeriodicalIF":2.0,"publicationDate":"2022-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12079","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75725455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recognition of human Iris for biometric identification using Daugman’s method","authors":"Reend Tawfik Mohammed, Harleen Kaur, Bhavya Alankar, Ritu Chauhan","doi":"10.1049/bme2.12074","DOIUrl":"10.1049/bme2.12074","url":null,"abstract":"<p>Iris identification is a well-known technology used to detect striking biometric identification procedures for recognizing human beings based on physical behaviour. The texture of the iris is unique and its anatomy varies from individual to individual. As we know, the physical features of human beings are unique, and they never change; this has led to a significant development in the field of iris recognition. Iris recognition tends to be a reliable domain of technology as it inherits the random variation of the data. In the proposed study of approach, we have designed and implemented a framework using various subsystems, where each phase relates to the other iris recognition system, and these stages are discussed as segmentation, normalisation, and feature encoding. The study is implemented using MATLAB where the results are outcast using the rapid application development (RAD) approach. We have applied the RAD domain, as it has an excellent computing power to generate expeditious results using complex coding, image processing toolbox, and high-level programing methodology. Further, the performance of the technology is tested on two informational groups of eye images MMU Iris database, CASIA V1, CASIA V2, MICHE I, MICHE II iris database, and images captured by iPhone camera and Android phone. The emphasis on the current study of approach is to apply the proposed algorithm to achieve high performance with less ideal conditions.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 4","pages":"304-313"},"PeriodicalIF":2.0,"publicationDate":"2022-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12074","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89044719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Breast mass classification based on supervised contrastive learning and multi-view consistency penalty on mammography","authors":"Lilei Sun, Jie Wen, Junqian Wang, Zheng Zhang, Yong Zhao, Guiying Zhang, Yong Xu","doi":"10.1049/bme2.12076","DOIUrl":"10.1049/bme2.12076","url":null,"abstract":"<p>Breast cancer accounts for the largest number of patients among all cancers in the world. Intervention treatment for early breast cancer can dramatically extend a woman's 5-year survival rate. However, the lack of public available breast mammography databases in the field of Computer-aided Diagnosis and the insufficient feature extraction ability from breast mammography limit the diagnostic performance of breast cancer. In this paper, A novel classification algorithm based on Convolutional Neural Network (CNN) is proposed to improve the diagnostic performance for breast cancer on mammography. A multi-view network is designed to extract the complementary information between the Craniocaudal (CC) and Mediolateral Oblique (MLO) mammographic views of a breast mass. For the different predictions of the features extracted from the CC view and MLO view of the same breast mass, the proposed algorithm forces the network to extract the consistent features from the two views by the cross-entropy function with an added consistent penalty term. To exploit the discriminative features from the insufficient mammographic images, the authors learnt an encoder in the classification model to learn the invariable representations from the mammographic breast mass by Supervised Contrastive Learning (SCL) to weaken the side effect of colour jitter and illumination of mammographic breast mass on image quality degradation. The experimental results of all the classification algorithms mentioned in this paper on Digital Database for Screening Mammography (DDSM) illustrate that the proposed algorithm greatly improves the classification performance and diagnostic speed of mammographic breast mass, which is of great significance for breast cancer diagnosis.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 6","pages":"588-600"},"PeriodicalIF":2.0,"publicationDate":"2022-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12076","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89203516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Masked face recognition: Human versus machine","authors":"Naser Damer, Fadi Boutros, Marius Süßmilch, Meiling Fang, Florian Kirchbuchner, Arjan Kuijper","doi":"10.1049/bme2.12077","DOIUrl":"10.1049/bme2.12077","url":null,"abstract":"<p>The recent COVID-19 pandemic has increased the focus on hygienic and contactless identity verification methods. However, the pandemic led to the wide use of face masks, essential to keep the pandemic under control. The effect of wearing a mask on face recognition (FR) in a collaborative environment is a currently sensitive yet understudied issue. Recent reports have tackled this by evaluating the masked probe effect on the performance of automatic FR solutions. However, such solutions can fail in certain processes, leading to the verification task being performed by a human expert. This work provides a joint evaluation and in-depth analyses of the face verification performance of human experts in comparison to state-of-the-art automatic FR solutions. This involves an extensive evaluation by human experts and 4 automatic recognition solutions. The study concludes with a set of take-home messages on different aspects of the correlation between the verification behaviour of humans and machines.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 5","pages":"512-528"},"PeriodicalIF":2.0,"publicationDate":"2022-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12077","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90566262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
IET BiometricsPub Date : 2022-05-05DOI: 10.1049/bme2.12073
Wardah Farrukh, Dustin van der Haar
{"title":"Lip print-based identification using traditional and deep learning","authors":"Wardah Farrukh, Dustin van der Haar","doi":"10.1049/bme2.12073","DOIUrl":"https://doi.org/10.1049/bme2.12073","url":null,"abstract":"<p>The concept of biometric identification is centred around the theory that every individual is unique and has distinct characteristics. Various metrics such as fingerprint, face, iris, or retina are adopted for this purpose. Nonetheless, new alternatives are needed to establish the identity of individuals on occasions where the above techniques are unavailable. One emerging method of human recognition is lip-based identification. It can be treated as a new kind of biometric measure. The patterns found on the human lip are permanent unless subjected to alternations or trauma. Therefore, lip prints can serve the purpose of confirming an individual's identity. The main objective of this work is to design experiments using computer vision methods that can recognise an individual solely based on their lip prints. This article compares traditional and deep learning computer vision methods and how they perform on a common dataset for lip-based identification. The first pipeline is a traditional method with Speeded Up Robust Features with either an SVM or K-NN machine learning classifier, which achieved an accuracy of 95.45% and 94.31%, respectively. A second pipeline compares the performance of the VGG16 and VGG19 deep learning architectures. This approach obtained an accuracy of 91.53% and 93.22%, respectively.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"12 1","pages":"1-12"},"PeriodicalIF":2.0,"publicationDate":"2022-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12073","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50121827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Time–frequency fusion learning for photoplethysmography biometric recognition","authors":"Chunying Liu, Jijiang Yu, Yuwen Huang, Fuxian Huang","doi":"10.1049/bme2.12070","DOIUrl":"https://doi.org/10.1049/bme2.12070","url":null,"abstract":"<p>Photoplethysmography (PPG) signal is a novel biometric trait related to the identity of people; many time- and frequency-domain methods for PPG biometric recognition have been proposed. However, the existing domain methods for PPG biometric recognition only consider a single domain or the feature-level fusion of time and frequency domains, without considering the exploration of the fusion correlations of the time and frequency domains. The authors propose a time–frequency fusion for a PPG biometric recognition method with collective matrix factorisation (TFCMF) that leverages collective matrix factorisation to learn a shared latent semantic space by exploring the fusion correlations of the time and frequency domains. In addition, the authors utilise the <i>ℓ</i><sub>2,1</sub> norm to constrain the reconstruction error and shared matrix, which can alleviate the influence of noise and intra-class variation, and ensure the robustness of learnt semantic space. Experiments demonstrate that TFCMF has better recognition performance than current state-of-the-art methods for PPG biometric recognition.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"11 3","pages":"187-198"},"PeriodicalIF":2.0,"publicationDate":"2022-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2.12070","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91827864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}