Nermeen Nader, Fatma El-Zahraa Ahmed El-Gamal, Mohammed Elmogy
{"title":"An efficient deep learning system for kinship verification based on ConvNext-EfficientNet-VIT feature fusion","authors":"Nermeen Nader, Fatma El-Zahraa Ahmed El-Gamal, Mohammed Elmogy","doi":"10.1016/j.eij.2025.100809","DOIUrl":null,"url":null,"abstract":"<div><div>Kinship verification has emerged as a compelling area of research within computer vision, driven by its critical role in real-world applications, such as forensic investigations and the search for missing persons. Despite recent progress, the task remains challenging due to subtle facial similarities across generations and variations in pose, lighting, and expression. Deep learning techniques have significantly advanced the field. Among them, feature fusion has proven to be a powerful tool for enhancing model performance by integrating complementary characteristics from multiple architectures. This research introduces a new kinship verification framework that harnesses the strengths of ConvNext-Base, EfficientNet-B0, and vision transformer (ViT) through an effective feature fusion strategy. By combining the local texture sensitivity of ConvNeXt, the parameter efficiency of EfficientNet-B0, and the global context modeling capabilities of ViT. The fused representation captures a more holistic and discriminative understanding of facial features relevant to kinship. To the best of the authors’ knowledge, this specific fusion of deep models has not yet been explored for kinship verification. The proposed framework is structured into six stages: image preprocessing, parent/child image pairing, feature extraction, feature normalization, feature fusion, and classification. It was evaluated using two standard benchmark datasets – KinFaceW-I (KinFWI) and KinFaceW-II (KinFWII) – achieving maximum accuracy rates of 84.85% and 91.65%, respectively. These results outperform several state-of-the-art (SOTA) methods and underscore the critical role of multi-model feature fusion in improving the accuracy and robustness of kinship verification systems. This research’s promising findings validate the proposed approach’s effectiveness and highlight the potential of deep feature fusion in addressing complex facial analysis problems.</div></div>","PeriodicalId":56010,"journal":{"name":"Egyptian Informatics Journal","volume":"32 ","pages":"Article 100809"},"PeriodicalIF":4.3000,"publicationDate":"2025-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Egyptian Informatics Journal","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1110866525002026","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Kinship verification has emerged as a compelling area of research within computer vision, driven by its critical role in real-world applications, such as forensic investigations and the search for missing persons. Despite recent progress, the task remains challenging due to subtle facial similarities across generations and variations in pose, lighting, and expression. Deep learning techniques have significantly advanced the field. Among them, feature fusion has proven to be a powerful tool for enhancing model performance by integrating complementary characteristics from multiple architectures. This research introduces a new kinship verification framework that harnesses the strengths of ConvNext-Base, EfficientNet-B0, and vision transformer (ViT) through an effective feature fusion strategy. By combining the local texture sensitivity of ConvNeXt, the parameter efficiency of EfficientNet-B0, and the global context modeling capabilities of ViT. The fused representation captures a more holistic and discriminative understanding of facial features relevant to kinship. To the best of the authors’ knowledge, this specific fusion of deep models has not yet been explored for kinship verification. The proposed framework is structured into six stages: image preprocessing, parent/child image pairing, feature extraction, feature normalization, feature fusion, and classification. It was evaluated using two standard benchmark datasets – KinFaceW-I (KinFWI) and KinFaceW-II (KinFWII) – achieving maximum accuracy rates of 84.85% and 91.65%, respectively. These results outperform several state-of-the-art (SOTA) methods and underscore the critical role of multi-model feature fusion in improving the accuracy and robustness of kinship verification systems. This research’s promising findings validate the proposed approach’s effectiveness and highlight the potential of deep feature fusion in addressing complex facial analysis problems.
期刊介绍:
The Egyptian Informatics Journal is published by the Faculty of Computers and Artificial Intelligence, Cairo University. This Journal provides a forum for the state-of-the-art research and development in the fields of computing, including computer sciences, information technologies, information systems, operations research and decision support. Innovative and not-previously-published work in subjects covered by the Journal is encouraged to be submitted, whether from academic, research or commercial sources.