Li Xu , Yingjie Zhou , Sitong Liu , Farong Wen , Yu Zhou , Xiaohong Liu , Jie Guo , Yu Wang , Jiezhang Cao
{"title":"数字人类面部质量评估:融合形态和谐与表达协调的双分支框架","authors":"Li Xu , Yingjie Zhou , Sitong Liu , Farong Wen , Yu Zhou , Xiaohong Liu , Jie Guo , Yu Wang , Jiezhang Cao","doi":"10.1016/j.displa.2025.103221","DOIUrl":null,"url":null,"abstract":"<div><div>With the rapid advancement of metaverse technologies, digital humans (DH), as core interactive entities in virtual-physical integrated ecosystems, face unique challenges in their quality assessment frameworks. Existing research predominantly focuses on quantifying natural image distortions but fails to address DH-specific issues such as facial morphological disharmony and expression incoherence. To bridge this gap, we propose a dual-branch quality assessment framework for digital humans: (1) Leveraging medical aesthetic priors, we construct structural features based on facial aesthetic subunits and model temporal dependencies using gated recurrent units, combined with a loss-averse pooling strategy to capture transient severe distortions. (2) We quantify expression coordination through multi-dimensional Action Unit (AU) topology graphs, proposing triple-edge definitions and regressing dynamic distortion levels via graph convolutional networks. Experiments on the multiple THQA datasets demonstrate that our framework significantly outperforms conventional methods in subjective mean opinion score consistency, with the dynamic branch playing a dominant role in performance optimization. This work establishes a quantifiable evaluation standard for DH modeling refinement and real-time rendering.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103221"},"PeriodicalIF":3.4000,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Facial quality assessment of digital humans: A dual-branch framework integrating morphological harmony and expressive coordination\",\"authors\":\"Li Xu , Yingjie Zhou , Sitong Liu , Farong Wen , Yu Zhou , Xiaohong Liu , Jie Guo , Yu Wang , Jiezhang Cao\",\"doi\":\"10.1016/j.displa.2025.103221\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>With the rapid advancement of metaverse technologies, digital humans (DH), as core interactive entities in virtual-physical integrated ecosystems, face unique challenges in their quality assessment frameworks. Existing research predominantly focuses on quantifying natural image distortions but fails to address DH-specific issues such as facial morphological disharmony and expression incoherence. To bridge this gap, we propose a dual-branch quality assessment framework for digital humans: (1) Leveraging medical aesthetic priors, we construct structural features based on facial aesthetic subunits and model temporal dependencies using gated recurrent units, combined with a loss-averse pooling strategy to capture transient severe distortions. (2) We quantify expression coordination through multi-dimensional Action Unit (AU) topology graphs, proposing triple-edge definitions and regressing dynamic distortion levels via graph convolutional networks. Experiments on the multiple THQA datasets demonstrate that our framework significantly outperforms conventional methods in subjective mean opinion score consistency, with the dynamic branch playing a dominant role in performance optimization. This work establishes a quantifiable evaluation standard for DH modeling refinement and real-time rendering.</div></div>\",\"PeriodicalId\":50570,\"journal\":{\"name\":\"Displays\",\"volume\":\"91 \",\"pages\":\"Article 103221\"},\"PeriodicalIF\":3.4000,\"publicationDate\":\"2025-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Displays\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0141938225002586\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Displays","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0141938225002586","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Facial quality assessment of digital humans: A dual-branch framework integrating morphological harmony and expressive coordination
With the rapid advancement of metaverse technologies, digital humans (DH), as core interactive entities in virtual-physical integrated ecosystems, face unique challenges in their quality assessment frameworks. Existing research predominantly focuses on quantifying natural image distortions but fails to address DH-specific issues such as facial morphological disharmony and expression incoherence. To bridge this gap, we propose a dual-branch quality assessment framework for digital humans: (1) Leveraging medical aesthetic priors, we construct structural features based on facial aesthetic subunits and model temporal dependencies using gated recurrent units, combined with a loss-averse pooling strategy to capture transient severe distortions. (2) We quantify expression coordination through multi-dimensional Action Unit (AU) topology graphs, proposing triple-edge definitions and regressing dynamic distortion levels via graph convolutional networks. Experiments on the multiple THQA datasets demonstrate that our framework significantly outperforms conventional methods in subjective mean opinion score consistency, with the dynamic branch playing a dominant role in performance optimization. This work establishes a quantifiable evaluation standard for DH modeling refinement and real-time rendering.
期刊介绍:
Displays is the international journal covering the research and development of display technology, its effective presentation and perception of information, and applications and systems including display-human interface.
Technical papers on practical developments in Displays technology provide an effective channel to promote greater understanding and cross-fertilization across the diverse disciplines of the Displays community. Original research papers solving ergonomics issues at the display-human interface advance effective presentation of information. Tutorial papers covering fundamentals intended for display technologies and human factor engineers new to the field will also occasionally featured.