{"title":"基于门控双投影融合的全向图像质量评估","authors":"ChengZhi Xiao, RuiKang Yu","doi":"10.1016/j.displa.2025.103173","DOIUrl":null,"url":null,"abstract":"<div><div>Existing omnidirectional image quality assessment (OIQA) models typically rely on the equirectangular projection (ERP) or cubemap projection (CMP) of omnidirectional images as inputs. However, the deformation in ERP and the discontinuities at the boundaries of CMP limit the network’s ability to represent image information, leading to information loss. Therefore, it is necessary to fuse these two projections of omnidirectional images to achieve comprehensive feature representation. Current OIQA models only integrate and interact high-level features extracted from different projection formats at the last stage of the network, overlooking potential information loss at each stage within the network. To this end, we consider the respective strengths and weaknesses of the two projections, and design a feature extraction and fusion module at each stage of the network to enhance the model’s representation capability. Specifically, the ERP features are first decomposed into two projection formats before being fed into each feature extraction stage of the network for separate processing. Subsequently, we introduce the gating mechanism and develop a Gated Dual-Projection Fusion module (GDPF) to interactively fuse the features computed from both the ERP and CMP projection formats. GDPF allows the developed model to enhance critical information while filtering out deformation and discontinuous information. The fused features are then input into the next stage, where the aforementioned operations are repeated. This process alleviates the issues of feature representation caused by deformation in ERP and discontinuities in CMP and the fused features are used for quality prediction. Experiments on three public datasets demonstrate the superior prediction accuracy of the proposed model.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"90 ","pages":"Article 103173"},"PeriodicalIF":3.4000,"publicationDate":"2025-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Omnidirectional image quality assessment with gated dual-projection fusion\",\"authors\":\"ChengZhi Xiao, RuiKang Yu\",\"doi\":\"10.1016/j.displa.2025.103173\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Existing omnidirectional image quality assessment (OIQA) models typically rely on the equirectangular projection (ERP) or cubemap projection (CMP) of omnidirectional images as inputs. However, the deformation in ERP and the discontinuities at the boundaries of CMP limit the network’s ability to represent image information, leading to information loss. Therefore, it is necessary to fuse these two projections of omnidirectional images to achieve comprehensive feature representation. Current OIQA models only integrate and interact high-level features extracted from different projection formats at the last stage of the network, overlooking potential information loss at each stage within the network. To this end, we consider the respective strengths and weaknesses of the two projections, and design a feature extraction and fusion module at each stage of the network to enhance the model’s representation capability. Specifically, the ERP features are first decomposed into two projection formats before being fed into each feature extraction stage of the network for separate processing. Subsequently, we introduce the gating mechanism and develop a Gated Dual-Projection Fusion module (GDPF) to interactively fuse the features computed from both the ERP and CMP projection formats. GDPF allows the developed model to enhance critical information while filtering out deformation and discontinuous information. The fused features are then input into the next stage, where the aforementioned operations are repeated. This process alleviates the issues of feature representation caused by deformation in ERP and discontinuities in CMP and the fused features are used for quality prediction. Experiments on three public datasets demonstrate the superior prediction accuracy of the proposed model.</div></div>\",\"PeriodicalId\":50570,\"journal\":{\"name\":\"Displays\",\"volume\":\"90 \",\"pages\":\"Article 103173\"},\"PeriodicalIF\":3.4000,\"publicationDate\":\"2025-07-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Displays\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0141938225002100\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Displays","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0141938225002100","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Omnidirectional image quality assessment with gated dual-projection fusion
Existing omnidirectional image quality assessment (OIQA) models typically rely on the equirectangular projection (ERP) or cubemap projection (CMP) of omnidirectional images as inputs. However, the deformation in ERP and the discontinuities at the boundaries of CMP limit the network’s ability to represent image information, leading to information loss. Therefore, it is necessary to fuse these two projections of omnidirectional images to achieve comprehensive feature representation. Current OIQA models only integrate and interact high-level features extracted from different projection formats at the last stage of the network, overlooking potential information loss at each stage within the network. To this end, we consider the respective strengths and weaknesses of the two projections, and design a feature extraction and fusion module at each stage of the network to enhance the model’s representation capability. Specifically, the ERP features are first decomposed into two projection formats before being fed into each feature extraction stage of the network for separate processing. Subsequently, we introduce the gating mechanism and develop a Gated Dual-Projection Fusion module (GDPF) to interactively fuse the features computed from both the ERP and CMP projection formats. GDPF allows the developed model to enhance critical information while filtering out deformation and discontinuous information. The fused features are then input into the next stage, where the aforementioned operations are repeated. This process alleviates the issues of feature representation caused by deformation in ERP and discontinuities in CMP and the fused features are used for quality prediction. Experiments on three public datasets demonstrate the superior prediction accuracy of the proposed model.
期刊介绍:
Displays is the international journal covering the research and development of display technology, its effective presentation and perception of information, and applications and systems including display-human interface.
Technical papers on practical developments in Displays technology provide an effective channel to promote greater understanding and cross-fertilization across the diverse disciplines of the Displays community. Original research papers solving ergonomics issues at the display-human interface advance effective presentation of information. Tutorial papers covering fundamentals intended for display technologies and human factor engineers new to the field will also occasionally featured.