Omnidirectional image quality assessment with gated dual-projection fusion

IF 3.4 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
ChengZhi Xiao, RuiKang Yu
{"title":"Omnidirectional image quality assessment with gated dual-projection fusion","authors":"ChengZhi Xiao,&nbsp;RuiKang Yu","doi":"10.1016/j.displa.2025.103173","DOIUrl":null,"url":null,"abstract":"<div><div>Existing omnidirectional image quality assessment (OIQA) models typically rely on the equirectangular projection (ERP) or cubemap projection (CMP) of omnidirectional images as inputs. However, the deformation in ERP and the discontinuities at the boundaries of CMP limit the network’s ability to represent image information, leading to information loss. Therefore, it is necessary to fuse these two projections of omnidirectional images to achieve comprehensive feature representation. Current OIQA models only integrate and interact high-level features extracted from different projection formats at the last stage of the network, overlooking potential information loss at each stage within the network. To this end, we consider the respective strengths and weaknesses of the two projections, and design a feature extraction and fusion module at each stage of the network to enhance the model’s representation capability. Specifically, the ERP features are first decomposed into two projection formats before being fed into each feature extraction stage of the network for separate processing. Subsequently, we introduce the gating mechanism and develop a Gated Dual-Projection Fusion module (GDPF) to interactively fuse the features computed from both the ERP and CMP projection formats. GDPF allows the developed model to enhance critical information while filtering out deformation and discontinuous information. The fused features are then input into the next stage, where the aforementioned operations are repeated. This process alleviates the issues of feature representation caused by deformation in ERP and discontinuities in CMP and the fused features are used for quality prediction. Experiments on three public datasets demonstrate the superior prediction accuracy of the proposed model.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"90 ","pages":"Article 103173"},"PeriodicalIF":3.4000,"publicationDate":"2025-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Displays","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0141938225002100","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0

Abstract

Existing omnidirectional image quality assessment (OIQA) models typically rely on the equirectangular projection (ERP) or cubemap projection (CMP) of omnidirectional images as inputs. However, the deformation in ERP and the discontinuities at the boundaries of CMP limit the network’s ability to represent image information, leading to information loss. Therefore, it is necessary to fuse these two projections of omnidirectional images to achieve comprehensive feature representation. Current OIQA models only integrate and interact high-level features extracted from different projection formats at the last stage of the network, overlooking potential information loss at each stage within the network. To this end, we consider the respective strengths and weaknesses of the two projections, and design a feature extraction and fusion module at each stage of the network to enhance the model’s representation capability. Specifically, the ERP features are first decomposed into two projection formats before being fed into each feature extraction stage of the network for separate processing. Subsequently, we introduce the gating mechanism and develop a Gated Dual-Projection Fusion module (GDPF) to interactively fuse the features computed from both the ERP and CMP projection formats. GDPF allows the developed model to enhance critical information while filtering out deformation and discontinuous information. The fused features are then input into the next stage, where the aforementioned operations are repeated. This process alleviates the issues of feature representation caused by deformation in ERP and discontinuities in CMP and the fused features are used for quality prediction. Experiments on three public datasets demonstrate the superior prediction accuracy of the proposed model.
基于门控双投影融合的全向图像质量评估
现有的全向图像质量评估(OIQA)模型通常依赖全向图像的等矩形投影(ERP)或立方映射投影(CMP)作为输入。然而,ERP中的变形和CMP边界处的不连续限制了网络对图像信息的表示能力,导致信息丢失。因此,有必要融合这两种全向图像的投影,以实现全面的特征表示。目前的OIQA模型只在网络的最后阶段整合和交互从不同投影格式中提取的高级特征,忽略了网络中每个阶段潜在的信息丢失。为此,我们考虑了两种投影各自的优缺点,并在网络的每个阶段设计了特征提取和融合模块,以增强模型的表示能力。具体来说,ERP特征首先被分解成两种投影格式,然后被送入网络的每个特征提取阶段进行单独处理。随后,我们引入了门控机制,并开发了一个门控双投影融合模块(GDPF),以交互融合从ERP和CMP投影格式计算的特征。GDPF允许开发的模型在过滤变形和不连续信息的同时增强关键信息。然后将融合的特征输入到下一阶段,在该阶段重复上述操作。该方法减轻了ERP中变形和CMP中不连续性引起的特征表示问题,并将融合的特征用于质量预测。在三个公共数据集上的实验表明,该模型具有较高的预测精度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Displays
Displays 工程技术-工程:电子与电气
CiteScore
4.60
自引率
25.60%
发文量
138
审稿时长
92 days
期刊介绍: Displays is the international journal covering the research and development of display technology, its effective presentation and perception of information, and applications and systems including display-human interface. Technical papers on practical developments in Displays technology provide an effective channel to promote greater understanding and cross-fertilization across the diverse disciplines of the Displays community. Original research papers solving ergonomics issues at the display-human interface advance effective presentation of information. Tutorial papers covering fundamentals intended for display technologies and human factor engineers new to the field will also occasionally featured.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信