{"title":"Opinion-unaware blind quality assessment of AI-generated omnidirectional images based on deep feature statistics","authors":"Xuelin Liu, Jiebin Yan, Yuming Fang, Jingwen Hou","doi":"10.1016/j.jvcir.2025.104461","DOIUrl":null,"url":null,"abstract":"<div><div>The advancement of artificial intelligence generated content (AIGC) and virtual reality (VR) technologies have prompted AI-generated omnidirectional images (AGOI) to gradually into people’s daily lives. Compared to natural omnidirectional images, AGOIs exhibit traditional low-level technical distortions and high-level semantic distortions, which can severely affect the immersive experience for users in practical applications. Consequently, there is an urgent need for thorough research and precise evaluation of AGOI quality. In this paper, we propose a novel opinion-unaware (OU) blind quality assessment approach for AGOIs based on deep feature statistics. Specifically, we first transform the AGOIs in equirectangular projection (ERP) format into a set of six cubemap projection (CMP)-converted viewport images, and extract viewport-wise multi-layer deep features from the pre-trained neural network backbone. Based on the deep representations, the multivariate Gaussian (MVG) models are subsequently fitted. The individual quality score for each CMP-converted image is calculated by comparing it against the corresponding fitted pristine MVG model. The final quality score for a testing AGOI is then computed by aggregating these individual quality scores. We conduct comprehensive experiments using the existing AGOIQA database and the experimental results show that the proposed OU-BAGOIQA model outperforms current state-of-the-art OU blind image quality assessment models. The ablation study has also been conducted to validate the effectiveness of our method.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"110 ","pages":"Article 104461"},"PeriodicalIF":2.6000,"publicationDate":"2025-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Communication and Image Representation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1047320325000756","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The advancement of artificial intelligence generated content (AIGC) and virtual reality (VR) technologies have prompted AI-generated omnidirectional images (AGOI) to gradually into people’s daily lives. Compared to natural omnidirectional images, AGOIs exhibit traditional low-level technical distortions and high-level semantic distortions, which can severely affect the immersive experience for users in practical applications. Consequently, there is an urgent need for thorough research and precise evaluation of AGOI quality. In this paper, we propose a novel opinion-unaware (OU) blind quality assessment approach for AGOIs based on deep feature statistics. Specifically, we first transform the AGOIs in equirectangular projection (ERP) format into a set of six cubemap projection (CMP)-converted viewport images, and extract viewport-wise multi-layer deep features from the pre-trained neural network backbone. Based on the deep representations, the multivariate Gaussian (MVG) models are subsequently fitted. The individual quality score for each CMP-converted image is calculated by comparing it against the corresponding fitted pristine MVG model. The final quality score for a testing AGOI is then computed by aggregating these individual quality scores. We conduct comprehensive experiments using the existing AGOIQA database and the experimental results show that the proposed OU-BAGOIQA model outperforms current state-of-the-art OU blind image quality assessment models. The ablation study has also been conducted to validate the effectiveness of our method.
期刊介绍:
The Journal of Visual Communication and Image Representation publishes papers on state-of-the-art visual communication and image representation, with emphasis on novel technologies and theoretical work in this multidisciplinary area of pure and applied research. The field of visual communication and image representation is considered in its broadest sense and covers both digital and analog aspects as well as processing and communication in biological visual systems.