Signal Processing-Image Communication最新文献

筛选
英文 中文
Few-shot image generation based on meta-learning and generative adversarial network 基于元学习和生成对抗网络的少镜头图像生成
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-28 DOI: 10.1016/j.image.2025.117307
Bowen Gu, Junhai Zhai
{"title":"Few-shot image generation based on meta-learning and generative adversarial network","authors":"Bowen Gu,&nbsp;Junhai Zhai","doi":"10.1016/j.image.2025.117307","DOIUrl":"10.1016/j.image.2025.117307","url":null,"abstract":"<div><div>Generative adversarial network (GAN) learns the latent distribution of samples through the adversarial training between discriminator and generator, then uses the learned probability distribution to generate realistic samples. Training a vanilla GAN requires a large number of samples and a significant amount of time. However, in practical applications, obtaining a large dataset and dedicating extensive time to model training can be very costly. Training a GAN with a small number of samples to generate high-quality images is a pressing research problem. Although this area has seen limited exploration, FAML (Fast Adaptive Meta-Learning) stands out as a notable approach. However, FAML has the following shortcomings: (1) The training time on complex datasets, such as VGGFaces and MiniImageNet, is excessively long. (2) It exhibits poor generalization performance and produces low-quality images across different datasets. (3) The generated samples lack diversity. To address the three shortcomings, we improved FAML in two key areas: model structure and loss function. The improved model effectively overcomes all three limitations of FAML. We conducted extensive experiments on four datasets to compare our model with the baseline FAML across seven evaluation metrics. The results demonstrate that our model is both more efficient and effective, particularly on the two complex datasets, VGGFaces and MiniImageNet. Our model outperforms FAML on six of the seven evaluation metrics, with only a slight underperformance on one metric. Our code is available at <span><span>https://github.com/BTGWS/FSML-GAN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"137 ","pages":"Article 117307"},"PeriodicalIF":3.4,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143760468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
OTPL: A novel measurement method of structural parallelism based on orientation transformation and geometric constraints OTPL:一种基于方位变换和几何约束的结构平行度测量方法
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-25 DOI: 10.1016/j.image.2025.117310
Weili Ding , Zhiyu Wang , Shuo Hu
{"title":"OTPL: A novel measurement method of structural parallelism based on orientation transformation and geometric constraints","authors":"Weili Ding ,&nbsp;Zhiyu Wang ,&nbsp;Shuo Hu","doi":"10.1016/j.image.2025.117310","DOIUrl":"10.1016/j.image.2025.117310","url":null,"abstract":"<div><div>Detecting parallel geometric structures from images is a significant step for computer vision tasks. In this paper, an algorithm called Orientation Transformation-based Parallelism Measurement (OTPL) is proposed in this paper to measure the parallelism of structures including both line structures and curve structures. The task is decomposed into measurements of parallel straight line and parallel curve structures due to the inherent geometric differences between them, where the parallelism between curve structures can be further transformed into a matching problem. For parallel straight lines, the angle constraints and the rate of overlapping projection are considered as the parallel relationship selection rules for the candidate lines. For the parallel curves, the approximate vertical growing (AVG) algorithm is proposed to accelerate the search of adjacent curves and each smooth curve is coded as a vector with different angle values. The matching pairs are extracted through cosine similarity transformation and convexity consistency. Finally, the parallel curves are extracted by a decision-making process. The proposed algorithm is evaluated in a comprehensive manner, encompassing both qualitative and quantitative approaches, with the objective of achieving a more robust assessment. The results demonstrate the algorithm's efficacy in identifying parallel structures in both synthetic and natural images.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"137 ","pages":"Article 117310"},"PeriodicalIF":3.4,"publicationDate":"2025-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143760467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bidirectional interactive multi-scale network using Wave-Conv ViT for single image deraining 基于波变换ViT的单幅图像双向交互多尺度网络
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-23 DOI: 10.1016/j.image.2025.117311
Siyan Fang, Bin Liu
{"title":"Bidirectional interactive multi-scale network using Wave-Conv ViT for single image deraining","authors":"Siyan Fang,&nbsp;Bin Liu","doi":"10.1016/j.image.2025.117311","DOIUrl":"10.1016/j.image.2025.117311","url":null,"abstract":"<div><div>To address the limitations of high-frequency information capture by Vision Transformer (ViT) and the loss of fine details in existing image deraining methods, we introduce a Bidirectional Interactive Multi-Scale Network (BIMNet) that employs newly developed Wave-Conv ViT (WCV). The WCV utilizes a wavelet transform to enable self-attention in both low-frequency and high-frequency domains, significantly enhancing ViT's capacity for diverse frequency-domain feature modeling. Additionally, by incorporating convolutional operations, WCV enhances the extraction and integration of local features across various spatial windows. BIMNet injects rainy images into deep network layers, enabling bidirectional propagation with shallow layer features that enrich skip connections with detailed and complementary information, thus improving the fidelity of detail recovery. Moreover, we present the CORain1000 dataset, tailored for the dual challenges of image deraining and object detection, which offers more diversity in rain patterns, image sizes, and volumes than the commonly used COCO350 dataset. Extensive experiments demonstrate the superiority of BIMNet over advanced methods. The code and CORain1000 dataset are available at <span><span>https://github.com/fashyon/BIMNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"137 ","pages":"Article 117311"},"PeriodicalIF":3.4,"publicationDate":"2025-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143792164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Feature Extraction and Knowledge Distillation Based Deep Learning Model for Human Activity Recognition System 基于混合特征提取和知识蒸馏的人体活动识别深度学习模型
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-21 DOI: 10.1016/j.image.2025.117308
Hetal Shah , Mehfuza S. Holia
{"title":"Hybrid Feature Extraction and Knowledge Distillation Based Deep Learning Model for Human Activity Recognition System","authors":"Hetal Shah ,&nbsp;Mehfuza S. Holia","doi":"10.1016/j.image.2025.117308","DOIUrl":"10.1016/j.image.2025.117308","url":null,"abstract":"<div><div>This article introduces the Generative Adversarial Network (GAN) framework model, which uses offline knowledge distillation (KD) to move spatio-deep data from a large teacher to a smaller student model. To achieve this, the teacher model named EfficientNetB7 embedded with spatial attention (E2SA) and a multi-layer Gated Recurrent Unit (GRU) is used. A hybrid feature extraction method known as Completed Hybrid Local Binary Pattern (ChLBP) is employed prior to the prediction process. After feature extraction, the hybrid features are parallelly given as input to both teacher and student models. In the teacher model, E2SA extracts both deep and spatio attention activity features, and these features are then input to the multi-layer GRU, which learns the human activity frame sequences overall. The proposed model obtains 98.50 % recognition accuracy on the UCF101 dataset and 79.21 % recognition accuracy on the HMDB51 dataset, which is considerably better than the existing models.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"137 ","pages":"Article 117308"},"PeriodicalIF":3.4,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143826393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain-guided multi-frequency underwater image enhancement network 域制导多频水下图像增强网络
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-17 DOI: 10.1016/j.image.2025.117281
Qingzheng Wang, Bin Li, Ge Shi, Xinyu Wang, Yiliang Chen
{"title":"Domain-guided multi-frequency underwater image enhancement network","authors":"Qingzheng Wang,&nbsp;Bin Li,&nbsp;Ge Shi,&nbsp;Xinyu Wang,&nbsp;Yiliang Chen","doi":"10.1016/j.image.2025.117281","DOIUrl":"10.1016/j.image.2025.117281","url":null,"abstract":"<div><div>The distribution of underwater images exhibits diverse due to the varied scattering and absorption of light in different water types. However, most existing methods have significant limitations as they cannot distinguish the difference between different water types during enhancement processing, and do not propose clear solutions for the different frequency information. Therefore, the key challenge is to achieve consistency between learned features and water types while preserving multi-frequency information. Thus, we propose a domain-guided multi-frequency underwater image enhancement network (DGMF), which generate high quality images by learning water-type-related features and capturing multi-frequency information. Specifically, we introduce a domain-aware module equipped with a water type classifier, which can distinguish the impacts of different water types, and guide the update of the model towards the specific domain. In addition, we design a multi-frequency mixer that couples Multi-Group Convolution (MGC) and Global Sparse Attention (GSA) to more effectively captures local and global information. Extensive experiments demonstrate that our method outperforms most state-of-the-art methods in both visual perception and evaluation metrics. The code is publicly available at <span><span>https://github.com/liyoucai699/DGMF.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117281"},"PeriodicalIF":3.4,"publicationDate":"2025-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143686524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DMCTDet: A density map-guided composite transformer network for object detection of UAV images DMCTDet:一种用于无人机图像目标检测的密度图引导复合变压器网络
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-14 DOI: 10.1016/j.image.2025.117284
Junjie Li , Si Guo , Shi Yi , Runhua He , Yong Jia
{"title":"DMCTDet: A density map-guided composite transformer network for object detection of UAV images","authors":"Junjie Li ,&nbsp;Si Guo ,&nbsp;Shi Yi ,&nbsp;Runhua He ,&nbsp;Yong Jia","doi":"10.1016/j.image.2025.117284","DOIUrl":"10.1016/j.image.2025.117284","url":null,"abstract":"<div><div>The application of unmanned aerial vehicles (UAVs) in urban scene object detection is a vital area of research in urban planning, intelligent monitoring, disaster prevention, and urban surveillance.e However, detecting objects in urban scenes captured by UAVs is a challenging task mainly due to the small size of the objects, the variability within the same class, and the diversity of objects. To design an object detection network that can be applied to complex urban scenes, this study proposes a novel composite transformer object detection network guided by a density map (DMCTDet) for urban scene detection in UAV images. The distributional a priori information of objects can be fully exploited by density maps. In the detection stage, a composite backbone feature extraction network is constructed by Swin Transformer combined with Vision Longformer, which can fully extract the scale-variation objects. Adaptive multiscale feature pyramid enhancement modules (AMFPEM) are inserted in the feature fusion stage between both Swin Transformer and Vision Longformer to learn the relationship between object scale variation and enhance the feature representation capacity of small objects. In this way, the accuracy of urban scene detection is significantly improved, and weak aggregated objects are successfully detected from UAV images. Extensive ablation experiments and comparison experiments. of the proposed network are conducted on publicly available urban scene detection datasets of UAV images. The experimental results demonstrate the effectiveness of the designed network structure and the superiority of the proposed network compared to state-of-the-art methods in terms of detection accuracy.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117284"},"PeriodicalIF":3.4,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143817208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards human society-inspired decentralized DNN inference 面向人类社会的去中心化DNN推理
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-13 DOI: 10.1016/j.image.2025.117306
Dimitrios Papaioannou, Vasileios Mygdalis, Ioannis Pitas
{"title":"Towards human society-inspired decentralized DNN inference","authors":"Dimitrios Papaioannou,&nbsp;Vasileios Mygdalis,&nbsp;Ioannis Pitas","doi":"10.1016/j.image.2025.117306","DOIUrl":"10.1016/j.image.2025.117306","url":null,"abstract":"<div><div>In human societies, individuals make their own decisions and they may select if and who may influence it, by e.g., consulting with people of their acquaintance or experts of a field. At a societal level, the overall knowledge is preserved and enhanced by individual person empowerment, where complicated consensus protocols have been developed over time in the form of societal mechanisms to assess, weight, combine and isolate individual people opinions. In distributed machine learning environments however, individual AI agents are merely part of a system where decisions are made in a centralized and aggregated fashion or require a fixed network topology, a practice prone to security risks and collaboration is nearly absent. For instance, Byzantine Failures may tamper both the training and inference stage of individual AI agents, leading to significantly reduced overall system performance. Inspired by societal practices, we propose a decentralized inference strategy where each individual agent is empowered to make their own decisions, by exchanging and aggregating information with other agents in their network. To this end, a “Quality of Inference” consensus protocol (QoI) is proposed, forming a single commonly accepted inference rule applied by every individual agent. The overall system knowledge and decisions on specific manners can thereby be stored by all individual agents in a decentralized fashion, employing e.g., blockchain technology. Our experiments in classification tasks indicate that the proposed approach forms a secure decentralized inference framework, that prevents adversaries at tampering the overall process and achieves comparable performance with centralized decision aggregation methods.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117306"},"PeriodicalIF":3.4,"publicationDate":"2025-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143643817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive spatially regularized target attribute-aware background suppressed deep correlation filter for object tracking 自适应空间正则化目标属性感知背景抑制深度相关滤波器
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-11 DOI: 10.1016/j.image.2025.117305
Sathiyamoorthi Arthanari, Sathishkumar Moorthy, Jae Hoon Jeong, Young Hoon Joo
{"title":"Adaptive spatially regularized target attribute-aware background suppressed deep correlation filter for object tracking","authors":"Sathiyamoorthi Arthanari,&nbsp;Sathishkumar Moorthy,&nbsp;Jae Hoon Jeong,&nbsp;Young Hoon Joo","doi":"10.1016/j.image.2025.117305","DOIUrl":"10.1016/j.image.2025.117305","url":null,"abstract":"<div><div>In recent years, deep feature-based correlation filters have attained impressive performance in robust object tracking. However, deep feature-based correlation filters are affected by undesired boundary effects, which reduce the tracking performance. Moreover, the tracker moves towards a region that is identical to the target due to the sudden variation in target appearance and complicated background areas. To overcome these issues, we propose an adaptive spatially regularized target attribute-aware background suppressed deep correlation filter (ASTABSCF). To do this, a novel adaptive spatially regularized technique is presented, which aims to learn an efficient spatial weight for a particular object and fast target appearance variations. Specifically, we present a target-aware background suppression method with dual regression approach, which utilizes a saliency detection technique to produce the target mask. In this technique, we employ the global and target features to get the dual filters known as the global and target filters. Accordingly, global and target response maps are produced by dual filters, which are integrated into the detection stage to optimize the target response. In addition, a novel adaptive attribute-aware approach is presented to emphasize channel-specific discriminative features, which implements a post-processing technique on the observed spatial patterns to reduce the influence of less prominent channels. Therefore, the learned adaptive spatial attention patterns significantly reduce the irrelevant information of multi-channel features and improve the tracker performance. Finally, we demonstrate the efficiency of the ASTABSCF approach against existing modern trackers using the OTB-2013, OTB-2015, TempleColor-128, UAV-123, LaSOT, and GOT-10K benchmark datasets.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117305"},"PeriodicalIF":3.4,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143686523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lost in light field compression: Understanding the unseen pitfalls in computer vision 迷失于光场压缩:理解计算机视觉中看不见的陷阱
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-03-10 DOI: 10.1016/j.image.2025.117304
Adam Zizien , Chiara Galdi , Karel Fliegel , Jean-Luc Dugelay
{"title":"Lost in light field compression: Understanding the unseen pitfalls in computer vision","authors":"Adam Zizien ,&nbsp;Chiara Galdi ,&nbsp;Karel Fliegel ,&nbsp;Jean-Luc Dugelay","doi":"10.1016/j.image.2025.117304","DOIUrl":"10.1016/j.image.2025.117304","url":null,"abstract":"<div><div>Could we be overlooking a fundamental aspect of light fields in our quest for efficient compression? The vast amount of data enclosed in a light field makes compression a necessity. Yet, from an application point of view, the focus is predominantly on visual consumption while light fields have properties that can potentially be used in various other tasks. This paper examines the impact of light field compression on the performance of subsequent computer vision tasks. We investigate the variations in quality across perspectives and their impact on face recognition systems and disparity estimation. By leveraging a diverse dataset of light field images, we thoroughly evaluate the performance of various face recognition algorithms when subjected to different conventional and learning-based compression techniques, such as JPEG Pleno, ALVC, and SADN-QVRF. Our findings reveal a noticeable decline in peak recognition performance as compression levels increase, given specific recognition frameworks. Furthermore, we identify a significant shift in the recognition threshold, particularly in response to higher degrees of compression. Secondly, by relying on a novel disparity estimation algorithm, we explore the loss of information across light field perspectives. Our results highlight a disconnect between the preservation of visual fidelity and the loss of minute detail crucial for the preservation of disparity information in light field images. The findings presented herein aim to contribute to the development of efficient compression strategies while emphasizing the delicate balance between compression efficiency, subjective quality, and feature preservation with the aim of increased accuracy in specialized light field systems.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117304"},"PeriodicalIF":3.4,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143620241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Underwater image quality assessment method via the fusion of visual and structural information 通过融合视觉和结构信息评估水下图像质量的方法
IF 3.4 3区 工程技术
Signal Processing-Image Communication Pub Date : 2025-02-27 DOI: 10.1016/j.image.2025.117285
Tianhai Chen , Xichen Yang , Tianshu Wang , Nengxin Li , Shun Zhu , Genlin Ji
{"title":"Underwater image quality assessment method via the fusion of visual and structural information","authors":"Tianhai Chen ,&nbsp;Xichen Yang ,&nbsp;Tianshu Wang ,&nbsp;Nengxin Li ,&nbsp;Shun Zhu ,&nbsp;Genlin Ji","doi":"10.1016/j.image.2025.117285","DOIUrl":"10.1016/j.image.2025.117285","url":null,"abstract":"<div><div>Underwater-captured images often suffer from quality degradation due to the challenging underwater environment, leading to information loss that significantly affects their usability. Therefore, accurately predicting the quality of underwater images is crucial. To tackle this issue, this study introduces a novel Underwater Image Quality Assessment method that combines visual and structural information. First, the CIELab map, gradient feature map, and Mean Subtracted Contrast Normalized feature map of the underwater image are obtained. Then, these feature maps are divided into non-overlapping 32x32 patches, and each patch is fed into the corresponding sub-network. This method allows for a comprehensive description of the changes in visual and structural information resulting from quality degradation in underwater images. Subsequently, the features extracted by the multipath network are fused using a feature fusion network to promote feature complementarity and overcome the limitations of individual features. Finally, the relationship between underwater image quality and fusion features was learned to obtain an evaluation model. Furthermore, the quality of the underwater image can be measured by combining the quality prediction scores of different patches. Experimental results on underwater image datasets demonstrate that the proposed method can achieve more accurate and stable quality measurement results with a more lightweight structure. Meanwhile, performance comparisons on natural image datasets and screen content image datasets confirm that the proposed method is more applicable for complex application scenarios than existing methods. The code is open-source and available at <span><span>https://github.com/dart-into/UIQAVSI</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49521,"journal":{"name":"Signal Processing-Image Communication","volume":"136 ","pages":"Article 117285"},"PeriodicalIF":3.4,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143527063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信