Lightweight Width-Depth Scalable Implicit Neural Representation for Progressive Image Compression

IF 10.9 2区 计算机科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Qingyu Mao;Wenming Wang;Yongsheng Liang;Chenhu Xiao;Fanyang Meng;Gwanggil Jeon
{"title":"Lightweight Width-Depth Scalable Implicit Neural Representation for Progressive Image Compression","authors":"Qingyu Mao;Wenming Wang;Yongsheng Liang;Chenhu Xiao;Fanyang Meng;Gwanggil Jeon","doi":"10.1109/TCE.2025.3565495","DOIUrl":null,"url":null,"abstract":"Image compression approaches using implicit neural representation (INR) have recently gained attention for their lightweight nature, compactness, and fast decoding, showing promise for edge computing in consumer devices. Specifically, INR-based image compression methods implicitly store each image within a lightweight neural network, which serves as a compact representation of the image. However, most existing methods are limited to representing single-quality images with fixed-size models, which necessitates training separate models independently for images at varying quality levels, leading to additional training and storage costs. To tackle this problem, we propose a progressive image compression method based on Width-Depth Scalable Implicit Neural Representation (WDS-INR), which are composed of executable sub-networks of varying scales. By adjusting the scale of the sub-networks, WDS-INR can represent images at different quality levels while supporting progressive transmission. The scalable architecture of WDS-INR makes it well-suited for deployment on mobile and IoTs devices. Furthermore, we propose a band-limited initialization scheme that enhances both the representation capabilities and training stability of the WDS-INR. Finally, we introduce a meta-learning approach to the base sub-network to accelerate encoding <inline-formula> <tex-math>$(4 \\times \\text { faster})$ </tex-math></inline-formula>. Experimental results demonstrate that the proposed method outperforms the baseline in rate-distortion performance <inline-formula> <tex-math>$(+ 0.28~dB {~\\text {PSNR}})$ </tex-math></inline-formula>, while enabling scalable bit-rates with progressive decoding.","PeriodicalId":13208,"journal":{"name":"IEEE Transactions on Consumer Electronics","volume":"71 2","pages":"5666-5675"},"PeriodicalIF":10.9000,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Consumer Electronics","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10979971/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Image compression approaches using implicit neural representation (INR) have recently gained attention for their lightweight nature, compactness, and fast decoding, showing promise for edge computing in consumer devices. Specifically, INR-based image compression methods implicitly store each image within a lightweight neural network, which serves as a compact representation of the image. However, most existing methods are limited to representing single-quality images with fixed-size models, which necessitates training separate models independently for images at varying quality levels, leading to additional training and storage costs. To tackle this problem, we propose a progressive image compression method based on Width-Depth Scalable Implicit Neural Representation (WDS-INR), which are composed of executable sub-networks of varying scales. By adjusting the scale of the sub-networks, WDS-INR can represent images at different quality levels while supporting progressive transmission. The scalable architecture of WDS-INR makes it well-suited for deployment on mobile and IoTs devices. Furthermore, we propose a band-limited initialization scheme that enhances both the representation capabilities and training stability of the WDS-INR. Finally, we introduce a meta-learning approach to the base sub-network to accelerate encoding $(4 \times \text { faster})$ . Experimental results demonstrate that the proposed method outperforms the baseline in rate-distortion performance $(+ 0.28~dB {~\text {PSNR}})$ , while enabling scalable bit-rates with progressive decoding.
渐进式图像压缩的轻量级宽度-深度可扩展隐式神经网络表示
使用隐式神经表示(INR)的图像压缩方法最近因其轻量级、紧凑性和快速解码而受到关注,在消费设备的边缘计算中显示出前景。具体来说,基于inr的图像压缩方法隐式地将每个图像存储在一个轻量级的神经网络中,该神经网络作为图像的紧凑表示。然而,大多数现有方法仅限于用固定大小的模型表示单一质量的图像,这需要为不同质量水平的图像独立训练单独的模型,从而导致额外的训练和存储成本。为了解决这个问题,我们提出了一种基于宽度-深度可扩展隐式神经表示(WDS-INR)的渐进式图像压缩方法,该方法由不同规模的可执行子网络组成。通过调整子网的规模,WDS-INR可以表示不同质量水平的图像,同时支持逐行传输。WDS-INR的可扩展架构使其非常适合在移动和物联网设备上部署。此外,我们提出了一种带限初始化方案,增强了WDS-INR的表示能力和训练稳定性。最后,我们在基本子网络中引入了一种元学习方法来加速编码$(4 \times \text {faster})$。实验结果表明,该方法在率失真性能$(+ 0.28~dB {~\text {PSNR}})$上优于基线,同时实现了可扩展的比特率和渐进解码。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.70
自引率
9.30%
发文量
59
审稿时长
3.3 months
期刊介绍: The main focus for the IEEE Transactions on Consumer Electronics is the engineering and research aspects of the theory, design, construction, manufacture or end use of mass market electronics, systems, software and services for consumers.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信