Alberto Presta;Enzo Tartaglione;Attilio Fiandrotti;Marco Grangetto
{"title":"STanH: Parametric Quantization for Variable Rate Learned Image Compression","authors":"Alberto Presta;Enzo Tartaglione;Attilio Fiandrotti;Marco Grangetto","doi":"10.1109/TIP.2025.3527883","DOIUrl":null,"url":null,"abstract":"In end-to-end learned image compression, encoder and decoder are jointly trained to minimize a <inline-formula> <tex-math>$\\boldsymbol {R} \\boldsymbol {+} \\boldsymbol {\\lambda } \\boldsymbol {D}$ </tex-math></inline-formula> cost function, where <inline-formula> <tex-math>$\\boldsymbol {\\lambda }$ </tex-math></inline-formula> controls the trade-off between rate of the quantized latent representation and image quality. Unfortunately, a distinct encoder-decoder pair with millions of parameters must be trained for each <inline-formula> <tex-math>$\\boldsymbol {\\lambda }$ </tex-math></inline-formula>, hence the need to switch encoders and to store multiple encoders and decoders on the user device for every target rate. This paper proposes to exploit a differentiable quantizer designed around a parametric sum of hyperbolic tangents, called STanH, that relaxes the step-wise quantization function. STanH is implemented as a differentiable activation layer with learnable quantization parameters that can be plugged into a pre-trained fixed rate model and refined to achieve different target bitrates. Experimental results show that our method enables variable rate coding with comparable efficiency to the state-of-the-art, yet with significant savings in terms of ease of deployment, training time, and storage costs.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"34 ","pages":"639-651"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10843163/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In end-to-end learned image compression, encoder and decoder are jointly trained to minimize a $\boldsymbol {R} \boldsymbol {+} \boldsymbol {\lambda } \boldsymbol {D}$ cost function, where $\boldsymbol {\lambda }$ controls the trade-off between rate of the quantized latent representation and image quality. Unfortunately, a distinct encoder-decoder pair with millions of parameters must be trained for each $\boldsymbol {\lambda }$ , hence the need to switch encoders and to store multiple encoders and decoders on the user device for every target rate. This paper proposes to exploit a differentiable quantizer designed around a parametric sum of hyperbolic tangents, called STanH, that relaxes the step-wise quantization function. STanH is implemented as a differentiable activation layer with learnable quantization parameters that can be plugged into a pre-trained fixed rate model and refined to achieve different target bitrates. Experimental results show that our method enables variable rate coding with comparable efficiency to the state-of-the-art, yet with significant savings in terms of ease of deployment, training time, and storage costs.