TY - JOUR
T1 - STanH
T2 - Parametric Quantization for Variable Rate Learned Image Compression
AU - Presta, Alberto
AU - Tartaglione, Enzo
AU - Fiandrotti, Attilio
AU - Grangetto, Marco
N1 - Publisher Copyright:
© 1992-2012 IEEE.
PY - 2025/1/1
Y1 - 2025/1/1
N2 - In end-to-end learned image compression, encoder and decoder are jointly trained to minimize a R + λ D cost function, where λ controls the trade-off between rate of the quantized latent representation and image quality. Unfortunately, a distinct encoder-decoder pair with millions of parameters must be trained for each λ, hence the need to switch encoders and to store multiple encoders and decoders on the user device for every target rate. This paper proposes to exploit a differentiable quantizer designed around a parametric sum of hyperbolic tangents, called STanH, that relaxes the step-wise quantization function. STanH is implemented as a differentiable activation layer with learnable quantization parameters that can be plugged into a pre-trained fixed rate model and refined to achieve different target bitrates. Experimental results show that our method enables variable rate coding with comparable efficiency to the state-of-the-art, yet with significant savings in terms of ease of deployment, training time, and storage costs.
AB - In end-to-end learned image compression, encoder and decoder are jointly trained to minimize a R + λ D cost function, where λ controls the trade-off between rate of the quantized latent representation and image quality. Unfortunately, a distinct encoder-decoder pair with millions of parameters must be trained for each λ, hence the need to switch encoders and to store multiple encoders and decoders on the user device for every target rate. This paper proposes to exploit a differentiable quantizer designed around a parametric sum of hyperbolic tangents, called STanH, that relaxes the step-wise quantization function. STanH is implemented as a differentiable activation layer with learnable quantization parameters that can be plugged into a pre-trained fixed rate model and refined to achieve different target bitrates. Experimental results show that our method enables variable rate coding with comparable efficiency to the state-of-the-art, yet with significant savings in terms of ease of deployment, training time, and storage costs.
KW - Learned image compression
KW - differentiable quantization
KW - quantizer annealing
KW - variable rate image coding
UR - https://www.scopus.com/pages/publications/85215608158
U2 - 10.1109/TIP.2025.3527883
DO - 10.1109/TIP.2025.3527883
M3 - Article
AN - SCOPUS:85215608158
SN - 1057-7149
VL - 34
SP - 639
EP - 651
JO - IEEE Transactions on Image Processing
JF - IEEE Transactions on Image Processing
ER -