Sühring, KarstenKarstenSühringSchäfer, MichaelMichaelSchäferPfaff, JonathanJonathanPfaffSchwarz, HeikoHeikoSchwarzMarpe, DetlevDetlevMarpeWiegand, ThomasThomasWiegand2023-06-282023-06-282022https://publica.fraunhofer.de/handle/publica/44490610.1109/ICIP46576.2022.98976852-s2.0-85146718319The performance of variational auto-encoders (VAE) for image compression has steadily grown in recent years, thus becoming competitive with advanced visual data compression technologies. These neural networks transform the source image into a latent space with a channel-wise representation. In most works, the latents are scalar quantized before being entropy coded. On the other hand, vector quantizers generally achieve denser packings of high-dimensional data regardless of the source distribution. Hence, low-complexity variants of these quantizers are implemented in the compression standards JPEG 2000 and Versatile Video Coding. In this paper we demonstrate coding gains by using trellis-coded quantization (TCQ) over scalar quantization. For the optimization of the networks with regard to TCQ, we employ a specific noisy representation of the features during the training stage. For variable-rate VAEs, we obtained 7.7% average BD-rate savings on the Kodak images by using TCQ over scalar quantization. When different networks per target bitrate are optimized, we report a relative coding gain of 2.4% due to TCQ.enAuto-EncoderDeep LearningRate-Distortion-OptimizationTrellis-Coded QuantizationTrellis-Coded Quantization for End-to-End Learned Image Compressionconference paper