Spatial Competition for Low-Complexity Learned Image Compression
Abstract
Autoencoder-based image codecs achieve state-of-the-art compression performance but often incur high computational complexity, particularly at decoding time. This work introduces a low-complexity learned image compression framework based on spatial competition between multiple specialized neural codecs. For each image region, the encoder selects the codec that best matches the local content according to a rate-distortion cost. A mode map is transmitted as side information to indicate the per-reg...
Description / Details
Autoencoder-based image codecs achieve state-of-the-art compression performance but often incur high computational complexity, particularly at decoding time. This work introduces a low-complexity learned image compression framework based on spatial competition between multiple specialized neural codecs. For each image region, the encoder selects the codec that best matches the local content according to a rate-distortion cost. A mode map is transmitted as side information to indicate the per-region codec selection. At decoding time, this mode map-based selection guides reconstruction while preserving the complexity of a single codec. This design enables per-image adaptation with low decoding complexity and fast encoding. On the CLIC 2020 dataset, our method achieves up to -14.5% rate reduction compared to a single codec and reaches HEVC-level performance with a decoding complexity of 1433 MACs per pixel.
Source: arXiv:2605.13243v1 - http://arxiv.org/abs/2605.13243v1 PDF: https://arxiv.org/pdf/2605.13243v1 Original Link: http://arxiv.org/abs/2605.13243v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
May 14, 2026
Biomedical Engineering
Engineering
0