TY - GEN
T1 - LEARNED IMAGE COMPRESSION FOR BOTH HUMANS AND MACHINES VIA DYNAMIC ADAPTATION
AU - Zhu, Lingyu
AU - Li, Binzhe
AU - Lu, Riyu
AU - Chen, Peilin
AU - Mao, Qi
AU - Wang, Zhao
AU - Yang, Wenhan
AU - Wang, Shiqi
N1 - Publisher Copyright:
© 2024 IEEE
PY - 2024
Y1 - 2024
N2 - Recent advancements in neural image compression have shown great potential in outperforming conventional standard codecs in terms of both rate-distortion and rate-analysis performance. However, there is an issue of divergent preferences in information preservation or reconstruction in the process of compression for humans and machines, respectively. Compression for humans tends to retain the signal fidelity or perceptual quality of visual appearance while compression for machines requires preserving critical semantic information, resulting in the limitation of the bitstream supporting only a single requirement during the compression. To bridge this gap, we propose a dynamic adaptation approach that generates a single bitstream serving both humans and machines. This approach aims to mitigate the domain gap among tasks, which facilitates maintaining the performance of out-of-scope tasks. Specifically, the proposed method concentrates on learning a dynamic adaptation process, i.e., optimizing the latent representation in the compressed domain in an end-to-end manner while adhering to the rate-performance constraint. Extensive results reveal that our paradigm significantly reduces the domain gap, surpassing existing codecs.
AB - Recent advancements in neural image compression have shown great potential in outperforming conventional standard codecs in terms of both rate-distortion and rate-analysis performance. However, there is an issue of divergent preferences in information preservation or reconstruction in the process of compression for humans and machines, respectively. Compression for humans tends to retain the signal fidelity or perceptual quality of visual appearance while compression for machines requires preserving critical semantic information, resulting in the limitation of the bitstream supporting only a single requirement during the compression. To bridge this gap, we propose a dynamic adaptation approach that generates a single bitstream serving both humans and machines. This approach aims to mitigate the domain gap among tasks, which facilitates maintaining the performance of out-of-scope tasks. Specifically, the proposed method concentrates on learning a dynamic adaptation process, i.e., optimizing the latent representation in the compressed domain in an end-to-end manner while adhering to the rate-performance constraint. Extensive results reveal that our paradigm significantly reduces the domain gap, surpassing existing codecs.
KW - dynamic adaptation
KW - human-machine collaboration
KW - Learned image compression
UR - https://www.scopus.com/pages/publications/85216901905
U2 - 10.1109/ICIP51287.2024.10647464
DO - 10.1109/ICIP51287.2024.10647464
M3 - 会议稿件
AN - SCOPUS:85216901905
T3 - Proceedings - International Conference on Image Processing, ICIP
SP - 1788
EP - 1794
BT - 2024 IEEE International Conference on Image Processing, ICIP 2024 - Proceedings
PB - IEEE Computer Society
T2 - 31st IEEE International Conference on Image Processing, ICIP 2024
Y2 - 27 October 2024 through 30 October 2024
ER -