Skip to content

kamisli-icpl/ImageCompressionLearnedLiftingandLearnedTreeBasedModels

 
 

Repository files navigation

IMAGE-COMPRESSION-WITH-LEARNED-LIFTING-BASED-DWT-AND-LEARNED-TREE-BASED-ENTROPY-MODELS

Abstract

This paper explores learned image compression based on traditional and learned discrete wavelet transform (DWT) architectures and learned entropy models for coding DWT subband coefficients. A learned DWT is obtained through the lifting scheme with learned nonlinear predict and update filters. Several learned entropy models are proposed to exploit inter and intra-DWT subband coefficient dependencies, akin to traditional EZW, SPIHT, or EBCOT algorithms. Experimental results show that when the proposed learned entropy models are combined with traditional wavelet filters, such as the CDF 9/7 filters, compression performance that far exceeds that of JPEG2000 can be achieved. When the learned entropy models are combined with the learned DWT, compression performance increases further. The computations in the learned DWT and all entropy models, except one, can be simply parallelized, and the systems provide practical encoding and decoding times on GPUs.

Citation

@article{sahin2023image,
  title={Image compression with learned lifting-based DWT and learned tree-based entropy models},
  author={Sahin, Ugur Berk and Kamisli, Fatih},
  journal={Multimedia Systems},
  volume={29},
  number={6},
  pages={3369--3384},
  year={2023},
  publisher={Springer}
}

Requirements

Required libraries can be found in requirements.txt pytorch-wavelets should be installed to use CDF97 filters with torch tensors. For experiments few weight files are uploaded can be downloaded via the link in weight_link.txt

It should be noted that this repo includes the Version1 files for the proposed work.

Note: Final updates will be updated right after the review process of the paper and will be shared as version2.

About

This repo includes the code files for IMAGE COMPRESSION WITH LEARNED LIFTING-BASED DWT AND LEARNED TREE-BASED ENTROPY MODELS paper

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 99.9%
  • Shell 0.1%