Compressing Deep Image Super-resolution Models

University of Bristol

Abstract

Deep learning techniques have been applied in the context of image super-resolution (SR), achieving remarkable advances in terms of reconstruction performance. Existing techniques typically employ highly complex model structures which result in large model sizes and slow inference speeds. This often leads to high energy consumption and restricts their adoption for practical applications. To address this issue, this work employs a three-stage workflow for compressing deep SR models which significantly reduces their memory requirement. Restoration performance has been maintained through teacher-student knowledge distillation using a newly designed distillation loss. We have applied this approach to two popular image super-resolution networks, SwinIR and EDSR, to demonstrate its effectiveness. The resulting compact models, SwinIRmini and EDSRmini, attain an 89% and 96% reduction in both model size and floating-point operations (FLOPs) respectively, compared to their original versions. They also retain competitive super-resolution performance compared to their original models and other commonly used SR approaches.


Proposed workflow


Evaluation results

(Top) Average PSNR scores on Set5 and Set14 for various models versus their corresponding numbers of model parameters. Results for our two compact models are marked with solid symbols. (Bottom) The plot between the performance and FLOPs based on Set5 and Set14


Source code

[Download] source code from github.

Citation

@article{jiang2023compressing,
  title={Compressing Deep Image Super-resolution Models},
  author={Jiang, Yuxuan and Nawala, Jakub and Zhang, Fan and Bull, David},
  journal={arXiv preprint arXiv:2401.00523},
  year={2023}
}[paper]