Yan Wang1, Tongtong Su1, Yusen Li1†, Jiuwen Cao2, Gang Wang1, Xiaoguang Liu1
1Nankai University, 2Hangzhou Dianzi University
Abstract: Recent research on deep convolutional neural networks (CNNs) has provided a significant performance boost on efficient super-resolution (SR) tasks by trading off the performance and applicability. However, most existing methods focus on subtracting feature processing consumption to reduce the parameters and calculations without refining the immediate features, which leads to inadequate information in the restoration. In this paper, we propose a lightweight network termed DDistill-SR, which significantly improves the SR quality by capturing and reusing more helpful information in a static-dynamic feature distillation manner. Specifically, we propose a plug-in reparameterized dynamic unit (RDU) to promote the performance and inference cost trade-off. During the training phase, the RDU learns to linearly combine multiple reparameterizable blocks by analyzing varied input statistics to enhance layer-level representation. In the inference phase, the RDU is equally converted to simple dynamic convolutions that explicitly capture robust dynamic and static feature maps. Then, the information distillation block is constructed by several RDUs to enforce hierarchical refinement and selective fusion of spatial context information. Furthermore, we propose a dynamic distillation fusion (DDF) module to enable dynamic signals aggregation and communication between hierarchical modules to further improve performance. Empirical results show that our DDistill-SR outperforms the baselines and achieves state-of-the-art results on most super-resolution domains with much fewer parameters and less computational overhead.
This repository contains PyTorch implementation for DDistill-SR (TMM 2023).
-
Training with EDSR framework or BasicSR framework.
-
Testing with five commenly used datasets.
- Pretrained models available at model-backup.
- Visual results available at Baidu Pan.
Our RDU is based on existing dynamic and reparameterized methods, thanks for their enlightening work!
@ARTICLE{9939085,
author={Wang, Yan and Su, Tongtong and Li, Yusen and Cao, Jiuwen and Wang, Gang and Liu, Xiaoguang},
journal={IEEE Transactions on Multimedia},
title={DDistill-SR: Reparameterized Dynamic Distillation Network for Lightweight Image Super-Resolution},
year={2022},
pages={1-13},
doi={10.1109/TMM.2022.3219646}}