SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution, ICLR 2024 Spotlight [Paper Link]
Wenlong zhang1,2, Xiaohui Li2,3, Xiangyu Chen2,4,5, Yu Qiao2,5, Xiaoming Wu1 and Chao Dong2,5
1The HongKong Polytechnic University
2Shanghai AI Laboratory
3Shanghai Jiao Tong University
4University of Macau
5Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences
Our relative, distributed evaluation approach SEAL serves as a complement to existing evaluation methods that solely rely on absolute, average performance, addressing their limitations and providing a valuable alternative perspective for evaluation.
We consider SEAL as the first step towards creating an unbiased and comprehensive evaluation platform, which can promote the development of real-SR.
Fig. Our SEAL consists of a clustering-based approach for degradation space modeling and a set of metrics based on representative degradation cases.
-
SE benchmark test sets, including:
- Set14-SE
- Urban100-SE
- DIV2K_val-SE
- ...
-
Two reference lines:
- Acceptance line
- Excellence line
-
Two systemactic metrics:
- AR (Acceptance Rate)
- RPR (Relative Performance Ratio)
-
A coarse-to-fine evaluation protocol
Fig. A coarse-to-fine evaluation protocol to rank different real-SR models with the proposed metrics.
- Visualization of distributional performance
Fig. Distribution results under our SEAL evaluation.
- 2023.09.07: Repo is released.
- Release code and pretrained models:computer:.
- Update SE test sets links:link:.
SysTest Set14 | PSNR-S |
AR |
RPR |
RPR |
RPR |
---|---|---|---|---|---|
SRResNet | 20.95 | 0.00 | 0.02 | 0.00 | 0.03 |
DASR | 21.08 | 0.00 | 0.01 | 0.00 | 0.02 |
BSRNet | 22.77 | 0.59 | 0.42 | 0.72 | 0.27 |
RealESRNet | 22.67 | 0.27 | 0.28 | 0.63 | 0.28 |
RDSR | 22.44 | 0.08 | 0.23 | 0.63 | 0.21 |
RealESRNet-GD | 22.82 | 0.43 | 0.37 | 0.74 | 0.33 |
SwinIR | 22.61 | 0.41 | 0.24 | 0.58 | 0.29 |
#
cd SEAL
pip install -r requirements.txt
python setup.py develop
Download SE test sets from google drive or quark pan. Put them to datasets/
.
or Generate NEW SE test sets by
python scripts/data_generation/data_generation.py
Download acceptance and excellence lines from google drive or quark pan. Put them in modelzoo/
.
- Inference Real-SR model on the SE test sets provided by us.
python scripts/inference/inference_SE.py
-
For new SE test sets:
python scripts/inference/inference_SE.py
python scripts/metrics/cal_psnr_ssim.py # It includes LPIPS and NIQE
The results are saved in a CSV file with each line named in form 'model name_ test metrics'(such as line.csv and model.csv).
python scripts/metrics/calculate_AR_RPR.py # It includes LPIPS and NIQE
@article{2023seal,
author = {Wenlong Zhang, Xiaohui Li, Xiangyu Chen, Yu Qiao, Xiao-Ming Wu, Chao Dong},
title = {SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution},
journal = {arxiv},
year = {2023},
}
If you have any question, please email wenlong.zhang@connect.polyu.hk.