Skip to content

ICLR 2024 (Spotlight) - SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution

Notifications You must be signed in to change notification settings

XPixelGroup/SEAL

Repository files navigation

SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution, ICLR 2024 Spotlight [Paper Link]

visitors

Wenlong zhang1,2, Xiaohui Li2,3, Xiangyu Chen2,4,5, Yu Qiao2,5, Xiaoming Wu1 and Chao Dong2,5

1The HongKong Polytechnic University 2Shanghai AI Laboratory 3Shanghai Jiao Tong University
4University of Macau 5Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences

Highlight

Our relative, distributed evaluation approach SEAL serves as a complement to existing evaluation methods that solely rely on absolute, average performance, addressing their limitations and providing a valuable alternative perspective for evaluation.

We consider SEAL as the first step towards creating an unbiased and comprehensive evaluation platform, which can promote the development of real-SR.



Fig. Our SEAL consists of a clustering-based approach for degradation space modeling and a set of metrics based on representative degradation cases.


This repository contains

  • SE benchmark test sets, including:

    • Set14-SE
    • Urban100-SE
    • DIV2K_val-SE
    • ...
  • Two reference lines:

    • Acceptance line
    • Excellence line
  • Two systemactic metrics:

    • AR (Acceptance Rate)
    • RPR (Relative Performance Ratio)
  • A coarse-to-fine evaluation protocol



Fig. A coarse-to-fine evaluation protocol to rank different real-SR models with the proposed metrics.


  • Visualization of distributional performance


Fig. Distribution results under our SEAL evaluation.


🆕Update

  • 2023.09.07: Repo is released.

🧗TODO

  • Release code and pretrained models:computer:.
  • Update SE test sets links:link:.

🔲Benchmark Results

SysTest Set14 PSNR-S $\uparrow$ AR $\uparrow$ RPR $_I\downarrow$ RPR $_A \downarrow$ RPR $_U \uparrow$
SRResNet 20.95 0.00 0.02 0.00 0.03
DASR 21.08 0.00 0.01 0.00 0.02
BSRNet 22.77 0.59 0.42 0.72 0.27
RealESRNet 22.67 0.27 0.28 0.63 0.28
RDSR 22.44 0.08 0.23 0.63 0.21
RealESRNet-GD 22.82 0.43 0.37 0.74 0.33
SwinIR 22.61 0.41 0.24 0.58 0.29

Installation

# 
cd SEAL
pip install -r requirements.txt
python setup.py develop

Data Preparation

Download SE test sets from google drive or quark pan. Put them to datasets/.

or Generate NEW SE test sets by

python scripts/data_generation/data_generation.py

Acceptance and Excellence lines

Download acceptance and excellence lines from google drive or quark pan. Put them in modelzoo/.

Inference the real-SR model on SE test sets

  • Inference Real-SR model on the SE test sets provided by us.
python scripts/inference/inference_SE.py
  • For new SE test sets:

    • Inference the acceptance and excellence lines on the new SE test sets.
    • Inference Real-SR model on the new SE test sets.
python scripts/inference/inference_SE.py

Get the common-used IQA performance on SE test sets

python scripts/metrics/cal_psnr_ssim.py # It includes LPIPS and NIQE

The results are saved in a CSV file with each line named in form 'model name_ test metrics'(such as line.csv and model.csv).

Get AR and RPR performance

python scripts/metrics/calculate_AR_RPR.py # It includes LPIPS and NIQE

Citation

@article{2023seal,
  author    = {Wenlong Zhang, Xiaohui Li, Xiangyu Chen, Yu Qiao, Xiao-Ming Wu, Chao Dong},
  title     = {SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution},
  journal   = {arxiv},
  year      = {2023},
}

Contact

If you have any question, please email wenlong.zhang@connect.polyu.hk.

About

ICLR 2024 (Spotlight) - SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages