In collaboration with the NTIRE workshop, we are hosting a challenge focused on Efficient Super-Resolution (NTIRE2024_ESR). This involves the task of enhancing the resolution of an input image by a factor of x4, utilizing a set of pre-existing examples comprising both low-resolution and their corresponding high-resolution images. The challenge encompasses one π main track which consists of three π sub-tracks, i.e., the Inference Runtime, FLOPs (Floating Point Operations Per Second), and Parameters. The baseline method in NTIRE2024_ESR is RLFN (Kong, et al, 2022), the winner of NTIRE2022 Efficient Super-Resolution Challenge. Details are shown below:
-
π Main-track: Overall Performance (Runtime, Parameters, FLOPs,) the aim is to obtain a network design / solution with the best overall performance in terms of inference runtime, FLOPS, and parameters on a common GPU (i.e., NVIDIA GeForce RTX 3090 GPU) while being constrained to maintain or improve the PSNR results.
-
π Sub-track 1: Inference Runtime, the aim is to obtain a network design / solution with the lowest inference time (runtime) on a common GPU (i.e., NVIDIA GeForce RTX 3090 GPU) while being constrained to maintain or improve over the baseline method RLFN in terms of number of parameters, FLOPs, and the PSNR result.
-
π Sub-track 2: FLOPs, the aim is to obtain a network design / solution with the lowest amount of FLOPs on a common GPU (i.e., NVIDIA GeForce RTX 3090 GPU) while being constrained to maintain or improve the inference runtime, the parameters, and the PSNR results of RLFN.
-
π Sub-track 3: Parameters, the aim is to obtain a network design / solution with the lowest amount of parameters on a common GPU (i.e., NVIDIA GeForce RTX 3090 GPU) while being constrained to maintain the FLOPs, the inference time (runtime), and the PSNR result of RLFN.
It's important to highlight that to determine the final ranking and challenge winners, greater weight will be given to teams or participants who demonstrate improvements in more than one aspect (runtime, FLOPs, and parameters) over the provided reference solution.
To ensure fairness in the evaluation process, it is imperative to adhere to the following guidelines:
-
Avoid Training with Specific Image Sets: Refrain from training your model using the validation LR images, validation HR images, or testing LR images. The test datasets will not be disclosed, making PSNR performance on the test datasets a crucial factor in the final evaluation.
-
PSNR Threshold and Ranking Eligibility: Methods with a PSNR below the specified threshold (i.e., 26.90 dB on DIV2K_LSDIR_valid and, 26.99 dB on DIV2K_LSDIR_test) will not be considered for the subsequent ranking process. It is essential to meet the minimum PSNR requirement to be eligible for further evaluation and ranking.
The evaluation environments adopted by us is recorded in the requirements.txt
. After you built your own basic Python setup via either virtual environment or anaconda, please try to keep similar to it via:
pip install -r requirements.txt
or take it as a reference based on your original environments.
After downloaded all the necessary validate dataset (DIV2K_LSDIR_valid_LR and DIV2K_LSDIR_valid_HR), please organize them as follows:
|NTIRE2024_ESR_Challenge/
|--DIV2K_LSDIR_valid_HR/
| |--000001.png
| |--000002.png
| |--...
| |--000100.png
| |--0801.png
| |--0802.png
| |--...
| |--0900.png
|--DIV2K_LSDIR_valid_LR/
| |--000001x4.png
| |--000002x4.png
| |--...
| |--000100x4.png
| |--0801x4.png
| |--0802x4.png
| |--...
| |--0900.png
|--NTIRE2024_ESR/
| |--...
| |--test_demo.py
| |--...
|--results/
|--......
git clone https://github.com/Amazingren/NTIRE2024_ESR.git
- Select the model you would like to test from
run.sh
CUDA_VISIBLE_DEVICES=0 python test_demo.py --data_dir [path to your data dir] --save_dir [path to your save dir] --model_id 0
- Be sure the change the directories
--data_dir
and--save_dir
.
- Be sure the change the directories
- More detailed example-command can be found in
run.sh
for your convenience.
As a reference, we provide the results of RLFN (baseline method) below:
-
Average PSNR on DIV2K_LSDIR_valid: 26.96 dB
-
Average PSNR on DIV2K_LSDIR_test: 27.07 dB
-
Number of parameters: 0.317 M
-
Runtime: 13.54 ms (Average runtime of 16.18 ms on DIV2K_LSDIR_valid data and 10.89 ms on DIV2K_LSDIR_test data)
-
FLOPs on an LR image of size 256Γ256: 19.67 G
Please note that the results reported above are the average of 5 runs, and each run is conducted on the same device (i.e., NVIDIA GeForce RTX 3090 GPU).
- Register your team in the Google Spreadsheet and get your team ID.
- Put your the code of your model in
./models/[Your_Team_ID]_[Your_Model_Name].py
- Please add only one file in the folder
./models
. Please do not add other submodules. - Please zero pad [Your_Team_ID] into two digits: e.g. 00, 01, 02
- Please add only one file in the folder
- Put the pretrained model in
./model_zoo/[Your_Team_ID]_[Your_Model_Name].[pth or pt or ckpt]
- Please zero pad [Your_Team_ID] into two digits: e.g. 00, 01, 02
- Add your model to the model loader
./test_demo/select_model
as follows:elif model_id == [Your_Team_ID]: # define your model and load the checkpoint
- Note: Please set the correct data_range, either 255.0 or 1.0
- Send us the command to download your code, e.g,
git clone [Your repository link]
- We will do the following steps to add your code and model checkpoint to the repository.
from utils.model_summary import get_model_flops, get_model_activation
from models.team00_RLFN import RLFN_Prune
from fvcore.nn import FlopCountAnalysis
model = RLFN_Prune()
input_dim = (3, 256, 256) # set the input dimension
activations, num_conv = get_model_activation(model, input_dim)
activations = activations / 10 ** 6
print("{:>16s} : {:<.4f} [M]".format("#Activations", activations))
print("{:>16s} : {:<d}".format("#Conv2d", num_conv))
# The FLOPs calculation in previous NTIRE_ESR Challenge
# flops = get_model_flops(model, input_dim, False)
# flops = flops / 10 ** 9
# print("{:>16s} : {:<.4f} [G]".format("FLOPs", flops))
# fvcore is used in NTIRE2024_ESR for FLOPs calculation
input_fake = torch.rand(1, 3, 256, 256).to(device)
flops = FlopCountAnalysis(model, input_fake).total()
flops = flops/10**9
print("{:>16s} : {:<.4f} [G]".format("FLOPs", flops))
num_parameters = sum(map(lambda x: x.numel(), model.parameters()))
num_parameters = num_parameters / 10 ** 6
print("{:>16s} : {:<.4f} [M]".format("#Params", num_parameters))
After the organizers receive all the submitted codes/checkpoints/results, four steps are adopted for the evaluation:
- Step1: The organizers will execute each model five times to reevaluate all submitted methods on the same device, specifically the NVIDIA GeForce RTX 3090. The average results of these five runs will be documented for each metric.
- Step2: To ensure PSNR consistency with the baseline method RLFN, PSNR checks will be conducted for all submitted methods. Any method with a PSNR below 26.90 dB on the DIV2K_LSDIR_valid dataset or less than 26.99 on the DIV2K_LSDIR_test datasets will be excluded from the comparison list for the remaining rankings.
- Step3: For the rest, the Score_Runtime, Score_FLOPs, and the Score_Params will be calculated as follows:
Score_Runtime = exp(2*Runtime / Runtime_RLFN)
Score_FLOPs = exp(2*FLOPs / FLOPs_RLFN)
Score_Params = exp(2*Params / Params_RLFN)
- Step4: The final comparison score will be calculated as follows:
Score_Final = 0.7*Score_Runtime + 0.15*Score_FLOPs + 0.15*Score_Params
Let's take the baseline as an example, given the results (i.e., average Runtime_RLFN = 13.54 ms, FLOPs_RLFN = 19.67 G, and Params_RLFN = 0.317 M) of RLFN, we have:
Score_Runtime = 7.3891
Score_FLOPs = 7.3891
Score_Params = 7.3891
Score_Final = 7.3891
βThe ranking for each sub-track will be generated based on the corresponding Score (i.e., Score_Runtime, Score_FLOPs, and Score_Params), while for the main track, the ranking will be determined by the Score_Final.
- Yawei Li (yawei.li@vision.ee.ethz.ch)
- Bin Ren (bin.ren@unitn.it)
- Nancy Mehta (nancy.mehta@uni-wuerzburg.de)
- Radu Timofte (Radu.Timofte@uni-wuerzburg.de)
If you have any question, feel free to reach out the contact persons and direct managers of the NTIRE challenge.
This code repository is release under MIT License.