Supereeeee / MSDAN

MSDAN: A lightweight multi-scale distillation attention network for image super-resolution

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

MSDAN: A lightweight multi-scale distillation attention network for image super-resolution

Environment in our experiments

[python 3.8]

[Ubuntu 20.04]

BasicSR 1.4.2

PyTorch 1.13.0, Torchvision 0.14.0, Cuda 11.7

Installation

git clone https://github.com/Supereeeee/MSDAN.git
pip install -r requirements.txt
python setup.py develop

How To Test

· Refer to ./options/test for the configuration file of the model to be tested and prepare the testing data.

· The pre-trained models have been palced in ./experiments/pretrained_models/

· Then run the follwing codes (taking MSDAN_x4.pth as an example):

python basicsr/test.py -opt options/test/test_MSDAN_x4.yml

The testing results will be saved in the ./results folder.

How To Train

· Refer to ./options/train for the configuration file of the model to train.

· Preparation of training data can refer to this page. All datasets can be downloaded at the official website.

· Note that the default training dataset is based on lmdb, refer to docs in BasicSR to learn how to generate the training datasets.

· The training command is like

python basicsr/train.py -opt options/train/train_MSDAN_x4.yml

For more training commands and details, please check the docs in BasicSR

Model Complexity

· The network structure of MSDAN is palced at ./basicsr/archs/MSDAN_arch.py

· We adopt thop tool to calculate model complexity, see ./basicsr/archs/model_complexity.py

Inference time

· We test the inference time on multiple benchmark datasets on a 140W fully powered 3060 laptop.

· You can run ./inference/inference_MSDAN.py on your decive.

Acknowledgement

This code is based on BasicSR toolbox. Thanks for the awesome work.

Contact

If you have any question, please email 1051823707@qq.com.

About

MSDAN: A lightweight multi-scale distillation attention network for image super-resolution


Languages

Language:Python 84.6%Language:Cuda 8.7%Language:C++ 5.8%Language:MATLAB 0.8%Language:Shell 0.1%