-
Python 3.6.12
-
Pytorch 1.2.0+
-
torchvision 0.4.0+
- We released our code for joint training with depth and appearance, which is also our best performance model.
git clone https://github.com/OLobbCode/NoiseLF.git
cd NoiseLF-code/
Download the following datasets and unzip them.
- DUT-LF dataset,fetch code is ‘vecy’.
- HFUT dataset.
- LFSD dataset.
- The .txt file link for testing and training is here, code is 'joaa'.
- Set the
c.DATA.TRAIN.ROOT
andc.DATA.TRAIN.LIST
path inconfig.py
correctly. - We demo using VGG-19 as network backbone and train with a initial lr of 1e-5 for 30 epoches.
- After training the result model will be stored under
snapshot/exp_noiself
folder.
Note:only support c.SOLVER.BATCH_SIZE=1
For single dataset testing: you should set c.PHASE='test'
in config.py, and set c.DATA.TEST.ROOT
, c.DATA.TEST.LIST
as yours.
python demo.py
For evaluate :
python evaluate.py
All results saliency maps will be stored under 'Test/Out/exp_noiself_30/'
folders in .png formats.
Thanks to MOLF.