tmlr-group / SIGUA

[ICML 2020] "SIGUA: Forgetting May Make Learning with Noisy Labels More Robust"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

SIGUA

ICML'20: SIGUA: Forgetting May Make Learning with Noisy Labels More Robust

======== This is the code for the paper: SIGUA: Forgetting May Make Learning with Noisy Labels More Robust

Presented at ICML'20 If you find this code useful in your research then please cite

@InProceedings{pmlr-v119-han20c,
  title = 	 {{SIGUA}: Forgetting May Make Learning with Noisy Labels More Robust},
  author =       {Han, Bo and Niu, Gang and Yu, Xingrui and Yao, Quanming and Xu, Miao and Tsang, Ivor and Sugiyama, Masashi},
  booktitle = 	 {International Conference on Machine Learning},
  pages = 	 {4006--4016},
  year = 	 {2020}
}

Setup

Istall Miniconda3, and then

conda create -f environment.yml 
conda activate pytorch1.0.0_py36

Running SIGUA on benchmark datasets (MNIST, CIFAR-10)

sh scripts/mnist_sigua_sl.sh
sh scripts/mnist_sigua_bc.sh
sh scripts/cifar10_sigua_sl.sh
sh scripts/cifar10_sigua_bc.sh

The other reproducible version

Please check the other reproducible code of SIGUA: https://github.com/yeachan-kr/pytorch-sigua

About

[ICML 2020] "SIGUA: Forgetting May Make Learning with Noisy Labels More Robust"


Languages

Language:Python 94.1%Language:Shell 5.9%