ndb796 / MultipleInputMixup

Multiple Input Mixup for Out of Distribution Detection

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Towards Reliable AI Model Deployments:
Multiple Input Mixup for Out-of-Distribution Detection (MIM)

arXiv

  • This repository provides PyTorch implementations of our proposed Multiple Input Mixup for Out-of-Distribution Detection (MIM) method, along with other state-of-the-art (SOTA) methods in OOD detection.
  • This work has been accepted to the AAAI 2024 workshop on Deployable AI

Authors

Dongbin Na, Dasol Choi

Abstract

Recent remarkable success in the deep-learning industries has unprecedentedly increased the need for reliable model deployment. For example, the model should alert the user if the produced model outputs might not be reliable. Previous studies have proposed various methods to solve the Out-of-distribution (OOD) detection problem, however, they generally require a burden of resources. In this work, we propose a novel and simple method, Multiple Input Mixup (MIM). Our method can help improve the OOD detection performance with only single epoch fine-tuning. Our method does not require training the model from scratch and can be attached to the classifier simply. Despite its simplicity, our MIM shows competitive performance. Our method can be suitable for various environments because our method only utilizes the In-Distribution (ID) samples to generate the synthesized OOD data. With extensive experiments with CIFAR10 and CIFAR100 benchmarks that have been largely adopted in out-of-distribution detection fields, we have demonstrated our MIM shows comprehensively superior performance compared to the SOTA method. Especially, our method does not need additional computation on the feature vectors compared to the previous studies. All source codes, models, and datasets will be publicly available if the paper is accepted.

OOD Datasets

We utilize a range of Out-of-Distribution (OOD) datasets for comprehensive testing, including Textures, LSUN-crop, LSUN-resize, Tiny-ImageNet-crop, Tiny-ImageNet-resize, SVHN, and Places365. These datasets encompass a diverse array of images, from natural textures and cropped scene images to digit images and various scene categories, providing a robust platform for evaluating OOD detection performance.

Metrics

  • AUROC (Area Under the Receiver Operating Characteristic): This measures the model's ability to distinguish between in-distribution (ID) and OOD samples, offering an overall assessment of OOD detection performance.
  • AUPR (Area Under the Precision-Recall Curve): This evaluates the precision-recall trade-off in OOD detection, indicating how well the model balances precision and recall in identifying OOD samples.

Performance

All experiments are conducted by pytorch-ood, the OOD detection benchmark framework.

  • Comparison with state-of-the-art methods and our MIM using a WideResNet-40-2 classifier.

  • Comparison of AUROC between Outlier Exposure (OE) and our MIM using a WideResNet-40-2 classifier.

Citation

@misc{choi2023reliable,
      title={Towards Reliable AI Model Deployments: Multiple Input Mixup for Out-of-Distribution Detection}, 
      author={Dasol Choi and Dongbin Na},
      year={2023},
      eprint={2312.15514},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

About

Multiple Input Mixup for Out of Distribution Detection


Languages

Language:Jupyter Notebook 100.0%