Skip to content

This is an official implementation for "Block Selection Method for Using Feature Norm in Out-of-distribution Detection".

License

Notifications You must be signed in to change notification settings

gist-ailab/block-selection-for-OOD-detection

Repository files navigation

Block Selection Method for Using Feature Norm in Out-of-distribution Detection (FeatureNorm)

Official Implementation of the "Block Selection Method for Using Feature Norm in Out-of-distribution Detection (CVPR 2023)" by Yeonguk Yu, Sungho Shin, Seongju Lee, Changhyun Jun, and Kyoobin Lee.

In this study, we propose a block selection method that utilizes the L2-norm of the activation map to detect out-of-distribution (OOD) samples. We were inspired to develop this method because we observed that the last block of neural networks can sometimes be overconfident, which can lead to deterioration in OOD detection performance. To select the block for OOD detection, we use NormRatio, which is a ratio of FeatureNorm for ID and pseudo-OOD. This ratio measures the OOD detection performance of each block. Specifically, we create Jigsaw puzzle images from ID training samples to simulate pseudo-OOD and calculate NormRatio. We choose the block with the largest value of NormRatio, which provides the biggest difference between FeatureNorm of ID and FeatureNorm of pseudo-OOD.

concept.png

[ArXiv] [Presentation]


Currently, this code supports for the CIFAR10 benchmark with ResNet18, WideResNet, and VGG architectures

Updates & TODO Lists

  • FeatureNorm has been released
  • pretrained checkpoints
  • Environment settings and Train & Evaluation Readme
  • Code for all architectures

Getting Started

Environment Setup

This code is tested under Linux 20.04 and Python 3.7.7 environment, and the code requires following packages to be installed:

  • Pytorch: Tested under 1.11.0 version of Pytorch-GPU.
  • torchvision: which will be installed along Pytorch. Tested under 0.6.0 version.
  • timm: Tested under 0.4.12 version.
  • scipy: Tested under 1.4.1 version.
  • scikit-learn: Tested under 0.22.1 version.

Dataset Preparation

Some public datasets are required to be downloaded for running evaluation. Required dataset can be downloaded in following links as in https://github.com/wetliu/energy_ood:

Config file need to be changed for your path to download. For example,

# conf/cifar10.json
{
    "epoch" : "100",
    "id_dataset" : "./cifar10",   # Your path to Cifar10
    "batch_size" : 128,
    "save_path" : "./cifar10/",   # Your path to checkpoint
    "num_classes" : 10
}

Also, you need to change the path of the OOD dataset in "eval.py" to conduct a OOD benchmark.

OOD_results(preds_in, model, get_svhn('/SSDe/yyg/data/svhn', batch_size), device, args.method+'-SVHN', f)
OOD_results(preds_in, model, get_ood('/SSDe/yyg/data/ood-set/textures/images'), device, args.method+'-TEXTURES', f)
OOD_results(preds_in, model, get_ood('/SSDe/yyg/data/ood-set/LSUN'), device, args.method+'-LSUN-crop', f)
OOD_results(preds_in, model, get_ood('/SSDe/yyg/data/ood-set/LSUN_resize'), device, args.method+'-LSUN-resize', f)
OOD_results(preds_in, model, get_ood('/SSDe/yyg/data/ood-set/iSUN'), device, args.method+'-iSUN', f)
OOD_results(preds_in, model, get_places('/SSDd/yyg/data/places256'), device, args.method+'-Places365', f)

How to Run

To train a model by our setting (i.e., ours) with ResNet18 architecture

python train.py -d 'data_name' -g 'gpu-num' -n resnet18 -s 'save_name'

for example,

python train_baseline.py -d cifar10 -g 0 -n resnet18 -s baseline

To detect OOD using norm of the penultimate block

python eval.py -n resnet18 -d 'data_name' -g 'gpu_num' -s 'save_name' -m featurenorm

for example,

python eval.py -n resnet18 -d cifar10 -g 0 -s baseline 

Also, you can try MSP method

python eval.py -n resnet18 -d 'data_name' -g 'gpu_num' -s 'save_name' -m msp

To calculate NormRatio of each block using generated Jigsaw puzzle images

python normratio.py -n resnet18 -d 'data_name' -g 'gpu_num' -s 'save_name' 

for example,

python normratio.py -n resnet18 -d cifar10 -g 0 -s baseline 

License

The source code of this repository is released only for academic use. See the license file for details.

Acknowledgement

This work was partially supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2022-0-00951, Development of Uncertainty-Aware Agents Learning by Asking Questions) and by ICT R&D program of MSIT/IITP[2020-0-00857, Development of Cloud Robot Intelligence Augmentation, Sharing and Framework Technology to Integrate and Enhance the Intelligence of Multiple Robots].

Citation

@InProceedings{Yu_2023_CVPR,
    author    = {Yu, Yeonguk and Shin, Sungho and Lee, Seongju and Jun, Changhyun and Lee, Kyoobin},
    title     = {Block Selection Method for Using Feature Norm in Out-of-Distribution Detection},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2023},
    pages     = {15701-15711}
}

About

This is an official implementation for "Block Selection Method for Using Feature Norm in Out-of-distribution Detection".

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages