Skip to content

Latest commit

 

History

History
136 lines (101 loc) · 6.08 KB

README.md

File metadata and controls

136 lines (101 loc) · 6.08 KB

Robustness Tokens

Training schema for Robustness Tokens

Official implementation of the paper "Robustness Tokens: Towards Adversarial Robustness of Transformers" (ECCV 2024).

Set-up

We build on top of the official DinoV2 implementation. In particular, our code:

  • Allows training robustness tokens
  • Implements the PGD adversarial attack
  • Converts obtained robustness tokens into a valid checkpoint for evaluation with the DinoV2 codebase

We use the official DinoV2 codebase for evaluation of robustness tokens on downstream tasks:

  • Classification
  • Semantic segmentation

For each task, we evaluate the performances of the original model and of the model with robustness tokens, both on clean and adversarial examples.

Conda environment

Create the conda environment using the provided environment.yml file:

conda env create -f environment.yml
conda activate rtokens

Alternatively, you can install the required packages using pip:

conda create -n rtokens python=3.11
conda activate rtokens

# Install PyTorch with the desired version, e.g.:
conda install pytorch torchvision torchaudio cudatoolkit=11.1 -c pytorch -c nvidia

# Install the remaining requirements
pip install -r requirements.txt

Python-Dotenv

Create a .env file in the root directory of the project with a variable IMAGENET_DIR pointing to the ImageNet dataset directory.

Datasets

We use the ImageNet dataset for training robustness tokens and to evaluate linear classification capabilities. For segmentation, we use the ADE20k 2016 dataset. You can get the test set from here.

For the ImageNet dataset, pre-process the dataset as described in the DinoV2 codebase:

from dinov2.data.datasets import ImageNet

for split in ImageNet.Split:
    dataset = ImageNet(split=split, root=your_imagenet_dir, extra=your_extra_dir)
    dataset.dump_extra()

you will need to use the labels.txt file to map the ImageNet labels to the class indices used in the codebase.

Training robustness tokens

To train robustness tokens for a pre-trained DinoV2 model, run:

PYTHONPATH=src/ python src/train.py --config $path_to_file

Examples of training configurations can be found in configs/train/.

Evaluating robustness

You can evaluate the robustness of features extracted by models with or without robustness tokens to adversarial attacks.

PYTHONPATH=src/ python src/robustness/feat.py --config $path_to_file

Examples of training configurations can be found in configs/robustness/features/.

The same can be done to evaluate robustness in the case of classification and segmentation with the scripts src/robustness/class.py and src/robustness/seg.py, respectively.

Evaluating downstream performance

We verify that downstream performance is not affected by the addition of robustness tokens.

For evaluation, we convert our checkpoints into a valid checkpoint for the DinoV2 codebase.

PYTHONPATH=src/ python src/eval/convert.py --checkpoint $path_to_file --output $path_to_file

The robustness tokens are converted into DinoV2 register tokens and appended before patch tokens. Please refer to the DinoV2 and MMSegmentation codebases for more details on how to evaluate the downstream performance.

Pre-trained tokens

We publicly release the (10) obtained robustness tokens. These tokens were obtained from the pre-trained backbone models without any change to their weights. You can download the pre-trained robustness tokens from the following links

Model Weights
DiNOv2 small tokens
DiNOv2 base tokens
DiNOv2 large tokens
DiNOv2 huge tokens
DiNOv2 small + reg tokens
DiNOv2 base + reg tokens
DiNOv2 large + reg tokens
DiNOv2 huge + reg tokens
Deit3 base tokens
Deit3 large tokens
Deit3 huge tokens
Openclip base tokens
Openclip large tokens
Openclip huge tokens

Citation

@InProceedings{10.1007/978-3-031-73202-7_7,
author="Pulfer, Brian
and Belousov, Yury
and Voloshynovskiy, Slava",
editor="Leonardis, Ale{\v{s}}
and Ricci, Elisa
and Roth, Stefan
and Russakovsky, Olga
and Sattler, Torsten
and Varol, G{\"u}l",
title="Robustness Tokens: Towards Adversarial Robustness of Transformers",
booktitle="Computer Vision -- ECCV 2024",
year="2025",
publisher="Springer Nature Switzerland",
address="Cham",
pages="110--127",
abstract="Recently, large pre-trained foundation models have become widely adopted by machine learning practitioners for a multitude of tasks. Given that such models are publicly available, relying on their use as backbone models for downstream tasks might result in high vulnerability to adversarial attacks crafted with the same public model. In this work, we propose Robustness Tokens, a novel approach specific to the transformer architecture that fine-tunes a few additional private tokens with low computational requirements instead of tuning model parameters as done in traditional adversarial training. We show that Robustness Tokens make Vision Transformer models significantly more robust to white-box adversarial attacks while also retaining the original downstream performances.",
isbn="978-3-031-73202-7"
}

License

The code is released with the MIT license.