Adversarial Robustness
Edit model card

MixedNUTS: Training-Free Accuracy-Robustness Balance via Nonlinearly Mixed Classifiers

This is the official model repository of the preprint paper
MixedNUTS: Training-Free Accuracy-Robustness Balance via Nonlinearly Mixed Classifiers
by Yatong Bai, Mo Zhou, Vishal M. Patel, and Somayeh Sojoudi in Transactions on Machine Learning Research.

MixedNUTS Results

TL;DR: MixedNUTS balances clean data classification accuracy and adversarial robustness without additional training via a mixed classifier with nonlinear base model logit transformations.

Model Checkpoints

MixedNUTS is a training-free method that has no additional neural network components other than its base classifiers.

All robust base classifiers used in the main results of our paper are available on RobustBench and can be downloaded automatically via the RobustBench API.

Here, we provide the download links to the standard base classifiers used in the main results.

Dataset Link
CIFAR-10 Download
CIFAR-100 Download
ImageNet Download

For code and detailed usage, please refer to our GitHub repository.

Citing our work (BibTeX)

@article{MixedNUTS,
  title={MixedNUTS: Training-Free Accuracy-Robustness Balance via Nonlinearly Mixed Classifiers},
  author={Bai, Yatong and Zhou, Mo and Patel, Vishal M. and Sojoudi, Somayeh},
  journal={Transactions on Machine Learning Research},
  year={2024}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Datasets used to train Bai-YT/MixedNUTS