Data-Centric Robust Training for Defending Against Transfer-Based Adversarial Attacks

Research output: Contribution to journalArticlepeer-review

Abstract

Transfer-based adversarial attacks pose a severe threat to real-world deep learning systems since they do not require access to target models. Adversarial training (AT), which is recognized as the most effective defense against white-box attacks, also ensures high robustness against (black-box) transfer-based attacks. However, AT suffers from significant computational overhead because it repeatedly generates adversarial examples (AEs) throughout the entire training process. In this paper, we demonstrate that such repeated generation is unnecessary to achieve robustness against transfer-based attacks. Instead, pre-generating AEs all at once before training is sufficient, as proposed in our new defense paradigm called Data-Centric Robust Training (DCRT). DCRT employs clean data augmentation and adversarial data augmentation techniques to enhance the dataset before training. Our experimental results show that DCRT outperforms widely-used AT techniques (e.g., PGD-AT, TRADES, EAT, and FAT) in terms of transfer-based black-box robustness and even surpasses the top-1 defense on RobustBench when combined with common model-centric techniques. We also highlight additional benefits of DCRT, such as improved training efficiency and class-wise fairness. Our code will be available on GitHub.

Original languageEnglish
Pages (from-to)10275-10287
Number of pages13
JournalIEEE Transactions on Information Forensics and Security
Volume20
DOIs
StatePublished - 2025

Keywords

  • Adversarial example
  • black-box defense
  • class-wise fairness
  • data-centric learning
  • transfer-based attack

Fingerprint

Dive into the research topics of 'Data-Centric Robust Training for Defending Against Transfer-Based Adversarial Attacks'. Together they form a unique fingerprint.

Cite this