Abstract
With the rapid deployment of deep neural networks (DNNs) across critical application domains, backdoor attacks have emerged as a significant security threat. However, most existing methods rely on access to the target model's original training data and require explicit triggers to activate malicious behavior, which limits their practicality and compromises stealth.This paper proposes a novel trigger-free and data-free backdoor attack framework that enhances both the practicality and concealment of attacks. Our approach leverages a fine-tuning strategy to embed the semantics of malicious data into the feature space of an attacker-specified target class, enabling adversarial samples to be misclassified consistently without any visible trigger.To preserve the model's performance on clean inputs, we incorporate a knowledge distillation mechanism in place of the original training data and design an elastic weight consolidation-based parameter importance estimation method to guide the injection process.Extensive experiments conducted on three real-world benchmark datasets demonstrate the effectiveness, stealthiness, and real-world feasibility of the proposed method. Additionally, we explore the potential of auxiliary data and model inversion techniques in further enhancing attack success.
| Translated title of the contribution | Trigger-free and data-free backdoor attacks on deep neural networks |
|---|---|
| Original language | Chinese (Traditional) |
| Pages (from-to) | 2798-2816 |
| Number of pages | 19 |
| Journal | Scientia Sinica Informationis |
| Volume | 55 |
| Issue number | 11 |
| DOIs | |
| State | Published - 1 Nov 2025 |