RedTachyon commited on
Commit
4c86d90
1 Parent(s): d5fc6a4

Upload folder using huggingface_hub

Browse files
XuOE99cmST/11_image_0.png ADDED

Git LFS Details

  • SHA256: ac2e912d8b7ed3683f8b7faa8710b5f46adcd007ff0d96704d7a3974fbf51920
  • Pointer size: 130 Bytes
  • Size of remote file: 39.1 kB
XuOE99cmST/19_image_0.png ADDED

Git LFS Details

  • SHA256: 69fe28873bcf2d99c2e558e30bb111792ebf3eead66a6384db6fe55d0925fdf2
  • Pointer size: 130 Bytes
  • Size of remote file: 93.4 kB
XuOE99cmST/1_image_0.png ADDED

Git LFS Details

  • SHA256: 1bf61a55b3f761a5ed5a0ec94e4868660be33bd39d5c5b1aa09c215ed12a70f3
  • Pointer size: 130 Bytes
  • Size of remote file: 78.5 kB
XuOE99cmST/20_image_0.png ADDED

Git LFS Details

  • SHA256: 82cbd76133ea7fe0f4912e5c8410caaaf9888e550cb8cdc2d2ee180fe72680bd
  • Pointer size: 130 Bytes
  • Size of remote file: 64 kB
XuOE99cmST/20_image_1.png ADDED

Git LFS Details

  • SHA256: 3b95837b3b28b8814f94e1173f9c5c027ec8d57a1df4c335d8180921e78df0a2
  • Pointer size: 131 Bytes
  • Size of remote file: 156 kB
XuOE99cmST/22_image_0.png ADDED

Git LFS Details

  • SHA256: 11e5eb2a202cf7cd692d900ebce4686a03046ad631578577de7a232b8c037328
  • Pointer size: 130 Bytes
  • Size of remote file: 26.2 kB
XuOE99cmST/4_image_0.png ADDED

Git LFS Details

  • SHA256: 38c2f75d98721230d1962c59d196ca4afe958ed19c0f4b265747176f674c2d1e
  • Pointer size: 130 Bytes
  • Size of remote file: 71.9 kB
XuOE99cmST/7_image_0.png ADDED

Git LFS Details

  • SHA256: 9b76bc6ac78709b86cfe39f53e3a101dd1dde325340c41911cab2a1db4dbe026
  • Pointer size: 130 Bytes
  • Size of remote file: 51.9 kB
XuOE99cmST/XuOE99cmST.md ADDED
@@ -0,0 +1,683 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Turning A Curse Into A Blessing: Enabling In-Distributiondata-Free Backdoor Removal Via Stabilized Model Inversion
2
+
3
+ Si Chen *chensi@vt.edu* Virginia Tech Yi Zeng yizeng@vt.edu Virginia Tech Jiachen T. Wang *tianhaowang@princeton.edu* Princeton University Won Park wonpark@umich.edu University of Michigan Xun Chen xun.chen@samsung.com Samsung Research America Lingjuan Lyu lingjuan.lv@sony.com Sony AI
4
+ Zhuoqing Mao *zmao@umich.edu* University of Michigan Ruoxi Jia *ruoxijia@vt.edu* Virginia Tech Reviewed on OpenReview: *https: // openreview. net/ forum? id= P880C39xAvM*
5
+
6
+ ## Abstract
7
+
8
+ The effectiveness of many existing techniques for removing backdoors from machine learning models relies on access to clean in-distribution data. However, given that these models are often trained on proprietary datasets, it may not be practical to assume that in-distribution samples will always be available. On the other hand, model inversion techniques, which are typically viewed as privacy threats, can reconstruct realistic training samples from a given model, potentially eliminating the need for in-distribution data. To date, the only prior attempt to integrate backdoor removal and model inversion involves a simple combination that produced very limited results. This work represents a first step toward a more thorough understanding of how model inversion techniques could be leveraged for effective backdoor removal. Specifically, we seek to answer several key questions: What properties must reconstructed samples possess to enable successful defense? Is perceptual similarity to clean samples enough, or are additional characteristics necessary? Is it possible for reconstructed samples to contain backdoor triggers?
9
+
10
+ We demonstrate that relying solely on perceptual similarity is insufficient for effective defenses. The stability of model predictions in response to input and parameter perturbations also plays a critical role. To address this, we propose a new bi-level optimization-based framework for model inversion that promotes stability in addition to visual quality. Interestingly, we also find that reconstructed samples from a pre-trained generator's latent space do not contain backdoors, even when signals from a backdoored model are utilized for reconstruction. We provide a theoretical analysis to explain this observation. Our evaluation shows that our stabilized model inversion technique achieves state-of-the-art backdoor removal performance without requiring access to clean in-distribution data, but solely by utilizing publicly available data from similar distributions. Furthermore, its performance is on par with or even better than using the same amount of clean samples.
11
+
12
+ ## 1 Introduction
13
+
14
+ ![1_Image_0.Png](1_Image_0.Png)
15
+
16
+ Figure 1: Loss landscape of clean data and synthetic data by a prior MI technique (Zhang et al., 2020).
17
+
18
+ x, y axis are two random change directions in data space (figure (a)) and model parameter space (figure
19
+ (b)); (0,0) represents the original data and model in figure (a) and (b). The synthesized data generated by prior MI is demonstrated to exhibit instability in the face of small data perturbations or changes in model parameters, unlike the stable nature observed in clean in-distribution data.
20
+ Deep neural networks have been shown to be vulnerable to backdoor attacks, in which attackers poison training data such that the trained model misclassifies any test input patched with some trigger pattern as an attacker-specified target class (Saha et al., 2020; Li et al., 2020a; Zeng et al., 2021). These attacks create a major hurdle to deploying deep networks in safety-critical applications.
21
+
22
+ Various techniques (Wang et al., 2019; Guo et al., 2019; Liu et al., 2018a) have been developed to remove the effects of backdoor attacks from a target poisoned model and turn it into a well-behaved model that does not react to the presence of a trigger. Most of the backdoor removal techniques rely on access to a set of clean samples drawn from the distribution that the poisoned model is trained on. These clean data are needed for synthesizing potential triggers and further fine-tuning the model to let the model unlearn the triggers.
23
+
24
+ However, accessing clean in-distribution samples might not always be feasible. Particularly, machine learning models are often trained on proprietary datasets which are not publicly released. For instance, various MLas-a-service platforms (e.g., Tensorflow Hub) offer trained models that users can download but often do not publish the corresponding training data. There have been a few attempts to lift the requirement on clean in-distribution data, yet suffering unstable performance across different triggers. CLP (Zheng et al., 2022) assumes that the backdoor-related neurons in a poisoned model have a large Lipschitz constant and prunes these neurons to repair the model. However, this assumption does not hold when the trigger induces a large change in the model input. Another line of ideas (Chen et al., 2019) is to reconstruct data from the target model and then use them as a proxy for clean in-distribution data needed by existing data-reliant defenses. This line has the *unique benefit* that it can take advantage of advances of those data-reliant defenses which have already demonstrated remarkable efficacy on various triggers.
25
+
26
+ In fact, the problem of reconstructing samples from a trained model has been extensively studied in the data privacy literature, known as *model inversion (MI)*. While it is natural to utilize MI to generate data for data-reliant backdoor defense, to the best of our knowledge, there is only one work doing so: Chen et al. (2019) utilizes the simplest MI technique, which synthesizes images for a given class by optimizing the likelihood of the model for predicting that class, and further performs backdoor removal. However, this MI technique is known to fall short in reconstructing high-dimensional input (e.g., RGB images) from a deep neural network, resulting in a noise-like pattern that lacks semantic information about a class.
27
+
28
+ We observe that feeding such low-quality reconstructed samples into even the most advanced (data-reliant)
29
+ backdoor removal technique leads to poor performance. Recently, a series of MI techniques have significantly improved the visual quality for high-dimensional data (Zhang et al., 2020; Chen et al., 2021; Wang et al.,
30
+ 2021; An et al., 2022; Struppek et al., 2022) by performing data synthesis in the latent space of a pre-trained neural network generator. Notably, this generator can be trained using publicly available data from similar distributions. These advanced techniques can often produce synthetic samples that largely retain the classspecific semantics and look perceptually similar to the original training data. However, many important questions remain unclear: *Is perceptual similarity to clean samples enough to enable successful defense, or* are additional characteristics necessary? Is it possible for reconstructed samples to contain backdoor triggers?
31
+
32
+ Intriguingly, despite the perceptual similarity between the samples synthesized by these advanced MI techniques and the original training data, we find that there is a significant gap in their resulting backdoor removal performance. Particularly, we find two factors that contribute to performance degradation (Figure 1). Firstly, we show that the model predictions at the synthesized samples are unstable to small input perturbations, which misleads downstream backdoor removal techniques to remove these perturbations instead of underlying backdoor triggers. Moreover, unlike clean samples for which the prediction loss of the model converges and thus is stable to local changes on the model parameters, the prediction loss at the synthesized samples is sensitive to small parameter changes; therefore, using them to fine-tune a poisoned model result in degradation in model accuracy. Based on these observations, we introduce an algorithmic framework for data reconstruction based on bi-level optimization, which promotes not only perceptual quality but the stability to perturbations in data and parameter space.
33
+
34
+ Moreover, the existing work Chen et al. (2019) that utilizes model inversion for backdoor removal has overlooked a critical question of whether reconstructed samples from a backdoored model could contain backdoors. Note that if the synthesized samples for the target class contain triggers, then the existing backdoor removal techniques would be nullified. Empirically, we find that as long as the pre-trained generator leveraged by MI is learned from clean data, the reconstructed samples from a poisoned model do not contain triggers. For a commonly used generator in MI literature—a generative adversarial network (GAN), we prove that backdoors are not in the range of the generator by analyzing the GAN's equilibrium.
35
+
36
+ We summarize our contributions as follows:
37
+ - We are the first to investigate the connection between MI and backdoor removal. We go beyond perceptual quality and reveal the dependence of defense performance on the stability of the inverted samples to input and parameter perturbations; and provide a theoretical understanding of why pre-trained generator-based MI does not generate backdoor-contaminated samples.
38
+
39
+ - We propose a novel bilevel optimization based data reconstruction approach for in-distribution data free defense (FreD), which maximizes the stability to input perturbation while encouraging perceptual similarity and the stability to model perturbation.
40
+
41
+ - On a range of datasets and models, employing the synthetic samples produced by FreD can lead to state-of-the-art data-free backdoor defense performance, which is comparable to or sometimes even better than using the same amount of clean data.
42
+
43
+ - FreD can be extended to match clean data's features when there is limited access to clean indistribution samples. For instance, we show that combining just one clean in-distribution point per class with FreD can lead to a better defense performance than directly supplying 20 clean points.
44
+
45
+ ## 2 Preliminaries
46
+
47
+ Attacker Model. Assume that an attacker performs a backdoor attack against a clean training set D
48
+ drawn from the distribution D. The attacker injects a set of poisoned samples into D to form a poisoned dataset Dpoi. We will refer to the model trained on the poisoned dataset as a poisoned model, denoted by fθpoi . The goal of the attacker is to poison the training set D such that for any clean test input x, adding a pre-defined trigger pattern δ to x will change the output of the trained classifier fθpoi to be an attackerdesired target class ytar. A standard technique to poison the dataset is to inject backdoored samples that are labeled as the target class and inject the trigger into their features. The model trained on such a poisoned dataset will learn the association between the trigger and the target class, thereby outputting the target class whenever a test input contains the trigger.
49
+
50
+ Backdoor Removal. We consider that the defender is given the poisoned model fθpoi . The goal of the defender is to remove the effects of backdoor triggers from fθpoi and obtain a new model fθ
51
+ ∗ that is robust to backdoor triggers, i.e., fθ
52
+ ∗ (x + δ) = fθ
53
+ ∗ (x). Many past backdoor removal techniques (including the stateof-the-art one) are based on the idea of fine-tuning the poisoned model with a set of samples, which will be referred to as the *base set*; furthermore, past techniques assume that the base set is clean and in-distribution, i.e., each sample there is drawn from D—the distribution generating the clean portion of the data that the poisoned model is trained on. Given the base set B = {(xi, yi)}
54
+ n i=1, Zeng et al. (2022) provides a minimax optimization framework that *unifies* a variety of different backdoor removal techniques (Wang et al., 2019; Chen et al., 2019; Guo et al., 2019):
55
+
56
+ $$\theta^{*}=\arg\min_{\theta}\max_{\delta}\frac{1}{|B|}\sum_{i\in B}L(f_{\theta}(x_{i}+\delta),y_{i}),\tag{1}$$
57
+ $$(1)$$
58
+
59
+ where the inner optimization is aimed at (approximate) *trigger synthesis*, i.e., finding a pattern that causes a high loss for predicting correct labels across all samples in the base set, and the outer optimization performs trigger unlearning, which seeks a model that maintains the correct label prediction yi when the synthesized trigger pattern is patched onto the input xi. Zeng et al. (2022) proposed I-BAU, which achieves state-ofthe-art backdoor removal performance by fine-tuning the poisoned model using mini-batch gradients of the objective in (1). Backdoor removal performance is typically measured by *attack success rate* (ASR), which measures the ratio of the backdoored samples predicted as the target class, and *clean accuracy* (ACC), which measures the ratio of the clean samples predicted as their original class. Despite the promising results, I-BAU
60
+ shows that the defense performance degrades quickly as the size of available clean in-distribution samples shrinks.
61
+
62
+ Connection between Data-Free Backdoor Removal and Model Inversion. How to remove backdoors from a given poisoned model without access to clean, in-distribution samples? A natural idea is that as the poisoned model is trained with some clean data, it may memorize the information about the data and therefore one can potentially reconstruct the clean data from the poisoned model. Reconstructing training data from a trained model is intensively studied in the privacy literature, known as *model inversion*
63
+ (MI) (Fredrikson et al., 2014; 2015). To recover training data from a given model fθ for any class y, the key idea of MI is to find an input that minimizes the prediction loss of y:
64
+
65
+ $$x_{\mathrm{syn}}\in\arg\operatorname*{min}_{x}L(f_{\theta}(x),y).$$
66
+ L(fθ(x), y). (2)
67
+ DeepInspect (Chen et al., 2019) solved (2) with gradient descent for multiple times, each of which uses a randomly selected initial value of x; then, the base set was formed by collecting the converged input xsyn for each initial value and pairing it with the corresponding label y. However, solving (2) over the high-dimensional space without any constraints generates noise-like features that lack semantic information about corresponding labels. Hence, using the samples synthesized by this way to form the base set gives unsatisfactory backdoor removal performance.
68
+
69
+ Recently, GMI (Zhang et al., 2020) proposed to optimize over the latent space of a pre-trained GAN instead:
70
+
71
+ $$x_{\mathrm{syn}}=G(z^{*}),z^{*}\in\arg\operatorname*{min}_{z}\underbrace{L(f_{\theta}(G(z)),y)}_{L_{\mathrm{cl}}(z)}\underbrace{-D(G(z))}_{L_{\mathrm{prior}}(z)},$$
72
+ $$\left(2\right)$$
73
+
74
+ $$(3)$$
75
+
76
+ , (3)
77
+ where G and D represent the generator and the discriminator of the GAN, respectively. Chen et al. (2021); An et al. (2022); Struppek et al. (2022) follow the idea of using GAN and further improve the quality of reconstructed images with different techniques, e.g., knowledge distillation from the target model; latent space disentanglement via a StyleGAN (Karras et al., 2019; 2020), etc. These works show that the samples synthesized by the GAN-based MI technique above can maintain high visual similarity to the original training data of fθ. It is natural to ask: Can we apply these more advanced MI techniques to recover samples from the poisoned model and use them as a substitute for the clean, in-distribution samples needed in backdoor removal? Also, it is critical for the effectiveness of backdoor removal that the target-class samples in the base set do not contain backdoor triggers; otherwise, the trigger unlearning step would reinforce the association between the trigger and the target class, instead of eliminating it. Hence, another critical question is: will MI recover backdoor triggers from the poisoned model? We will answer these questions in the following section. And a more detailed discussion of related works on backdoor removal and MI can be found in Appendix A.
78
+
79
+ ## 3 Using Model Inversion To Form The Base Set
80
+
81
+ MI synthesized data may suffer from visual quality degradation if the GAN is not well trained. Will better visual quality lead to a more successful defense? Besides, in Fig 1, we observe that clean data are more robust to perturbations in both data and model parameters space than GMI synthesized data; could stability also be an important property? It's important to understand key factors that contribute to removal performance and to develop a method accordingly.
82
+
83
+ To tailor backdoor removal, We formalize the optimization goal of model inversion attacks as minzi Lprior(zi)+
84
+ λ1Lcl(zi) + λ2Lmp(zi) + λ4Ldp(zi). Each newly added loss term in this equation corresponds to a significant factor contributing to the backdoor removal performance. We will elaborate on the motivation for each loss term in the subsequent section.
85
+
86
+ ## 3.1 Understanding Factors That Contribute To Backdoor Removal Performance
87
+
88
+ Question 1: Does visual quality affect backdoor removal performance? Existing MI techniques mostly focus on improving visual quality. To study how visual quality impacts backdoor removal performance, we adopt the GAN-based model inversion technique (Zhang et al., 2020) to form a base set (GMI), and follow the idea in An et al. (2022); Struppek et al. (2022) to use a StyleGAN generator to form a base set with higher quality (GMI+). We then use these base sets to perform backdoor removal with I-BAU (Zeng et al.,
89
+ 2022), which is the state-of-the-art backdoor removal technique. Specifically, the poisoned model is trained on a traffic sign dataset (Houben et al., 2013). The backdoor attack in Li et al. (2020a) is considered and the target class is a randomly chosen class. Figure 2 (a) shows the reconstructed samples and the original training data. We find that the samples synthesized by GMI can in general successfully recover the semantics of the clean samples, samples generated by GMI+ even achieve almost perfect visual quality. However, there is still a big gap between clean performance and GMI+ performance: the unlearning performance among different runs is quite different and the average ASR is more than 10 times higher (Figure 2 (b)). This motivates us to find other missing factors contribute to backdoor removal performance.
90
+
91
+ ![4_image_0.png](4_image_0.png)
92
+
93
+ Figure 2: (a) and (b) show the example images and defense performance of the four base sets. (c) is misclassification rate when adding UAP to the four sets respectively. For each set, an optimal UAP is obtained and normalized to 1. We then gradually scale up the four UAPs and test the corresponding misprediction rate. (d) is distribution of samples given its model stability ∥∇θL(fθ(x), y)|θ=θpoi∥1.
94
+
95
+ | Clean | GMI | GMI+ | FreD | |
96
+ |--------------|---------|---------|---------|---------|
97
+ | Trigger Norm | 41.4484 | 32.1581 | 35.6812 | 39.2282 |
98
+
99
+ Table 1: Norm of synthesized triggers using different base sets.
100
+
101
+ Question 2: Does stability to small perturbations affect backdoor performance? Many backdoor removal techniques, including the state-of-the-art one, rely on trigger synthesis. Recall the backdoor unlearning optimization Eq. 1, the inner loop aims to find a trigger causing the highest loss, and the trigger is initialized to be all zeros. However, if synthetic samples are sensitive to small universal adversarial perturbation (UAP), then these UAPs could be synthesized instead of the actual trigger, leading to poor or unstable defense performance. To validate our hypothesis, we use synthesized trigger norm as a metric to evaluate stability of different base sets - a smaller synthesized trigger implies that a smaller perturbation is required to flip the prediction on a batch of samples. As shown in Table 1, clean base set leads to the largest norm. Figure 2(c) also shows that clean data need to be perturbed with a stronger UAP to reach the same misprediction rate as GMI-synthesized data, implying that clean data is more robust to UAPs.
102
+
103
+ | Ldp | Lmp | | | | | |
104
+ |-------|-------|--------|------|-------|--------|------|
105
+ | × 10 | × 100 | × 1000 | × 10 | × 100 | × 1000 | |
106
+ | ACC | 0.75 | 0.82 | 0.72 | 0.94 | 0.96 | 0.97 |
107
+ | ASR | 0.27 | 0.08 | 0.02 | 0.42 | 0.49 | 0.37 |
108
+
109
+ To test whether improving stability to small pertuabtions could help decrease ASR, we introduce an additional loss in Equation 3, named data-perturbation loss: Ldp(z) = −CosSim(fθpoi (G(z)), fθpoi (G(z) + δ)), where CosSim(·, ·) stands for cosine similarity. This loss calculates the change of the model output logits when a synthesized sample G(z) is perturbed by δ. Here δ is randomly sampled from a normal Gaussian distribution.
110
+
111
+ From Table 2 we can see that ASR drops when increasing weight of Ldp, which validates our hypothesis. However, Ldp alone does not have much impact on ACC. How can we further improve the accuracy during unlearning?
112
+
113
+ Table 2: Backdoor unlearning performance with varied weights of data-perturbation (Ldp) and modelperturbation (Lmp) loss.
114
+
115
+ Question 3: Does stability to model parameters affect backdoor performance? The poisoned model is usually trained to be optimal on the original training data, meaning that the gradient with respect to the model parameter on the data should be close to zero: ∥∇θL(fθ(x), y)|θ=θpoi ∥1 ≈ 0, as shown in Figure 1(b). However, Figure 2(d) shows that, while 90% of the clean sample has ∥∇θL(fθ(x), y)|θ=θpoi ∥1 ≤ 0.001, GMI/GMI+ reconstructed samples distribute more diversely, and the gradient norm based on GMI generated samples are relatively higher. We observe that ACC quickly drops when performing unlearning on these samples, which is consistent with catastrophic forgetting phenomenon (Kirkpatrick et al., 2017). Intuitively, we can maintain a high ACC by regularizing the model gradient during inversion.
116
+
117
+ To test this hypothesis, we introduce an additional loss in Equation 3, named model-perturbation loss:
118
+ Lmp(z) = ∥∇θL(fθ(G(z)), y))|θ=θpoi∥1, which measures the stability of the prediction for the synthesized sample G(z) to small changes on the parameters of the poisoned model. Table 2 shows that this loss helps improve ACC during unlearning.
119
+
120
+ ## 3.2 Proposed Approach
121
+
122
+ Given the findings above, we propose FreD, an approach to reconstructing the training data from a trained model. FreD differs from recent MI techniques in that its synthesis goal not only considers synthetic data quality and recovery of class-specific semantics, but also addresses the specific challenges of non-converging prediction and small universal perturbation that hinder successful application to backdoor removal. We introduce following new loss terms critical to enable the downstream task of backdoor removal:
123
+
124
+ - The model-perturbation loss Lmp(z) = ∥∇θL(fθ(G(z)), y))|θ=θpoi∥1.
125
+ - The data-perturbation loss Ldp(*z, δ*) = −CosSim(fθpoi (G(z)), fθpoi (G(z) + δ)).
126
+
127
+ - (Optional) The feature consistency loss Lcon(z) = P(x′,y′)∈Dclean,y′=y
128
+ ∥gθpoi (G(z)) − gθpoi (x
129
+ ′)∥2.
130
+
131
+ gθpoi represents the feature extractor of the poisoned model fθpoi , i.e., the output of the penultimate layer. The loss is only used when we extend our approach to defense setting where a set of clean in-distribution samples Dclean is available. It measures the feature distance between the synthesized sample and the available clean samples.
132
+
133
+ ## Algorithm 1: Algorithm Of Fred.
134
+
135
+ Input : Generator G, target model T, batch size B, clean data x (optional), max iterations N,
136
+ learning rate α1, α2.
137
+
138
+ 1 for each class y ∈ (1, K) do 2 Initialize z: z
139
+ (1) ∼ N(0, I).
140
+
141
+ 3 Initialize δ: δ
142
+ (1) = 0 1×d where d is the dimension of synthesized sample G(z).
143
+
144
+ 4 for *each iteration* i ∈ (1, N) do 5 Temporary Update z: zˆ
145
+ (i) = z
146
+ (i) − α1 1 B
147
+ PB
148
+ b=1 ∇zLdp(z
149
+ (i), δ(i)|*y, x*).
150
+
151
+ 6 Update δ: δ
152
+ (i+1) = δ
153
+ (i) + α2 1 B
154
+ PB
155
+ b=1 ∇δLdp(ˆz i, δ(i)).
156
+
157
+ 7 Update z: z
158
+ (i+1) = z
159
+ (i) − α1 1 B
160
+ PB
161
+ b=1 ∇zL*total*(z
162
+ (i), δ(i+1)|*y, x*).
163
+
164
+ 8 end 9 zy = z
165
+ (N)
166
+ 10 end 11 **return** z1*, . . . , z*K
167
+ Instead of using simple Gaussian noise in Ldp, we propose a bilevel-optimization algorithm 1 to find the most potent universal perturbation for B synthesized samples:
168
+
169
+ $$\delta^{*}=\arg\max_{\delta}\sum_{i=1}^{B}L_{\mathrm{dp}}(z_{i}^{*}(\delta),\delta),$$
170
+ $\mathbf{s}.\mathbf{t}_i$
171
+ s.t. z
172
+
173
+ i (δ) = arg min zi
174
+ {Lprior(zi) + λ1Lcl(zi) + λ2Lmp(zi) + λ3Lcon(zi) + λ4Ldp(zi, δ)}
175
+ ∀i ∈ {1*, . . . , B*}
176
+
177
+ $$\left({4}\right)$$
178
+
179
+ However, it could be computationally expensive as the inner optimization at any δ requires synthesizing a batch of samples. To tackle this challenge, we propose an online approximation algorithm 1 to update z and δ alternatively through a single optimization loop. This type of algorithm is often used in meta learning (Shu et al., 2019; Madaan et al., 2021). The output of 1 provides the optimal perturbation and a batch of synthesized samples robustified against the perturbation.
180
+
181
+ Hyperparameter Tuning: λ1 to λ4 are the weights associated with each loss in Equation 4. In our experiments, we choose λ1 = 1000 following the prior works on MI. We find λ3 = 1000, λ4 = 10 work well across different datasets and models. The best value for λ2 is task-dependant, chosen by grid search that yields the smallest value of Lprior + Lcl + Lmp + Ldp. Table 2 provides a sensitive analysis of λ2 and λ4, which are the weights of our proposed loss terms. Detailed choice of other hyperparameters are provided in Appendix C.
182
+
183
+ ## 4 Mi Does Not Recover Backdoors 4.1 Empirical Study
184
+
185
+ We perform experiments on the GTSRB dataset (Houben et al., 2013) to study whether the synthesized samples would contain backdoors. Specifically, we train poisoned models under various attacks e.g., invisible attack (Li et al., 2020a), smooth (Zeng et al., 2021), trojan square (Liu et al., 2018b), WaNet (Nguyen &
186
+ Tran, 2020b); and apply FreD to reconstruct a set of samples from this model. We generate 100 images for each class. To detect whether or not our synthesized data contain the backdoor trigger, we train a binary trigger detection classifier on a clean GTSRB training set, a GAN-generated set, and their poisoned correspondence. The trained trigger classifier has 100% accuracy on a held-out test set. Applying this trigger classifier to our synthesized data, we get that no images are detected to contain the trigger. However, it is still possible that backdoored data are within the support of the GAN generated images but just not discovered by our synthesis technique as the underlying optimization does not directly lead the synthesized data to recover the backdoored samples. To empirically verify whether the backdoored data are not found by our optimization or they are not in the support of the GAN, after regular optimization as shown in Algorithm 1, we continue to optimize these synthesized data to minimize the mean square error (MSE) between them and their poisoned version. However, even after the optimization, backdoor detection rate is still 0. The intuition here is to encourage GAN to generate trigger pattern by maximizing the pixel-level similarity between GAN
187
+ generated images and poison images. We adopted MSE as our loss function because it's commonly used for pixel-level recovery in the literature. While cosine similarity can also be used to measure the pixel-level similarity, it is not easy to optimize as MSE.
188
+
189
+ Above experiments are all based on the assumption that the auxiliary dataset Daux used for training the GAN is clean. What if some poisoned data is mixed into the GAN's training data? We further poison Daux with different ratio and test the trigger detection rate respectively. As shown in Figure 3, detection rate remains 0 when Daux is poisoned by invisible attack (Li et al., 2020a) using 1% and 2% poison rate, which is typically used for backdoor attacks. Even when the poison rate is increased to an uncommonly large ratio
190
+ (i.e., 50%), the detection rate remains low at 0.02. Bau et al. (2019) shows similar findings of the wholesale omission of GAN on generating some objects. These objects are either complex patterns that include many pixels (e.g., large human figure), or are under-represented in the GAN training distribution. Based on these findings, even if backdoors exist in GAN's training data, it would be hard for them to take effect unless using a substantial poison rate, which becomes unrealistic. Also, recent backdoor data detection methods
191
+ (Ma et al., 2022) have achieved great efficacy. They can serve as pre-processing steps to screen out poisoned samples in GAN's training set. Figure 3: (a) Trigger detection rate when increasing data poison rate of samples from a single class in
192
+
193
+ ![7_image_0.png](7_image_0.png)
194
+
195
+ auxiliary dataset Daux. Trigger detection rate here refers to the percentage of images that were detected containing triggers among all generated images, while the poison rate refers to the percentage of poison images with the target-class images only. (b) Example of images generated from latent code h with different scales of norm.
196
+
197
+ ## 4.2 Theoretical Justification Based On Equilibrium In Gan
198
+
199
+ Here, we provide the theoretical justification of why reconstructing samples from a poisoned model based on a (clean) pre-trained GAN does not recover backdoored samples. We require the following two assumptions.
200
+
201
+ Assumption 1. The generator G is L*-Lipschitz in latent vector input* h ∈ R
202
+ p.
203
+
204
+ Assumption 2. *For all* h ∈ R
205
+ p, if ∥h∥ ≥ B for some B > 0, then G(h) *has no semantic meaning.*
206
+ The Assumption 2 is justified by Figure 3 (b): We observe an apparent quality degradation of the generated images when increasing the norm of the latent vector h. When the norm is 106larger, the generated images do not contain semantic meaning. Note that both clean and backdoor data are considered to have semantic meaning. We are interested in whether the range of the generator range(G) contains backdoored data. Since the backdoored images still have semantic meaning by definition, if range(G) contains backdoored data they will be within the high-density region where the corresponding latent vector h has ∥h∥ ≤ B. By the Lipschitzness (which implies continuity) of G, it means the density of the distribution induced by the generator (G(h), h ∼
207
+ N (0, I)) on the backdoored data points > 0.
208
+
209
+ 1 Thus, we reduce the question of "*whether the range of* G
210
+ contains backdoored data" to "*whether the generator distribution has a non-zero density on backdoored data* points."
211
+ We proceed by formulating GAN training as a two-player game between the generator and discriminator.
212
+
213
+ The game terminates only when the two players reach a min-max solution where neither party has the incentive to deviate from the current state. Such a min-max solution is called *pure Nash equilibrium*. Based on the game-theoretic framework, we show the following result, and the proof is outlined in Appendix B.
214
+
215
+ Theorem 3 (Informal). *When the generator learns a distribution with non-negligible density on backdoored* data, the generator and discriminator cannot achieve pure strategy Nash equilibrium. This result implies that, no backdoored data point can appear in range(G) when the GAN is trained properly where the generator and discriminator reach equilibrium. Thus, no matter how we search over G(h) for different latent vector h during the MI step, it is impossible to find an h such that G(h) is a backdoored image.
216
+
217
+ ## 5 Evaluation
218
+
219
+ Our evaluation focuses on the following aspects:
220
+ - Assess the effectiveness of FreD to enable data-free backdoor removal, and the benefit of FreD
221
+ with a limited amount of clean samples available (Sec. 5.2, Sec. D);
222
+ - Study application of FreD to adversarial fine-tuning (Jeddi et al., 2020), which aims to robustify a pre-trained model against adversarial examples (Goodfellow et al., 2014) (Sec. 5.2, Sec. D);
223
+ - Ablation study on several design choices of FreD, including different loss terms and the number of synthesized samples (Sec. 5.2);
224
+ - Study performance of FreD with varing distribution shifts OOD datasets (Sec. C.1);
225
+ - Study the effectiveness of FreD supplying other existing backdoor cleansing methods (Sec. E).
226
+
227
+ - Study the effectiveness of FreD against adaptive attacks (Sec. F).
228
+
229
+ Our evaluation setting reflects real-world scenarios, where the training data of pre-trained GANs often differs from the target model's training set in terms of distribution (e.g., pre-trained GAN is usually built with large, open data, yet the target model could be built by some stakeholder with proprietary datasets). We also considered out-of-distribution data with various distributional shifts from the target-model training data to enable a comprehensive evaluation (refer to Section C.1).
230
+
231
+ ## 5.1 Experimental Setup
232
+
233
+ Data. We evaluate datasets built for different prediction tasks, including face recognition, traffic sign classification, and general object recognition. For each task, we choose two datasets, one used for training the poisoned model and another for learning a pre-trained GAN. Detailed usage of the datasets is shown in Table 3.
234
+
235
+ Backdoor Attacks. We evaluate nine different kinds of backdoor attacks in all-to-one settings (the target model will misclassify all other classes' samples patched with the trigger as the target class), including the 1Note that the set of latent vectors with semantic meaning {h : ∥h∥ < B} is an open set.
236
+
237
+ | Face | Traffic Sign | General Object | |
238
+ |-----------------|----------------------------|-----------------------------|------------------------------|
239
+ | Recognition | Classification | Detection | |
240
+ | Poisoned Model | PubFig(Pinto et al., 2011) | GTSRB (Houben et al., 2013) | CIFAR-10 (Krizhevsky et al.) |
241
+ | Pre-trained GAN | CelebA (Liu et al., 2015) | TSRD(Huang) | STL-10 (Coates et al., 2011) |
242
+
243
+ Table 3: Datasets.
244
+
245
+ hidden trigger backdoor attack (Hidden) (Saha et al., 2020), input-aware backdoor (IAB) attack (Nguyen
246
+ & Tran, 2020a), WaNet (Nguyen & Tran, 2020b), L0 invisible (L0 inv) (Li et al., 2020a), L2 invisible (L2 inv) (Li et al., 2020a), the frequency invisible smooth (Smooth) attack (Zeng et al., 2021), trojan watermark
247
+ (Troj-WM) (Liu et al., 2018b), trojan square (Troj-SQ) (Liu et al., 2018b), and blend attack (Chen et al.,
248
+ 2017). The implementation details of the attacks are deferred to Appendix C.
249
+
250
+ Baselines. We compare FreD with five baselines, where the first four baselines differ in kinds of samples contained in the base set and share the same downstream backdoor removal technique, namely I-BAU,
251
+ which achieves the state-of-the-art backdoor removal performance given a clean base set. 1) Clean: The base set is formed by clean samples drawn from the original training data of the poisoned model. 2) Out-ofthe-distribution (OOD): The base set consists of ALL of the OOD samples that are used for learning the pre-trained GAN. 3) Naive: The base set contains samples synthesized by the MI adopted in Chen et al.
252
+
253
+ (2019) which directly optimizes in the pixel space. 4) GMI: The base set is formed by the synthetic samples from GMI (Zhang et al., 2020). The comparison between FreD and GMI will demonstrate the effectiveness of our designed loss terms. 5) CLP (Zheng et al., 2022): The last baseline is a recent data-free backdoor removal technique that does not utilize the idea of data synthesis. Instead, it prunes the neurons directly based on corresponding Lipschitz constants.
254
+
255
+ Our baselines provide comparison from different perspectives: 1) Clean is the original backdoor removal method with an extra clean in-distribution set provided, serving as the upper bound for defense performance. 2) OOD, Naive, GMI share the same assumption as our defense that the defender can leverage clean publicly available data from a related domain. This setting is practical in real-world scenarios considering the large amount of public data available online. 3) CLP is the state-of-the-art backdoor removal defense that is not base-set-reliant compared to this type of baselines, our method offers a unique plug-and-play benefit.
256
+
257
+ | | L0 inv | L2 inv | | | | | | | | | | | | |
258
+ |---------|----------|----------|-------|------|------|------|---------|-------|------|--------|------|------|------|------|
259
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP | Initial | Clean | OOD | Naive | GMI | FreD | CLP | |
260
+ | ACC | 0.97 | 0.98 | 0.88 | 0.88 | 0.93 | 0.95 | 0.94 | 0.97 | 0.98 | 0.9330 | 0.95 | 0.94 | 0.94 | 0.94 |
261
+ | ASR | 1.0 | 0.03 | 0.02 | 0.08 | 0.09 | 0.02 | 0.02 | 0.998 | 0.06 | 0.832 | 0.06 | 0.48 | 0.01 | 0.01 |
262
+ | | Smooth | Wanet | | | | | | | | | | | | |
263
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP | Initial | Clean | OOD | Naive | GMI | FreD | CLP | |
264
+ | ACC | 0.97 | 0.98 | 0.83 | 0.82 | 0.96 | 0.97 | 0.18 | 0.98 | 0.94 | 0.26 | 0.17 | 0.81 | 0.94 | 0.01 |
265
+ | ASR | 0.998 | 0.1 | 0.40 | 0.05 | 0.03 | 0.02 | 0.02 | 0.99 | 0.05 | 0.34 | 0.97 | 0.16 | 0.05 | 0.99 |
266
+ | | IAB | Troj-SQ | | | | | | | | | | | | |
267
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP | Initial | Clean | OOD | Naive | GMI | FreD | CLP | |
268
+ | ACC | 0.94 | 0.97 | 0.81 | 0.45 | 0.89 | 0.91 | 0.92 | 0.98 | 0.96 | 0.40 | 0.45 | 0.86 | 0.94 | 0.81 |
269
+ | ASR | 1.0 | 0.02 | 0.10 | 0.11 | 0.11 | 0.10 | 0.10 | 1.0 | 0.01 | 0.06 | 0.16 | 0.10 | 0.06 | 0.23 |
270
+ | | Troj-WM | Blend | | | | | | | | | | | | |
271
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP | Initial | Clean | OOD | Naive | GMI | FreD | CLP | |
272
+ | ACC | 0.98 | 0.96 | 0.47 | 0.62 | 0.84 | 0.87 | 0.87 | 0.98 | 0.96 | 0.47 | 0.68 | 0.82 | 0.92 | 0.75 |
273
+ | ASR | 1.0 | 0.01 | 0.30 | 0.09 | 0.30 | 0.09 | 0.12 | 1.0 | 0.08 | 0.51 | 0.55 | 0.48 | 0.22 | 0.85 |
274
+
275
+ Protocol. For all baselines except OOD, we draw/ generate 20 samples per class for PubFig and GTSRB;
276
+ 40 for CIFAR-10. A detailed study of choosing the number of samples to be generated for each class is shown in Section 5.2. For the hyperparameters, we fix λ1 = 1000, λ3 = 1000, λ4 = 10, set λ2 = 10 for PubFig and GTSRB, and λ2 = 100 for CIFAR-10. The defense performance is averaged over three random-initialized runs of I-BAU.
277
+
278
+ Table 4: Results of FreD boosted backdoor unlearning on GTSRB.
279
+
280
+ ## 5.2 Results
281
+
282
+ Data-Free Backdoor Defense. Table 4 shows that FreD outperforms naive MI, OOD, GMI and CLP
283
+ against various backdoor attacks on GTSRB. Results for the other datasets can be found in Appendix D and FreD remains the best. Figure 7 visualizes the samples synthesized by Naive, GMI, and FreD.
284
+
285
+ GMI, and FreD can generate samples with better visual quality, whereas Naive generates merely noise-like samples. This visualization explains the significant defense performance improvement achieved by GMI and FreD upon Naive. The performance of FreD is mostly on par with Clean. Interestingly, FreD achieves a higher ACC and comparable ASR than baseline utilizing clean data when defending against the IAB attack performed on the CIFAR-10 dataset. This may be because the model is overfitted to the clean training samples, and samples generated by FreD reduce the degree of overfitting by providing more abundant features. Note that CLP fails in defending against Smooth, Wanet, and Blend attack: Either the ACC drops to close to zero (Smooth and Wanet), or ASR remains high (Wanet and Blend), or both occurs (Wanet). As the above triggers have large norm and hence induce large changes in the model input, CLP's assumption that backdoor-related neurons in the poisoned model have a large Lipschitz constant does not hold. To better interpret the data synthesis process of FreD, we show a series of samples generated at different iterations in Figure 6. We observe that the appearance of the generated images varies significantly over the first ten optimization iterations and stabilizes afterwards.
286
+
287
+ Data-Limited Backdoor Defense. Here, we evaluate the benefits of FreD when there exists a tiny amount of clean samples. Particularly, we consider a stress test with 1 sample. With a single sample, even the state-of-the-art data-reliant backdoor removal technique works poorly as shown in the CIFAR-10 and PubFig83 results in Table 5). To evaluate FreD, we use FreD with the proposed feature consistency loss Lcon to generate 20 additional samples for each class, and the final result (FreD-Booster) is obtained by using the combination of both 1 clean sample and 20 generated samples for each class. Table 5 shows that FreD can significantly boost the defense performance compared to solely using the available clean sample(s).
288
+
289
+ Moreover, using 20 samples from FreD plus one clean sample gives better defense performance than 20 clean samples. As a final note, compared to FreD, CLP, the data-free backdoor removal baseline based on model pruning, cannot be benefited from additional clean samples.
290
+
291
+ | Clean(20) | | Clean(1) | FreD-Booster | |
292
+ |-------------|-----|------------|----------------|------|
293
+ | GTSRB | ACC | 0.98 | 0.93 | 0.98 |
294
+ | Smooth | ASR | 0.01 | 1 | 0 |
295
+ | CIFAR-10 | ACC | 0.82 | 0.52 | 0.85 |
296
+ | IAB | ASR | 0.03 | 0.18 | 0.01 |
297
+ | PubFig | ACC | 0.86 | 0.44 | 0.86 |
298
+ | Troj-WM | ASR | 0.03 | 0.35 | 0 |
299
+
300
+ | Initial | GMI | Lmp | Ldp | Lmp + Ldp | | |
301
+ |-----------|--------|-------|-------|-------------|-------|-------|
302
+ | GTSRB | ACC(%) | 98.8 | 86.52 | 92.70 | 89.32 | 94.49 |
303
+ | ASR(%) | 100.0 | 10.06 | 8.63 | 8.25 | 5.90 | |
304
+ | PubFig | ACC(%) | 92.21 | 70.52 | 72.32 | 70.61 | 83.53 |
305
+ | ASR(%) | 100.0 | 26.06 | 3.25 | 1.63 | 2.88 | |
306
+
307
+ Table 5: Results of backdoor unlearning performance with a small amount of clean data and generated samples.
308
+
309
+ Table 6: Ablation study of proposed model-perturbation loss Lmp and data-perturbation loss Ldp.
310
+
311
+ Ablation of Loss Terms. We proposed two loss terms 1) model-perturbation loss Lmp and 2) dataperturbation loss Ldp to improve the utility of the synthesized samples in the data-free backdoor defense setting. Table 6 presents an ablation study of the two losses on a poisoned model trained on GTSRB under the trojan square attack as well as a model trained on Celeba under the trojan watermark attack. We observe that ldp improves the ASR more than lmp while lmp is a more critical driver of maintaining the ACC
312
+ compared to ldp. This observation aligns with our design objectives. Recall that ldp is designed to enable effective synthesis of backdoor trigger and thus directly related to the reduction of ASR. On the other hand, lmp encourages the stability of prediction to small parameter changes, which in turn mitigates catestrophic forgetting during unlearning; hence, it is directly related to maintaining the clean accuracy. Ablation of Base Set Sizes. We study the impact of the number of the synthesized samples contained in the base set. We choose the number of samples for each class to be [1, 5, 10, 15, 20, 25, 30] and evaluate the defense performance by averaging over 3 runs of the defense. To better interpret the performance of FreD,
313
+ we also compare with Clean, Naive and GMI using the same amount of samples. Note that this experiment excludes the OOD baseline: we use the poisoned model to generate pseudo-labels for the OOD samples but because the label space of the OOD samples and that of the poisoned model may not overlap, the labeled OOD
314
+ samples are insufficient or even none for some classes. Figure 4 shows that the defense performance keeps increasing as the number of samples increases and converges to optimum when the number of samples for each class is above 20. FreD, GMI, and Clean can maintain a high ACC when a larger number of generated samples are used, but Naive suffers a significant ACC drop. We also observe during the experiments that performance of Naive has a large variance when evaluated over base set with different size. The variance could induce from the inconsistent quality among generated samples, or instability of the samples against input perturbation, leading to synthesizing inaccurate triggers. On the other hand, the variance of FreD
315
+ is similar to the variance of Clean, indicating good generalizability of FreD-enabled defenses. Another interesting finding is that when performing unlearning with a small amount of samples (i.e., 1 or 5 per class),
316
+ FreD even achieves higher ACC than Clean.
317
+
318
+ Figure 4: Ablation study of the number of samples used on backdoor defenses on the GTSRB with L0 inv
319
+
320
+ ![11_image_0.png](11_image_0.png)
321
+
322
+ attack.
323
+
324
+ ## 6 Conclusion
325
+
326
+ In this paper, we investigate the connection between model inversion and backdoor defense, and present FreD to generate synthetic samples that can be used as a substitute for clean in-distribution data to support backdoor removal. FreD can also be used to boost defense performance when only limited clean in-distribution data are available. This work sets a foundation towards developing highly effective indistribution-data-free backdoor defenses. In particular, one can potentially supply our synthetic data to other future defenses to enable their data-free mode of usage or improve their performance in the limited data setting.
327
+
328
+ ## 7 Acknowledgement
329
+
330
+ RJ and the ReDS lab acknowledge support through grants from the Amazon-Virginia Tech Initiative for Efficient and Robust Machine Learning, the National Science Foundation under Grant No. IIS-2312794, NSF IIS-2313130, NSF OAC-2239622, and the Commonwealth Cyber Initiative. JTW is supported by Princeton's Gordon Y. S. Wu Fellowship.
331
+
332
+ ## References
333
+
334
+ David Alvarez-Melis and Nicolo Fusi. Geometric dataset distances via optimal transport. *NeurIPS*, 33:
335
+ 21428–21439, 2020. 19 Shengwei An, Guanhong Tao, Qiuling Xu, Yingqi Liu, Guangyu Shen, Yuan Yao, Jingwei Xu, and Xiangyu Zhang. Mirror: Model inversion for deep learning network with high fidelity. In Proceedings of the Network and Distributed Systems Security Symposium (NDSS 2022), 2022. 3, 4, 5 Eugene Bagdasaryan and Vitaly Shmatikov. Blind backdoors in deep learning models. In *USENIX Security*,
336
+ pp. 1505–1521, 2021. 21, 22 David Bau, Jun-Yan Zhu, Jonas Wulff, William Peebles, Hendrik Strobelt, Bolei Zhou, and Antonio Torralba. Seeing what a gan cannot generate. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4502–4511, 2019. 8 Eitan Borgnia, Valeriia Cherepanova, Liam Fowl, Amin Ghiasi, Jonas Geiping, Micah Goldblum, Tom Goldstein, and Arjun Gupta. Strong data augmentation sanitizes poisoning and backdoor attacks without an accuracy tradeoff, 2020. 16 Bryant Chen, Wilka Carvalho, Nathalie Baracaldo, Heiko Ludwig, Benjamin Edwards, Taesung Lee, Ian Molloy, and Biplav Srivastava. Detecting backdoor attacks on deep neural networks by activation clustering.
337
+
338
+ CEUR Workshop Proceedings, 2018. 16 Huili Chen, Cheng Fu, Jishen Zhao, and Farinaz Koushanfar. Deepinspect: A black-box trojan detection and mitigation framework for deep neural networks. In *IJCAI*, volume 2, pp. 8, 2019. 2, 3, 4, 10, 16 Si Chen, Mostafa Kahla, Ruoxi Jia, and Guo-Jun Qi. Knowledge-enriched distributional model inversion attacks. In *Proceedings of the IEEE/CVF International Conference on Computer Vision*, pp. 16178–16187, 2021. 3, 4, 16 Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. Targeted backdoor attacks on deep learning systems using data poisoning, 2017. 10, 20 Edward Chou, Florian Tramer, and Giancarlo Pellegrino. Sentinet: Detecting localized universal attacks against deep learning systems. In *2020 IEEE Security and Privacy Workshops (SPW)*, pp. 48–54. IEEE,
339
+ 2020. 16 Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In *Proceedings of the fourteenth international conference on artificial intelligence and statistics*,
340
+ pp. 215–223. JMLR Workshop and Conference Proceedings, 2011. 10 Bao Gia Doan, Ehsan Abbasnejad, and Damith C Ranasinghe. Februus: Input purification defense against trojan attacks on deep neural network systems. In *Annual Computer Security Applications Conference*,
341
+ pp. 897–912, 2020. 16 Matt Fredrikson, Somesh Jha, and Thomas Ristenpart. Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pp. 1322–1333, 2015. 4, 16 Matthew Fredrikson, Eric Lantz, Somesh Jha, Simon Lin, David Page, and Thomas Ristenpart. Privacy in pharmacogenetics: An {End-to-End} case study of personalized warfarin dosing. In 23rd USENIX
342
+ Security Symposium (USENIX Security 14), pp. 17–32, 2014. 4 Yansong Gao, Change Xu, Derui Wang, Shiping Chen, Damith C. Ranasinghe, and Surya Nepal. Strip:
343
+ A defence against trojan attacks on deep neural networks. In *ACM ACSAC*, pp. 113–125, 2019. ISBN
344
+ 9781450376280. doi: 10.1145/3359789.3359790. 16 Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples.
345
+
346
+ arXiv preprint arXiv:1412.6572, 2014. 9 Wenbo Guo, Lun Wang, Xinyu Xing, Min Du, and Dawn Song. Tabor: A highly accurate approach to inspecting and restoring trojan backdoors in ai systems. *arXiv preprint arXiv:1908.01763*, 2019. 2, 4, 16, 22 Sebastian Houben, Johannes Stallkamp, Jan Salmen, Marc Schlipsing, and Christian Igel. Detection of traffic signs in real-world images: The German Traffic Sign Detection Benchmark. In *International Joint* Conference on Neural Networks, number 1288, 2013. 5, 7, 10 LinLin Huang. Chinese traffic sign database. URL http://www.nlpr.ia.ac.cn/pal/trafficdata/
347
+ recognition.html. 10 Ahmadreza Jeddi, Mohammad Javad Shafiee, and Alexander Wong. A simple fine-tuning is all you need:
348
+ Towards robust deep learning via adversarial fine-tuning. *arXiv preprint arXiv:2012.13628*, 2020. 9 Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In *Proceedings of the IEEE/CVF conference on computer vision and pattern recognition*, pp.
349
+
350
+ 4401–4410, 2019. 5 Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Training generative adversarial networks with limited data. *Advances in Neural Information Processing Systems*,
351
+ 33:12104–12114, 2020. 5 James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. *Proceedings of the national academy of sciences*, 114(13):3521–3526, 2017.
352
+
353
+ 6 Pang Wei Koh and Percy Liang. Understanding black-box predictions via influence functions. In Doina Precup and Yee Whye Teh (eds.), *Proceedings of the 34th International Conference on Machine Learning*, volume 70 of *Proceedings of Machine Learning Research*, pp. 1885–1894, International Convention Centre, Sydney, Australia, 06–11 Aug 2017. PMLR. URL http://proceedings.mlr.press/v70/koh17a.html.
354
+
355
+ 16 Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. Cifar-10 (canadian institute for advanced research).
356
+
357
+ URL http://www.cs.toronto.edu/~kriz/cifar.html. 10 Shaofeng Li, Minhui Xue, Benjamin Zhao, Haojin Zhu, and Xinpeng Zhang. Invisible backdoor attacks on deep neural networks via steganography and regularization. *IEEE Transactions on Dependable and Secure* Computing, 2020a. 2, 5, 7, 8, 10, 20 Yige Li, Xixiang Lyu, Nodens Koren, Lingjuan Lyu, Bo Li, and Xingjun Ma. Neural attention distillation: Erasing backdoor triggers from deep neural networks. In *International Conference on Learning* Representations, 2020b. 16 Kang Liu, Brendan Dolan-Gavitt, and Siddharth Garg. Fine-pruning: Defending against backdooring attacks on deep neural networks. In *International Symposium on Research in Attacks, Intrusions, and Defenses*,
358
+ pp. 273–294. Springer, 2018a. 2 Yingqi Liu, Shiqing Ma, Yousra Aafer, Wen-Chuan Lee, Juan Zhai, Weihang Wang, and Xiangyu Zhang.
359
+
360
+ Trojaning attack on neural networks. In *25nd Annual Network and Distributed System Security Symposium,* NDSS. The Internet Society, 2018b. 7, 10, 20 Yingqi Liu, Wen-Chuan Lee, Guanhong Tao, Shiqing Ma, Yousra Aafer, and Xiangyu Zhang. Abs: Scanning neural networks for back-doors by artificial brain stimulation. In Proceedings of the 2019 ACM SIGSAC
361
+ Conference on Computer and Communications Security, pp. 1265–1282, 2019. 16 Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp. 3730–3738, 2015. 10 Wanlun Ma, Derui Wang, Ruoxi Sun, Minhui Xue, Sheng Wen, and Yang Xiang. The" beatrix"resurrections:
362
+ Robust backdoor detection via gram matrices. *NDSS*, 2022. 8 Divyam Madaan, Jinwoo Shin, and Sung Ju Hwang. Learning to generate noise for multi-attack robustness.
363
+
364
+ In *International Conference on Machine Learning*, pp. 7279–7289. PMLR, 2021. 7 Tuan Anh Nguyen and Anh Tran. Input-aware dynamic backdoor attack. *NeurIPS*, 33:3454–3464, 2020a.
365
+
366
+ 10, 20 Tuan Anh Nguyen and Anh Tuan Tran. Wanet-imperceptible warping-based backdoor attack. In *ICLR*,
367
+ 2020b. 7, 10, 20 Nicolas Pinto, Zak Stone, Todd Zickler, and David Cox. Scaling up biologically-inspired computer vision:
368
+ A case study in unconstrained face recognition on facebook. In *CVPR 2011 WORKSHOPS*, pp. 35–42.
369
+
370
+ IEEE, 2011. 10 Han Qiu, Yi Zeng, Shangwei Guo, Tianwei Zhang, Meikang Qiu, and Bhavani Thuraisingham. Deepsweep:
371
+ An evaluation framework for mitigating dnn backdoor attacks using data augmentation. In *Proceedings* of the 2021 ACM Asia Conference on Computer and Communications Security, pp. 363–377, 2021. 16 Aniruddha Saha, Akshayvarun Subramanya, and Hamed Pirsiavash. Hidden trigger backdoor attacks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp. 11957–11965, 2020. 2, 10, 20 Guangyu Shen, Yingqi Liu, Guanhong Tao, Shengwei An, Qiuling Xu, Siyuan Cheng, Shiqing Ma, and Xiangyu Zhang. Backdoor scanning for deep neural networks through k-arm optimization. In *International* Conference on Machine Learning, pp. 9525–9536. PMLR, 2021. 16 Jun Shu, Qi Xie, Lixuan Yi, Qian Zhao, Sanping Zhou, Zongben Xu, and Deyu Meng. Meta-weight-net:
372
+ Learning an explicit mapping for sample weighting. *Advances in neural information processing systems*,
373
+ 32, 2019. 7 Hossein Souri, Liam Fowl, Rama Chellappa, Micah Goldblum, and Tom Goldstein. Sleeper agent: Scalable hidden trigger backdoors for neural networks trained from scratch. Advances in Neural Information Processing Systems, 35:19165–19178, 2022. 21, 22 Lukas Struppek, Dominik Hintersdorf, Antonio De Almeida Correia, Antonia Adler, and Kristian Kersting. Plug & play attacks: Towards robust and flexible model inversion attacks. arXiv preprint arXiv:2201.12179, 2022. 3, 4, 5, 19 Brandon Tran, Jerry Li, and Aleksander Madry. Spectral signatures in backdoor attacks. In *NeurIPS*, pp.
374
+
375
+ 8000–8010, 2018. 16 Bolun Wang, Yuanshun Yao, Shawn Shan, Huiying Li, Bimal Viswanath, Haitao Zheng, and Ben Y Zhao.
376
+
377
+ Neural cleanse: Identifying and mitigating backdoor attacks in neural networks. In 2019 IEEE Symposium on Security and Privacy (SP), pp. 707–723. IEEE, 2019. 2, 4, 16, 22 Kuan-Chieh Wang, Yan Fu, Ke Li, Ashish Khisti, Richard Zemel, and Alireza Makhzani. Variational model inversion attacks. *Advances in Neural Information Processing Systems*, 34, 2021. 3 Yi Zeng, Won Park, Z Morley Mao, and Ruoxi Jia. Rethinking the backdoor attacks' triggers: A frequency perspective. In *ICCV*, 2021. 2, 7, 10, 16, 20 Yi Zeng, Si Chen, Won Park, Zhuoqing Mao, Ming Jin, and Ruoxi Jia. Adversarial unlearning of backdoors via implicit hypergradient. In *International Conference on Learning Representations*, 2022. 4, 5, 16 Yuheng Zhang, Ruoxi Jia, Hengzhi Pei, Wenxiao Wang, Bo Li, and Dawn Song. The secret revealer: Generative model-inversion attacks against deep neural networks. In *Proceedings of the IEEE/CVF Conference* on Computer Vision and Pattern Recognition, pp. 253–261, 2020. 2, 3, 4, 5, 10, 16 Runkai Zheng, Rongjun Tang, Jianze Li, and Li Liu. Data-free backdoor removal based on channel lipschitzness. *arXiv preprint arXiv:2208.03111*, 2022. 2, 10, 16
378
+
379
+ ## A Related Works
380
+
381
+ Backdoor Defenses. Backdoor defenses normally can be performed on two levels: data-level and modellevel. For data-level detection or cleaning, the defender aims to identify (Gao et al., 2019; Chen et al.,
382
+ 2018; Tran et al., 2018; Koh & Liang, 2017; Chou et al., 2020; Zeng et al., 2021) or purify (Doan et al., 2020) the poison input in the training set, thus requiring the access to training data. Model-level detection
383
+ (Liu et al., 2019; Shen et al., 2021) or cleaning, instead, aims to detect if a pre-trained model is poisoned or mitigate vulnerabilities of the models. In this paper, we focus on model-level cleaning. Most of the prior works on this line (Wang et al., 2019; Chen et al., 2019; Guo et al., 2019; Zeng et al., 2022; Li et al.,
384
+ 2020b; Borgnia et al., 2020; Qiu et al., 2021) require a small set of clean data to synthesize triggers and further perform unlearning. Among them, I-BAU (Zeng et al., 2022) has achieved the state-of-the-art defense performance against a wide range of existing attacks. However, the performance of I-BAU degrades as the number of clean samples reduces. We aim to enable I-BAU to function effectively without any clean data.
385
+
386
+ A recent work (Zheng et al., 2022) proposes a method to perform backdoor removal without using clean data. They identify model channels with high Lipschitz constants, which are directly calculated from the weight matrices, as backdoor related channels; and do simple pruning to repair the model. However, their method only applies to backdoor scenarios and cannot benefit from clean samples if available. Our method, by contrast, also applies to evasion attacks and is able to leverage available clean samples to further boost defense performance. Above all, the performance of our method is more favorable. Model Inversion. The goal of model inversion (MI) is similar to ours. But from an attack perspective, MI
387
+ aims to divulge sensitive attributes in the training data, and to achieve this goal, the generated data should have good visual quality. Fredrikson *et al.* (Fredrikson et al., 2015) follows the maximum likelihood principle and performs MI by searching over the image space for a sample with highest likelihood under the given target model. DeepInspect (Chen et al., 2019) employs this simple MI to generate a surrogate training set for backdoor unlearning and achieves good results on MNIST and GTSRB. However, we find that samples generated by this naive MI approach have bad visual quality and usually fail in downstream defenses on high-dimensional datasets (e.g., PubFig and CIFAR-10). In this paper, we build upon the idea of recent MI works (Zhang et al., 2020; Chen et al., 2021) that search for a synthetic sample in the latent space of a pre-trained GAN instead of the image space. Even when the GAN is not trained on the in-distribution data, this idea can greatly help improve the visual quality of synthesized samples. The key innovations that set our work apart from the MI attacks is that we go beyond the traditional "high-likelihood" assumption made in all existing MI works about clean data and further formalize other plausible assumptions, especially those related to data- and model-stability. We show that enforcing the synthetic data to satisfy these assumptions can significantly improve their utility for defenses.
388
+
389
+ ## B Why Is Backdoored Data Not On Gan'S Range?
390
+
391
+ Notations. Throughout the section, we use d for the dimension of samples, and p for the dimension of the latent vector. We denote discriminator D : R
392
+ d → [0, 1], generator G : R
393
+ p → R
394
+ d. We use Preal to denote the real distribution the GAN aims to learn. The generator G defines a distribution PG as follows: generate latent vector h ∼ N (0, Ip) from p-dimensional spherical standard Gaussian distribution, and then apply G
395
+ on h and generate a sample x = G(h). We denote the class of discriminators as D = {D} and the class of generators G = {G}. Ideally, D is the class of all 1-Lipschitz functions. For a distribution P, we use P(x) to denote the density of P on x, and P(S) denotes the Lebesgue integration RI[x ∈ S]dP(x). We use EP [D]
396
+ as an abbreviation for Ex∼P [D(x)]. We use supp(·) to denote the support of distribution. We use dW(·, ·) to denote 1-Wasserstein distance with ℓ2-norm, i.e., the Earth Mover distance.
397
+
398
+ The two required assumptions are demonstrated in previous section 4.2. To formally state our theorem, we formulate the training of GAN as a game between generator and discriminator.
399
+
400
+ Definition 4 (Payoff). For a class of generators G = {G} and a class of discriminators D = {D}, we define the payoff F(D, G) *of the game between generator and discriminator as*
401
+
402
+ $$F(D,G)=\operatorname*{\mathbb{E}}_{x\sim{\mathcal{P}}_{\mathrm{real}}}[D(x)]-\operatorname*{\mathbb{E}}_{x\sim{\mathcal{P}}_{G}}[D(x)]$$
403
+ [D(x)] (5)
404
+ $\left(5\right)$.
405
+ The generator and discriminator aims at reaching a min-max solution, i.e., the *pure Nash equilibrium*, where neither party has the incentive to deviate from the current state..
406
+
407
+ Definition 5 (pure equilibrium). A pair of strategy (D∗, G∗) *is a pure equilibrium if for some value* V ,
408
+
409
+ $$\forall D\in{\mathcal{I}}$$
410
+ $$\leq V$$
411
+
412
+ ∀D ∈ D, F(*D, G*∗) ≤ V
413
+
414
+ $\mathbf{\hat{\hat{\varepsilon}}}$ V.
415
+ ∀G ∈ G, F(D∗, G) ≥ V
416
+ However, such an equilibrium may not be achievable for a pure strategy setting. We introduce a natural relaxation for quantifying the extent of equilibrium between a pair of generator/discriminator.
417
+
418
+ Definition 6 (ε-approximate pure equilibrium). A pair of strategy (D∗, G∗) is an ε*-approximate pure equilibrium if for some value* V ,
419
+
420
+ $$\begin{array}{l}{{\forall D\in{\mathcal{D}},F(D,G^{*})\leq V+\varepsilon}}\\ {{\forall G\in{\mathcal{G}},F(D^{*},G)\geq V-\varepsilon}}\end{array}$$
421
+
422
+ We are now ready to state our main results.
423
+
424
+ General Idea: game-theoretic interpretation of GAN. From a game theory perspective, the generator and discriminator in GAN are engaged in a two-player zero-sum game. In game theory, a Nash Equilibrium is a state in which neither player can improve their own outcome by changing their strategy, assuming the other player keeps their strategy fixed. In the context of GANs, a Nash Equilibrium is achieved when the Generator produces data that is indistinguishable from real data, and the Discriminator is no better than random guessing at telling the difference between real and generated data. In Theorem 3 (or Theorem 7), we show that when the generator's learned distribution contains a significant probability mass on backdoored images, which is out-of-distribution (OOD) data, the equilibrium cannot be established between the generator and discriminator, which forces the generator to reduce the probability mass on the backdoored images.
425
+
426
+ Theorem 7 (Formal version of Theorem 3). Given any two distributions P1,P2 *s.t. for the set* SOOD =
427
+ {x ∈ *supp*(P2) : miny∈supp(P1)∪*supp*(Preal) ∥x − y∥ ≥ 1}*, we have* P2(SOOD) ≥ 1 − q
428
+ ′*for some* q
429
+ ′ ∈ [0, 1)*. Let* D∗ = arg maxD∈D EPreal [D]−EP1
430
+ [D] and D∗ ∈ D∗. If G *induce a mixture distribution* PG = (1−q)P1 +qP2 for some q ∈ (0, 1), then there exists no D ∈ D s.t. (D, G) is ε-approximate pure equilibrium for any ε <
431
+ 1 2 q(EP1
432
+ [D∗]−q
433
+ ′)*. Moreover, when* Preal = P1, we have ε < 12 q(1−q
434
+ ′). Further more, given Assumption 1 and 2, we can lower bound q if range(G) contains backdoored data, which leads to ε < 12
435
+ (1−q
436
+ ′)
437
+ 1 L
438
+ √2 pexp(− 12 B2)
439
+ Γ(p/2+1) ,
440
+ where Γ *is the Gamma function.*
441
+ Interpretation. To interpret the above theorem statement, one can regard P1 as a clean distribution (not necessarily Preal), and P2 as a distribution that contains backdoor data on its support. Since backdoored images are separated from clean images (i.e., out-of-distribution (OOD) data), we can assume that all backdoored images are within the set SOOD = {x ∈ supp(P2) : miny∈supp(P1)∪supp(Preal) ∥x − y∥ ≥ 1}.
442
+
443
+ The above theorem thus states that no equilibrium could be achieved if PG has non-negligible density on SOOD. The theorem essentially reveals the tension between a generator's goal to trick the discriminator and its propensity to generate backdoored (OOD) data. In simple terms, think of P1 as a distribution representing normal, clean images, and P2 representing malicious, backdoored images. The generator mixes these distributions to create PG, which it uses to try to fool the discriminator. Now, if the PG mixture has a significant probability of sampling backdoored (OOD) images from P2, it is much easier for the discriminator to spot the differences between the generated and real data. Hence, there exists a better move for the generator to increase its "payoff" by sampling less from P2, hence an equilibrium becomes unattainable under such PG (captured by the upper bound of ε). In essence, the generator is in a quandary:
444
+ including backdoored data in the mixture makes it easier for the discriminator to win, which is contrary to what the generator wants.
445
+
446
+ Remark. It is also possible that P1 *also supports on* {x : miny∈*supp*(Preal) ∥x − y∥ ≥ 1}, but this leads to vacuous results.
447
+
448
+ ## B.1 Proof Of The Formal Theorem
449
+
450
+ Proof Overview. In Lemma 8, we derive the upper bound of the degree of equilibrium ε when PG =
451
+ (1 − q)P1 + qP2 in terms of the best discriminator in distinguishing between Preal and P1, i.e., D∗ =
452
+ arg maxD∈D EPreal [D] − EP1
453
+ [D]. Then in Lemma 9 and 10, we further bound different terms in the upper bound of ε derived in Lemma 8 under Assumption 1 and 2 stated in the maintext.
454
+
455
+ Lemma 8. Given any two distributions P1,P2*, let* D∗ = arg maxD∈D EPreal [D]−EP1
456
+ [D]. For any D∗ ∈ D∗,
457
+ if G *induce a distribution* PG = (1 − q)P1 + qP2, then there exists no D ∈ D s.t. (D, G) is ε*-approximate* pure equilibrium for any ε < 12 q(EP1
458
+ [D∗] − EP2
459
+ [D∗]).
460
+
461
+ Proof. We define an alternative generator G∗s.t. PG∗ = P1. Given any discriminator D, the payoff gain of G by switching strategy to G∗is
462
+
463
+ $$F(D,G)-F(D,G^{*})$$ $$=(1-q)(\mathbb{E}_{\mathcal{P}_{1}\mathrm{res}}[D]-\mathbb{E}_{\mathcal{P}_{1}}[D])+q(\mathbb{E}_{\mathcal{P}_{1}}[D]-\mathbb{E}_{\mathcal{P}_{2}}[D])-(\mathbb{E}_{\mathcal{P}_{1}\mathrm{res}}[D]-\mathbb{E}_{\mathcal{P}_{2}}[D])\tag{6}$$ $$=q(\mathbb{E}_{\mathcal{P}_{1}}[D]-\mathbb{E}_{\mathcal{P}_{2}}[D])$$
464
+
465
+ Given any discriminator D, the payoff gain of D by switching strategy to D∗is
466
+
467
+ F(D∗, G) − F(D, G) = EPreal [D∗] − (1 − q)EP1 [D∗] − qEP2 [D∗] − (EPreal [D] − (1 − q)EP1 [D] − qEP2 [D]) = EPreal [D∗] − EP1 [D∗] + q(EP1 [D∗] − EP2 [D∗]) − (EPreal [D] − EP1 [D]) − q(EP1 [D] − EP2 [D]) = dW(Preal,P1) + q(EP1 [D∗] − EP2 [D∗]) − (EPreal [D] − EP1 [D]) − q(EP1 [D] − EP2 [D]) (7)
468
+
469
+ $$({\boldsymbol{\delta}})$$
470
+ $$(10)$$
471
+ By Definition 6, (*D, G*) cannot be ε-approximate equilibrium for
472
+
473
+ $${}^{*}),F(D^{*},G)-F(D,G))$$
474
+
475
+ ε < max (F(D, G) − F(D, G∗), F(D∗, G) − F(*D, G*)) (8)
476
+ since otherwise at least one of D and G will gain more than ε by changing its strategy to D∗ or G∗. Therefore, we are interested in lower bounding minD max (F(D, G) − F(D, G∗), F(D∗, G) − F(*D, G*)). Note that the minimum can only be achieved when F(D, G) − F(*D, G*∗) = F(D∗, G) − F(*D, G*), where we have
477
+
478
+ $$LHS=q(\mathbb{E}_{\mathcal{P}_{1}}[D]-\mathbb{E}_{\mathcal{P}_{2}}[D])\tag{9}$$ $$=d_{\mathbb{E}}(\mathcal{P}_{\text{real}},\mathcal{P}_{1})+q(\mathbb{E}_{\mathcal{P}_{1}}[D^{*}]-\mathbb{E}_{\mathcal{P}_{2}}[D^{*}])-(\mathbb{E}_{\mathcal{P}_{\text{real}}}[D]-\mathbb{E}_{\mathcal{P}_{1}}[D])-q(\mathbb{E}_{\mathcal{P}_{1}}[D]-\mathbb{E}_{\mathcal{P}_{2}}[D])$$ $$=RHS$$
479
+
480
+ and we have
481
+
482
+ $$2LHS=d_{\mathbb{E}}(\mathcal{P}_{\mathrm{real}},\mathcal{P}_{1})+q(\mathbb{E}_{\mathcal{P}_{1}}[D^{\star}]-\mathbb{E}_{\mathcal{P}_{2}}[D^{\star}])-(\mathbb{E}_{\mathcal{P}_{\mathrm{real}}}[D]-\mathbb{E}_{\mathcal{P}_{1}}[D])$$ $$\geq q(\mathbb{E}_{\mathcal{P}_{1}}[D^{\star}]-\mathbb{E}_{\mathcal{P}_{2}}[D^{\star}])$$
483
+
484
+ where the last inequality is due to
485
+
486
+ $$\mathbb{E}_{\mathcal{P}_{\text{real}}}[D]-\mathbb{E}_{\mathcal{P}_{1}}[D]\leq\sup_{D\in\mathcal{D}}\mathbb{E}_{\mathcal{P}_{\text{real}}}[D]-\mathbb{E}_{\mathcal{P}_{1}}[D]$$ $$=d_{\mathcal{U}}(\mathcal{P}_{\text{real}},\mathcal{P}_{1})$$
487
+ $$(11)$$
488
+
489
+ Therefore,
490
+
491
+ $$\min_{D}\max\left(F(D,G)-F(D,G^{*}),F(D^{*},G)-F(D,G)\right)$$ $$\geq\frac{1}{2}q\left(\mathbb{E}\rho_{1}[D^{*}]-\mathbb{E}\rho_{2}[D^{*}]\right)$$
492
+ $$(12)$$
493
+
494
+ Remark. *This result may be of independent interest.*
495
+ Lemma 9. *Consider the set* SOOD = {x ∈ *supp*(P2) : miny∈supp(P1)∪*supp*(Preal) ∥x − y∥ ≥ 1}*. If* P2(SOOD) ≥
496
+ 1 − q
497
+ ′*for some* q
498
+ ′ ∈ [0, 1]*, then we have* EP2
499
+ [D∗] ≤ q
500
+ ′.
501
+
502
+ Proof. The value of D on supp(P2) \ (supp(Preal) ∪ supp(P1)) does not affect EPreal [D] − EP1
503
+ [D], therefore we only need to ensure that D∗satisfies the Lipschitz assumption. It is easy to see that D∗(x) can be 0 for all x ∈ SOOD. Since P2(SOOD) ≥ 1 − q
504
+ ′, we know that EP2
505
+ [D∗] ≤ q
506
+ ′.
507
+
508
+ Therefore, we know that (*G, D*) is ε-approximate pure equilibrium only for ε ≥
509
+ 1 2 q (EP1
510
+ [D∗] − q
511
+ ′). Moreover, when Preal = P1, it is impossible to distinguish between Preal and P1 and thus D∗contains function D that output D(x) = 1 for all x ∈ supp(Preal). Thus ε ≥
512
+ 1 2 q (1 − q
513
+ ′).
514
+
515
+ Now we lower bound q, based on Assumption 1 and 2.
516
+
517
+ $\text{Lemma10.}q\geq\left(\frac{1}{L\sqrt{2}}\right)^p\frac{\mathrm{e}}{q}$
518
+ pexp(− 12 B2)
519
+ $\frac{(-\frac{1}{2}B^2)}{p/2+1)}$.
520
+ Proof. Suppose for some h ∈ R
521
+ p we have G(h) ∈ SOOD, then for all ∥h
522
+ ′ − h∥ we have ∥G(h
523
+ ′) − G(h)∥ ≤
524
+ L∥h
525
+ ′ − h∥ ≤ 1. Therefore, G(h
526
+ ′) ∈ supp(P2). Therefore, h
527
+ ′ within the ball centered at h with radius 1/L
528
+ will all have G(h
529
+ ′) ∈ supp(P2), and thus
530
+
531
+ $$q\geq\frac{1}{(2\pi)^{p/2}}\exp(-B^{2}/2)\frac{\pi^{p/2}}{\Gamma(p/2+1)}(1/L)^{p}\tag{13}$$ $$=\left(\frac{1}{L\sqrt{2}}\right)^{p}\frac{\exp(-\frac{1}{2}B^{2})}{\Gamma(p/2+1)}$$
532
+ $\square$
533
+ Plugging this lower bound back to the original bound for ε leads to the final result in Theorem 7.
534
+
535
+ ## C Experimental Settings C.1 Choice Of Ood
536
+
537
+ Clean in-distribution data may not always accessible in real-world applications, however, given the access of the target model, one may make inference about the type of data the model is trained on, e.g., whether it is a face recognition model or digit classification model, etc. This offers us the chance to make use of public available data (OOD) of the same type. FreD does not assume the overlap of the label space between OOD
538
+ and the private data, as the OOD is only accessed during the GAN training stage.
539
+
540
+ Performance of FreD given different OOD is shown in Table 7, and the distance between OOD and the private distribution is calculated in Optimal Transport Dataset Distances (OTDD) (Alvarez-Melis & Fusi, 2020), which gives a meassurement of dataset distance even if the label sets are completely disjoint. As we can see, by using the in-distribution data, which in this case is CIFAR-10, FreD achieves ASR=0.01 while ACC is 0.87. Even when using OOD data which has a much larger distance, FreD can still achieve a comparable ASR and a even better ACC. The only exception is using Caltech-256 as OOD, the ASR is relatively higher compared with other settings. One possible reason is that Caltech-256 only has 30,609 samples, which is smallest dataset among the four datasets. GAN may not be well-trained on this dataset.
541
+
542
+ Given the fact that State-of-the-art MI provides could provide more advanced techniques, which relax the dependency between the target model and OOD (Struppek et al., 2022), FreD can also be benefit from it and leads to a even better performance.
543
+
544
+ | CIFAR-10 | STL-10 | Tiny-ImageNet | Caltech-256 | |
545
+ |------------|------------|-----------------|---------------|---------|
546
+ | ⇒ CIFAR-10 | ⇒ CIFAR-10 | ⇒ CIFAR-10 | ⇒ CIFAR-10 | |
547
+ | # Classes | 10 | 10 | 200 | 257 |
548
+ | OTDD | 324.3 | 3486.48 | 4068.28 | 3844.61 |
549
+ | ACC* | 0.96 | 0.33 | 0.74 | 0.57 |
550
+ | ACC | 0.87 | 0.85 | 0.89 | 0.80 |
551
+ | ASR | 0.01 | 0.05 | 0.02 | 0.24 |
552
+
553
+ Table 7: Results of FreD when using different OOD datasets. ACC* gives corresponding transfer accuracy respectively, and OTDD gives the distribution shift. The target model is trained on CIFAR-10 under IAB
554
+ attack.
555
+
556
+ ## C.2 Backdoor Attacks Implementation Details
557
+
558
+ We evaluate nine different kinds of backdoor attacks in all-to-one settings (the target model will misclassify all other classes' samples patched with the trigger as the target class), including the hidden trigger backdoor attack (Hidden) (Saha et al., 2020), input-aware backdoor (IAB) attack (Nguyen & Tran, 2020a), WaNet
559
+ (Nguyen & Tran, 2020b), L0 invisible (L0 inv) (Li et al., 2020a), L2 invisible (L2 inv) (Li et al., 2020a), the frequency invisible smooth (Smooth) attack (Zeng et al., 2021), trojan watermark (Troj-WM) (Liu et al.,
560
+ 2018b), trojan square (Troj-SQ) (Liu et al., 2018b), and blend attack (Chen et al., 2017).
561
+
562
+ On CIFAR, we test all the nine attacks listed above. Note that initially, Hidden can only work in one-toone attack settings where the goal is to fool one class with the trigger, thereby resulting in a low ASR in all-to-one settings. To address this issue, we manually increase the norm bound to 50/255 with one round of fine-tuning of a pre-trained clean model to achieve an acceptable ASR. However, the ASR of Hidden on GTSRB is still less than 10%, hence, we exclude it from evaluation on GTSRB. On PubFig, we adopt Trojan watermark (Troj-WM), Trojan square (Troj-SQ), and blend attack. The implementations of each attack follow the original works which propose them. The adopted trigger and the target label on each dataset is visualized in Fig. 5.
563
+
564
+ Figure 5: Datasets and examples of backdoor attacks incorporated. We consider three different datasets in
565
+
566
+ ![19_image_0.png](19_image_0.png)
567
+
568
+ this work: (1) CIFAR-10, (2) GTSRB, and (3) PubFig. Nine different backdoor attack triggers are included in the experimental part as listed. Above, we also show the target label used during the evaluated attacks
569
+ (e.g., Hidden targeting at label 8 of the CIFAR-10 dataset).
570
+
571
+ ## C.3 Fred Implementation Details
572
+
573
+ In our experiments, we choose the batch size B in Algorithm 1 to be 40, max iterations N = 4500, learning rate α2 = 1e − 3. We choose learning rate α1 = 2e − 3 for CIFAR-10 and α1 = 2e − 2 for the rest. For parameters in Equation 4, we choose λ1 = 1000 following the prior works on MI. We find λ3 = 1000, λ4 = 10 work well across different datasets and models. The best value for λ2 is task-dependant, and chose using grid search that yields the smallest value of Lprior + Lcl + Lmp + Ldp.
574
+
575
+ In our experiments, we use 8 NVIDIA GeForce RTX 2080 Ti. The synthesize base set generation takes approximately 3 hours for private dataset CIFAR-10 and GTSRB, and 7 hours for PubFig.The computational time of the backdoor removal depends on the downstream defense method used. For I-BAU, it takes 6.82 s on average on CIFAR-10, 7.84 s on GTSRB and 13.2s on PubFig for a single iteration, and can effectively mitigate backdoor trigger within 100 iterations.
576
+
577
+ ## D More Fred Boosted Defense Results
578
+
579
+ ![20_Image_0.Png](20_Image_0.Png)
580
+
581
+ Figure 6: Examples of model-specific synthesize by FreD at first 10 iterations for Identity 12 in PubFig
582
+
583
+ ![20_image_1.png](20_image_1.png)
584
+
585
+ dataset. Rightmost in the lower row is the real image of this identity from PubFig.
586
+
587
+ Figure 7: Examples of images obtained by FreD and naive MI. Each subplot represents randomly generated samples for the same class. The upper row shows images generated by FreD and the lower row shows images generated by naive MI.
588
+
589
+ More results of FreD boosted backdoor defense are given in Table 8,10, which FreD yields the best performance besides the clean baseline. One interesting finding is that FreD achieves a lower ASR and comparable ACC than baseline utilizing clean data when defending against the L2 inv attack performed on the CIFAR-10 dataset. This may be because the model is overfitted to the clean training samples, and samples generated by FreD reduce the degree of overfitting with more abundant features. To provide a better understanding of how FreD works, we give the images generated by FreD in the first 10 iterations for Identity 12 in PubFig dataset. As shown in Figure 6, appearance of the generated fake person gradually changed to be closer to the real identity.
590
+
591
+ We also conduct additional experiments to provide results on recent backdoor attacks, namely Blind(Bagdasaryan & Shmatikov, 2021) and SleeperAgent(Souri et al., 2022), to enhances the comprehensiveness of our evaluation. Specifically, we utilized the open-source implementation 2 of these two attacks.
592
+
593
+ It's worth noting that neither of these attacks was originally evaluated on the GTSRB dataset, and their performance was found to be poor. The ASR of SleeperAgent remained consistently low (below 10%), while Blind could not effectively balance ASR and benign ACC 3, where in either case, the ASR was very low or the ACC was very low. Hence, we excluded the evaluation of these two attacks on the GTSRB dataset.
594
+
595
+ | L0 inv | | L2 inv | | Smooth | | | | | | | | | | | | | | |
596
+ |----------|---------|----------|-------|----------|------|-------------|-------|------|-------|------|------|-------------|-------|------|-------|------|------|-----------|
597
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP |
598
+ | ACC | 0.93 | 0.89 | 0.47 | 0.44 | 0.70 | 0.76 0.69 | 0.94 | 0.90 | 0.45 | 0.87 | 0.89 | 0.90 0.67 | 0.93 | 0.83 | 0.48 | 0.54 | 0.83 | 0.83 0.23 |
599
+ | ASR | 0.97 | 0.15 | 0.07 | 0.09 | 0.09 | 0.06 0.06 | 0.99 | 0.07 | 0.12 | 0.06 | 0.07 | 0.01 0.05 | 0.95 | 0.18 | 0.02 | 0.94 | 0.20 | 0.18 0.86 |
600
+ | Wanet | | IAB | | Hidden | | | | | | | | | | | | | | |
601
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP |
602
+ | ACC | 0.94 | 0.91 | 0.84 | 0.23 | 0.79 | 0.81 0.75 | 0.94 | 0.82 | 0.11 | 0.34 | 0.84 | 0.85 0.70 | 0.76 | 0.89 | 0.11 | 0.66 | 0.86 | 0.89 0.35 |
603
+ | ASR | 0.99 | 0.01 | 0.32 | 0.32 | 0.03 | 0.03 0.05 | 0.99 | 0.03 | 0 | 0.08 | 0.06 | 0.05 0.03 | 0.88 | 0.09 | 0 | 0.16 | 0.13 | 0.09 0.94 |
604
+ | Troj-SQ | Troj-WM | | | Blend | | | | | | | | | | | | | | |
605
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP |
606
+ | ACC | 0.94 | 0.81 | 0.51 | 0.14 | 0.71 | 0.71 0.71 | 0.94 | 0.80 | 0.50 | 0.23 | 0.74 | 0.75 0.66 | 0.94 | 0.80 | 0.85 | 0.65 | 0.76 | 0.79 0.46 |
607
+ | ASR | 1.0 | 0.02 | 0.07 | 0.37 | 0.28 | 0.06 0.05 | 1.0 | 0.04 | 0.11 | 0.22 | 0.02 | 0.01 0.58 | 0.99 | 0.05 | 0.71 | 0.05 | 0.06 | 0.06 0.28 |
608
+
609
+ Table 8: Results of FreD boosted backdoor unlearning on CIFAR-10.
610
+
611
+ 2https://github.com/THUYimingLi/BackdoorBox 3https://github.com/ebagdasa/backdoors101/issues/17)
612
+
613
+ SleeperAgent **Blind**
614
+
615
+ Initial Clean OOD Naive GMI FreD CLP **Initial Clean OOD Naive GMI FreD CLP**
616
+
617
+ ACC 0.94 0.91 0.30 0.81 0.85 **0.86** 0.86 0.67 0.79 0.16 0.35 0.44 **0.54** 0.50
618
+
619
+ ASR 0.91 0.02 0.31 0.10 0.10 **0.02** 0.08 0.78 0.09 0.23 0.25 0.10 **0.07** 0.20
620
+
621
+ Table 9: Results of FreD boosted backdoor unlearning on CIFAR-10 on recent backdoor attacks SleeperAgent(Souri et al., 2022) and Blind (Bagdasaryan & Shmatikov, 2021).
622
+
623
+ | Troj-WM | Troj-SQ | Blend | | | | | | | | | | | | | | | | | | | |
624
+ |-----------|-----------|---------|-------|------|------|-------------|-------|------|-------|-------|------|-------------|-------|------|-------|------|------|------|------|------|------|
625
+ | Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP Initial | Clean | OOD | Naive | GMI | FreD | CLP | | | |
626
+ | ACC | 0.92 | 0.86 | 0.06 | 0.13 | 0.83 | 0.83 | 0.01 | 0.92 | 0.84 | 0.02 | 0.18 | 0.74 | 0.78 | 0.01 | 0.91 | 0.88 | 0.06 | 0.12 | 0.83 | 0.84 | 0.03 |
627
+ | ASR | 1.0 | 0.03 | 0.04 | 0.23 | 0.10 | 0.03 | 0.82 | 1.0 | 0.04 | 0.002 | 0.01 | 0.06 | 0.06 | 0.89 | 1.0 | 0.44 | 0.06 | 0.78 | 0.82 | 0.52 | 0.93 |
628
+
629
+ ## Table 10: Results Of Fred Boosted Backdoor Unlearning On Pubfig. E Fred Combined With Other Backdoor Defenses
630
+
631
+ | | IAB | | | | | Smooth | | | | | | |
632
+ |---------|---------|------|-------|------|--------------|----------|------|-------|------|------|------|------|
633
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
634
+ | ACC | 0.95 | 0.89 | 0.89 | 0.88 | 0.89 | 0.89 | 0.93 | 0.87 | 0.81 | 0.86 | 0.85 | 0.87 |
635
+ | ASR | 0.99 | 0.03 | 0.08 | 0.06 | 0.06 | 0.04 | 0.95 | 0.43 | 0.69 | 0.89 | 0.59 | 0.45 |
636
+ | | L0 inv | | | | | L2 inv | | | | | | |
637
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
638
+ | ACC | 0.93 | 0.91 | 0.80 | 0.81 | 0.87 | 0.89 | 0.94 | 0.91 | 0.82 | 0.81 | 0.94 | 0.88 |
639
+ | ASR | 0.97 | 0.03 | 0.85 | 0.91 | 0.05 | 0.04 | 1.00 | 0.31 | 0.63 | 0.59 | 0.88 | 0.28 |
640
+ | | Troj-SQ | | | | | Troj-WM | | | | | | |
641
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
642
+ | ACC | 0.94 | 0.88 | 0.69 | 0.89 | 0.72 | 0.91 | 0.94 | 0.86 | 0.77 | 0.75 | 0.85 | 0.85 |
643
+ | ASR | 1.00 | 0.08 | 0.98 | 0.99 | 0.10 | 0.00 | 1.00 | 0.09 | 0.98 | 0.98 | 0.16 | 0.07 |
644
+
645
+ While FreD shows satisfying results when combined with I-BAU, which is one of the most popular backdoor defense framework, in this paper we also show the application of FreD combined with other backdoor defense techniques. Note that we focus on model-level cleansing in this paper. Specifically, we supply Neural Cleanse
646
+ (Wang et al., 2019) and Tabor (Guo et al., 2019) with FreD synthesized data. As shown in Table 11 and 12, FreD outperforms OOD, Naive, GMI; and even achieves better results than Clean baseline under some settings.
647
+
648
+ Table 11: Results of FreD combined with Neural Cleanse on defending against various attacks on CIFAR10.
649
+
650
+ ## F Fred Against Adaptive Attacks
651
+
652
+ Another interesting case is the attacker may have knowledge about the pre-trained GAN and its corresponding training data, and design an adaptive attack to surpass our proposed defense. To investigate such setting, we designed the following attack pipeline:
653
+ 1) Selecting a more promising source class: When performing a one-to-one attack, the attacker can choose a source class (i.e., the class to be poisoned) exhibiting greater similarity to the OOD data. For example, when targeting a GTSRB dataset-trained target model, the attacker may select class 6 ('80 lifted') as the source class, which has a similar pattern to class 7 in the OOD (TSRD) dataset.
654
+
655
+ 2) Designing a trigger pattern closer to the OOD distribution: To achieve this, we utilize G(z) as our trigger, where z = arg maxz D(G(z)+source image), where D is the discriminator of the GAN and G is the generator.
656
+
657
+ The rationale behind this approach is that we expect the poisoned source image to exhibit higher confidence under the Discriminator, making it easier to be generated by GAN.
658
+
659
+ We compare our designed adaptive attack to all the eight non-adaptive attacks we evaluate in this paper. As the baseline. As shown by Figure 8 below, the attack performance (ASR) of the adaptive procedure is lower compared to the non-adaptive ones without any defense. The reason is that this trigger is constrained to be on a specific embedding space, limiting its attack capability. However, when comparing 'After Defense' ASR
660
+ of FreD, the adaptive attack indeed demonstrates better attack performance than the non-adaptive baseline, but the improvement is marginal. It is important to note that the current design is a simple and intuitive
661
+
662
+ | | IAB | | Smooth | | | | | | | | | |
663
+ |---------|---------|------|----------|--------|--------------|-------|------|-------|------|------|------|------|
664
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
665
+ | ACC | 0.93 | 0.90 | 0.89 | 0.89 | 0.89 | 0.89 | 0.93 | 0.91 | 0.85 | 0.85 | 0.84 | 0.86 |
666
+ | ASR | 0.95 | 0.03 | 0.10 | 0.06 | 0.06 | 0.04 | 0.95 | 0.35 | 0.93 | 0.84 | 0.66 | 0.48 |
667
+ | | L0 inv | | | L2 inv | | | | | | | | |
668
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
669
+ | ACC | 0.93 | 0.91 | 0.83 | 0.80 | 0.85 | 0.88 | 0.94 | 0.88 | 0.82 | 0.77 | 0.86 | 0.87 |
670
+ | ASR | 0.97 | 0.07 | 0.56 | 0.94 | 0.03 | 0.01 | 1.00 | 0.11 | 0.97 | 0.25 | 0.47 | 0.21 |
671
+ | | Troj-SQ | | Troj-WM | | | | | | | | | |
672
+ | Initial | Clean | OOD | Naive | GMI | FreD Initial | Clean | OOD | Naive | GMI | FreD | | |
673
+ | ACC | 0.94 | 0.89 | 0.81 | 0.86 | 0.81 | 0.90 | 0.94 | 0.88 | 0.80 | 0.81 | 0.84 | 0.88 |
674
+ | ASR | 1.00 | 0.13 | 0.99 | 0.77 | 0.00 | 0.00 | 1.00 | 0.40 | 0.94 | 0.95 | 0.20 | 0.16 |
675
+
676
+ Table 12: Results of FreD combined with Tabor on defending against various attacks on CIFAR-10.
677
+
678
+ ![22_image_0.png](22_image_0.png)
679
+
680
+ example. FreD could be vulnerable to a more advanced adaptive attack, which is one of the limitations of this paper. Further improvements will be explored in future work.
681
+
682
+ Figure 8: Comparison of the attack performance of non-adaptive attacks (L0 inv, L2 inv, IAB, Troj-WM,
683
+ Troj-SQ, Wanet, Blend) and adaptive attacks (Adaptive). Figure on the left gives attack results without any defense, and figure on the right gives results of attack performance after defense with FreD.
XuOE99cmST/XuOE99cmST_meta.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "languages": null,
3
+ "filetype": "pdf",
4
+ "toc": [],
5
+ "pages": 23,
6
+ "ocr_stats": {
7
+ "ocr_pages": 0,
8
+ "ocr_failed": 0,
9
+ "ocr_success": 0,
10
+ "ocr_engine": "none"
11
+ },
12
+ "block_stats": {
13
+ "header_footer": 23,
14
+ "code": 0,
15
+ "table": 11,
16
+ "equations": {
17
+ "successful_ocr": 29,
18
+ "unsuccessful_ocr": 4,
19
+ "equations": 33
20
+ }
21
+ },
22
+ "postprocess_stats": {
23
+ "edit": {}
24
+ }
25
+ }