RedTachyon commited on
Commit
a01994c
1 Parent(s): 2dd5b69

Upload folder using huggingface_hub

Browse files
FEEKR0Vl9s/10_image_0.png ADDED

Git LFS Details

  • SHA256: 2603373801c4a5247c8d9e0a337a0d737a6f333ea211e8f491e3ca5fa7a61379
  • Pointer size: 130 Bytes
  • Size of remote file: 74 kB
FEEKR0Vl9s/12_image_0.png ADDED

Git LFS Details

  • SHA256: c5200a5bbd8c2bd241077736c4d27bbab2526e95f374fd144f275d4625397e92
  • Pointer size: 130 Bytes
  • Size of remote file: 41.9 kB
FEEKR0Vl9s/13_image_0.png ADDED

Git LFS Details

  • SHA256: f7bb2789da0f53f658fe453e9b90a6de63ed6156c204872ea3e4afc7df64a89d
  • Pointer size: 131 Bytes
  • Size of remote file: 230 kB
FEEKR0Vl9s/14_image_0.png ADDED

Git LFS Details

  • SHA256: 2f84c4290cdea0f1cb1109478a8ea29dd5a5a6e7c95d7337220fbcc4c0e51506
  • Pointer size: 131 Bytes
  • Size of remote file: 249 kB
FEEKR0Vl9s/15_image_0.png ADDED

Git LFS Details

  • SHA256: 772c7b17ca7e1ec8c8dac6d0f5da9c70922e71d737d96ac819a23ea63533e1e5
  • Pointer size: 131 Bytes
  • Size of remote file: 288 kB
FEEKR0Vl9s/FEEKR0Vl9s.md ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Reproducibility Study On Adversarial Attacks Against Robust Transformer Trackers
2
+
3
+ Fatemeh Nourilenjan Nokabadi fatemeh.nourilenjan-nokabadi.1@ulaval.ca IID, Université Laval & *Mila* Jean-François Lalonde *jflalonde@gel.ulaval.ca* IID, Université Laval Christian Gagné christian.gagne@gel.ulaval.ca IID, Université Laval Canada CIFAR AI Chair, Mila Reviewed on OpenReview: *https: // openreview. net/ forum? id= FEEKR0Vl9s*
4
+
5
+ ## Abstract
6
+
7
+ New transformer networks have been integrated into object tracking pipelines and have demonstrated strong performance on the latest benchmarks. This paper focuses on understanding how transformer trackers behave under adversarial attacks and how different attacks perform on tracking datasets as their parameters change. We conducted a series of experiments to evaluate the effectiveness of existing adversarial attacks on object trackers with transformer and non-transformer backbones. We experimented on 7 different trackers, including 3 that are transformer-based, and 4 which leverage other architectures. These trackers are tested against 4 recent attack methods to assess their performance and robustness on VOT2022ST, UAV123 and GOT10k datasets. Our empirical study focuses on evaluating adversarial robustness of object trackers based on bounding box versus binary mask predictions, and attack methods at different levels of perturbations. Interestingly, our study found that altering the perturbation level may not significantly affect the overall object tracking results after the attack. Similarly, the sparsity and imperceptibility of the attack perturbations may remain stable against perturbation level shifts. By applying a specific attack on all transformer trackers, we show that new transformer trackers having a stronger cross-attention modeling achieve a greater adversarial robustness on tracking datasets, such as VOT2022ST and GOT10k. Our results also indicate the necessity for new attack methods to effectively tackle the latest types of transformer trackers. The codes necessary to reproduce this study are available at https://github.com/fatemehN/ReproducibilityStudy.
8
+
9
+ ## 1 Introduction
10
+
11
+ Adversarial perturbations deceive neural networks, leading to inaccurate outputs. Such adversarial attacks have been studied for vision tasks ranging from image classification (Mahmood et al., 2021; Shao et al.,
12
+ 2022) to object segmentation (Gu et al., 2022) and tracking (Guo et al., 2020; Jia et al., 2020; Yan et al.,
13
+ 2020; Jia et al., 2021). In this context, transformer-based networks have surpassed other deep learning-based trackers (Li et al., 2019; Zhu et al., 2018), showing a very robust performance on the state-of-the-art benchmarks (Kristan et al., 2023). However, the adversarial robustness of these trackers has not been thoroughly studied in the literature. First, transformer trackers relied on relatively light relation modeling (Chen et al.,
14
+ 2021), using a shallow feature extraction and fusion modeling. Based on a mixed attention module, the MixFormer (Cui et al., 2022) expanded the road for deeper relation modeling. Consequently, the Robust Object Modeling Tracker (ROMTrack) (Cai et al., 2023) proposed variation tokens to capture and preserve the object deformation across frames. Using transformers, especially those with deep relation modeling (Cui et al., 2022; Cai et al., 2023), the object tracker backbones made these models robust to many existing attack approaches (Guo et al., 2020; Jia et al., 2020). Indeed, the underlying concept of adversarial attacks against object trackers is to manipulate the tracker's output. By omitting the multi-head pipelines and substituting them with the transformer backbones (Cui et al., 2022; Cai et al., 2023), the tracker's output no longer contains object candidates, classification labels and/or regression labels, which previously were typical targets for attacks. As a result, it is not straightforward to transfer adversarial attacks dealing with classification or regression labels (Guo et al., 2020; Jia et al., 2020) on these new transformer trackers. The question also remains on whether other transferable attacks (Jia et al., 2021; Yan et al., 2020) can represent a sufficient challenge to transformer trackers.
15
+
16
+ This paper presents a study on the reproducibility of existing attack approaches for transformer trackers.
17
+
18
+ We aim to recreate the attack outcomes on transformer trackers using the VOT2022ST (Kristan et al.,
19
+ 2023), UAV123 (Mueller et al., 2016), DAVIS2016 (Perazzi et al., 2016) and GOT10k (Huang et al., 2019)
20
+ datasets following two different evaluation protocols, namely *anchor-based short-term tracking* and One Pass Evaluation (OPE) protocol. We focus on transformer trackers susceptible to adversarial attacks on their prediction outputs, including the object bounding box and binary mask. Then, we analyzed two whitebox attacks on a transformer tracker by varying the perturbation levels and checked its vulnerability to different levels of noise. For the black-box setting, we conducted a similar experiment to assess the attack performance on various noise levels by changing the upper bound of the added noise. In addition, we tested a set of transformer and non-transformer trackers before and after applying the adversarial attacks to discuss the role of transformers in boosting the visual tracking robustness and adversarial robustness. In short, our contributions can be summarized as follows:
21
+ 1. We extend the application of adversarial attacks, originally designed for non-transformer trackers like SPARK and RTAA, to assess their effectiveness against transformer-based trackers.
22
+
23
+ 2. We thoroughly evaluate the adversarial robustness of transformer-based trackers across various output scenarios, perturbation levels, changes in upper bounds, and in comparison to the non-transformer trackers.
24
+
25
+ ## 2 Related Work 2.1 Visual Object Trackers
26
+
27
+ For decades, the tracking task in computer vision has been extensively explored, taking into account various factors and issues. Before the advent of transformers, deep learning-based trackers achieved notable success by leveraging correlation in the form of Siamese networks (Li et al., 2019; Zhu et al., 2018), and discriminative pipelines for trackers (Bhat et al., 2019; Danelljan et al., 2020). However, new correlation modules in object tracker backbones are being devised through the emergence of transformers. By using transformers in different architectures (Chen et al., 2021; Cui et al., 2022; Chen et al., 2023; Cai et al., 2023), object trackers are capable of inferring object bounding box, binary mask and a prediction score with high robustness values over tracking benchmarks (Kristan et al., 2023; Mueller et al., 2016). These promising results though need to be revisited by assessing transformer trackers in handling the adversarial perturbations.
28
+
29
+ The first transformer tracker, TransT (Chen et al., 2021), used the cross-attention and self-attention blocks to mix features of the moving target and the search region of the tracker. TransT presents a multi-head pipeline with classification and regression heads, unlike other transformer trackers. In TransT-SEG (Chen et al., 2023), the segmentation head is included in the pipeline. The multi-head pipelines follow the Siamesebased trackers (Li et al., 2019) in dividing each task from target classification to discriminative tracking processing into individual blocks and fusing the results at the end of the tracker structure. Some light relation modeling layers called the Ego Context Augment (ECA) and Cross Feature Augment (CFA) are introduced by TransT to infer the output from combining the multi-head outputs. Next, the MixFormer (Cui et al., 2022) introduced Mixed Attention Module (MAM) to jointly extract and relate the information from video frames for the object tracking task. By attaching the MixFormer (Cui et al., 2022) tracker to the AlphaRefine (Yan et al., 2021), MixFormerM (Kristan et al., 2023) can provide a binary object mask per frame for mask oriented evaluations (Kristan et al., 2023). The One-Stream Tracking (OSTrack) (Ye et al.,
30
+ 2022) developed a tracking pipeline that jointly extracts features and models the relation between the search region and the template by bidirectional information flows. In contrast, with the Attention in Attention (AiA) mechanism, the AiATrack (Gao et al., 2022) suggested a three stream framework with long-term and short-term cross-attention modules for relation modeling. Following the further relation modeling, the Robust Object Modeling Tracker (ROMTrack) (Cai et al., 2023) is proposed to enable the interactive template learning using both self-attention and cross-attention modules. The ROMTrack has two main streams to learn discriminative features from hybrid (template and search) and inherent template. The newly introduced variation tokens enable ROMTrack with heavier relation modeling rather TransT (Chen et al., 2021) and MixFormer (Cui et al., 2022). The variation token carries the contextual appearance change to tackle object deformation in visual tracking task.
31
+
32
+ ## 2.2 Adversarial Attacks Against Trackers
33
+
34
+ The adversarial attack has been proposed in white-box (Guo et al., 2020; Jia et al., 2020) or black-box (Jia et al., 2021) attack settings. In black-box attacks, the perturbations are generated without relying on the tracker's gradients, whereas in white-box attacks, adversarial losses are backpropagated through the networks to create the adversarial frame patches (search or template regions). Adversarial attacks against object trackers adopt tracker outputs such as object candidates or classification labels as an attack proxy to generate the adversarial perturbations. For instance, spatial-aware online incremental attack (SPARK) (Guo et al., 2020) creates perturbations by manipulating the classification labels and Intersection of the Union
35
+ (IoU) (Jia et al., 2021). It is developed to mislead trackers in providing an altered object bounding box based on the predicted bounding box. In Robust Tracking against Adversarial Attack (RTAA) (Jia et al., 2020), both classification and regression labels are used to generate the adversarial samples, similar to SPARK (Guo et al., 2020). In the RTAA (Jia et al., 2020) algorithm, the positive gradient sign is used to generate the adversarial frames. However, in SPARK, the gradient direction is set to negative following the decoupling of the norm and the direction of gradients in white-box attacks (Rony et al., 2019). Based on the decoupling direction and norm for efficient gradient-based attacks, the direction of the gradient is set in such a way that the generated perturbation has a smaller norm value and greater impact on the results. Some other attacks, such as the Cooling-Shrinking Attack (CSA) (Yan et al., 2020) is developed specifically to impact the output of Siamese-based trackers. In CSA (Yan et al., 2020), two GANs are trained to cool the hottest regions in the final heatmap of Siamese-based trackers and shrink the object bounding box. Due to dependency on the Siamese-based architecture and loss function, the generalization of the CSA attack (Yan et al., 2020) for other scenarios is harder. The black-box attack, called IoU attack (Jia et al., 2021), adds two types of noise into the frame to make the tracker predict another bounding box rather than the target bounding box. By considering the object motion in historical frames, the direction of added noise is adjusted according to the IoU scores of the predicted bounding box.
36
+
37
+ Before the emergence of vision transformers, some object trackers typically integrated several heads, each assigned to specific vision tasks like classification or segmentation (Li et al., 2019; 2018; Zhu et al., 2018; Danelljan et al., 2020). These deep features were fused to predict the final object position and size, with regression computed over features. In these models, the ultimate decision was made at the network's end based on a set of object candidates. This architectural setup presented various opportunities for crafting adversarial attacks (Guo et al., 2020; Jia et al., 2020) against these models, including manipulation of object candidates, object probabilities, and other features, thereby exploiting vulnerabilities in the system. Vision transformers facilitate the integration of features within the deep architecture, enabling direct prediction of the final output, without exposing intermediate outputs that were previously exploitable for attacks. Consequently, white-box attacks utilizing these intermediate outputs (namely, object candidates and their labels) to compute the adversarial loss are no longer applicable to the new transformer backbones via the transformer's gradients themselves. Although they can be transferred in a black-box way (i.e., making adversarial samples with other backbones or other losses), our focus is to employ the transformer gradients in generating adversarial examples in a white-box setting.
38
+
39
+ ## 3 Object Trackers And Adversarial Attacks
40
+
41
+ In this section, we briefly review the transformer trackers used in our experiments. Also, we explain the adversarial attack methods which are used to attack transformer trackers. The codes and networks of all of the investigated models are publicly available. The implementations of every tracker and every attack approach are the official repository announced by authors. We are also using the fine-tuned and released networks from the authors of the original works.
42
+
43
+ ## 3.1 Object Trackers
44
+
45
+ For the object trackers, we considered three types of the robust single object trackers with transformer backbone, two types of the Siamese-based trackers and two discriminative trackers as follows. TransT and TransT-SEG In our studies, we used both Transformer Tracker (TransT) (Chen et al.,
46
+ 2021) and TransT with mask prediction ability (TransT-SEG) (Chen et al., 2023). By two discriminative streams and a lightweight cross-attention modeling in the end, the TransT introduced the first transformerbased tracker. Similar to the Siamese-based trackers (Li et al., 2018; 2019), the TransT tracker has two convolutional streams to extract features of the template and the search regions. By proposing Ego-Context Augment (ECA) and Cross-Feature Augment (CFA) modules, a feature fusing network is developed to effectively fuse the extracted features. The ECA infers output as
47
+
48
+ $$X_{\mathrm{ECA}}=X+\mathrm{MultiHead}(X+P_{x},\,X+P_{x},\,X)\,,$$
49
+ $\downarrow$ .
50
+
51
+ XECA = X + MultiHead(X + Px, X + Px, X), (1)
52
+ where "MultiHead" represents a multi-head attention module (Vaswani et al., 2017) and Px is the spatial positional encoding. On the other hand, the CFA module is defined as a Feed-Forward Network (FFN) attached to an ECA module. This FFN includes two linear transformations with a ReLU in between. The CFA's output is computed as:
53
+
54
+ $$\begin{array}{c}{{\tilde{X}_{\mathrm{CF}}=X_{q}+\mathrm{MultiHead}(X_{q}+P_{q},\,X_{k v}+P_{k v},\,X_{k v}),}}\\ {{X_{\mathrm{CFA}}=\tilde{X}_{\mathrm{CF}}+\mathrm{FFN}(\tilde{X}_{\mathrm{CF}}),}}\end{array}$$
55
+
56
+ $$\mathbf{\Sigma}$$
57
+ XCFA = X˜CF + FFN(X˜CF), (2)
58
+ where Xq is the input, and Pq is the spatial positional encoding of Xq. The input of the cross branch is Xkv and its spatial positional encoding is Pkv. Using two ECAs and two CFAs, the extracted features of the template and search regions are first fused with themselves by ECA and with each other by CFA.
59
+
60
+ Next, another cross attention (CFA) is used to integrate the outputs of the two streams. In the final prediction head, the regression and classification labels are generated to determine the target bounding box by finding the maximum score. The TransT-SEG (Chen et al., 2023) has a segmentation branch which uses the template vectors corresponding to the middle position of the template region to obtain an attention map via a multi-head attention module. The object binary mask is, then, predicted by fusing the low-level feature pyramid (Lin et al., 2017) of the search region.
61
+
62
+ MixFormer and MixFormerM The MixFormer (Cui et al., 2022) is based on a flexible attention operation named Mixed Attention Module (MAM) to interactively exploit features and integrate them in a deep layer of the tracker. The MixFormer coupled with the AlphaRefine network has been proposed for the VOT2022 challenge (Kristan et al., 2023) as MixFormerM. It enables the original tracker to provide the object mask as an extra output. In our experiments, we tested both MixFormer and MixFormerM trackers.
63
+
64
+ In the MixFormer tracker, the three stages of mixed attention is the core design named MAM which extracts and fuses the features of the template and the search regions. Given the target key, query and value
65
+ (kt, qt, vt) and the search key, query and value (ks, qs, vs), the mixed attention is defined as:
66
+
67
+ $$\begin{array}{r l}{{k_{m}=\operatorname{contact}(k_{t},k_{s}),}}&{{v_{m}=\operatorname{contact}(v_{t},v_{s}),}}\\ {{\mathrm{Attention}_{t}=\operatorname{softmax}(q_{t}k_{m}^{T}/{\sqrt{d}})\,v_{m},}}&{{\mathrm{Attention}_{s}=\operatorname{softmax}(q_{s}k_{m}^{T}/{\sqrt{d}})\,v_{m}\,,}}\end{array}$$
68
+
69
+ where d is the key dimension. The Attentiont and Attentions are the attention maps of the target and search regions. The target and search tokens are, then, concatenated and linearly projected to the output
70
+
71
+ $$\mathbf{\Sigma}$$
72
+
73
+ as the mixed attention output. To reduce the computational cost of MAM, the unnecessary cross-attention between the target query and search region is pruned by using the asymmetric mixed attention scheme. It is defined as follows:
74
+
75
+ $$\mathrm{Attention}_{t}=\mathrm{softmax}(q_{t}k_{t}^{T}/\sqrt{d})\,v_{t},\quad\mathrm{Attention}_{s}=\mathrm{softmax}(q_{s}k_{m}^{T}/\sqrt{d})\,v_{m}\,.$$
76
+ $$\left(4\right)$$
77
+
78
+ The Attentiont is updated to efficiently avoid the distractors in the search region by fixing the template tokens in the tracking process. In MixFormer, the Score Prediction Module (SPM) is developed by two attention blocks and a three-layer perceptron to identify the target confidence score and update a more reliable online template. ROMTrack The ROMTrack (Cai et al., 2023) is developed to generalize the idea of MixFormer (Cui et al.,
79
+ 2022) by providing the template learning procedure. The template feature is processed both in self-attention
80
+ (inherent template) and cross-attention (hybrid template) between template and search regions. This mixed feature avoids distraction in challenging frames and provides a more robust performance compared to TransT
81
+ and MixFormer. The backbone of ROMTrack is a vision transformer (Dosovitskiy et al., 2020) as an object encoder and a prediction head which is a fully convolutional center-based localization head (Zhou et al.,
82
+ 2019). Two essential elements of ROMTrack are variation tokens and robust object modeling. The variation tokens are developed to handle the object appearance change and deformation in the tracking process. Considering F
83
+ t k as the output features of k-th encoder in frame It and htk,t is the hybrid template part of F
84
+ t k
85
+ , a variation token is defined as
86
+
87
+ $$\begin{array}{l}{{v t_{k,t}=h t_{k,t-1},}}\\ {{F_{k+1}^{t}=\mathrm{ObjectEncoder}_{k+1}(\mathrm{concat}(v t_{k,t},F_{k}^{t}))\,,}}\end{array}$$
88
+
89
+ where F represents the output features, k is the encoder index and t denotes the t-th frame. In Equation 5, ROMTrack saves the hybrid template in the variation token and by Equation 6, it embeds the hybrid template into the output feature. The other essential element of ROMTrack is the robust object modeling containing four parts: inherit template it, hybrid template ht, search region sr and variation tokens vt. The inherit template is a self-attention module on the linear projected feature (qit, kit, vit) to learn the pure template features Ait. The hybrid template features and search regions features are obtained via cross-attention.
90
+
91
+ Considering the triplet of cross-attention (qz, kz, vz) as follows:
92
+
93
+ (5) (6) $\frac{1}{2}$
94
+ $q_{z}=\left[q_{ht},q_{sr}\right],$ $k_{z}=\left[k_{vt},k_{it},k_{ht},k_{sr}\right],$ $v_{z}=\left[v_{vt},v_{it},k_{ht},v_{sr}\right],$
95
+
96
+ $\downarrow$ .
97
+
98
+ $$\mathbf{\Sigma}$$
99
+
100
+ vz = [vvt, vit, kht, vsr] , (7)
101
+ where the different parts of features (*it, ht, sr, vt*) are rearranged and concatenated to create the crossattention input. The output of the cross-attention is obtained as:
102
+
103
+ $$A_{z}=\mathrm{softmax}(q_{z}k_{z}^{T}/\sqrt{d})\,v_{z}\,.$$
104
+
105
+
106
+ The features of inherit template and variation tokens are fused effectively with the hybrid template and search region tokens during the tracking process to keep the tracker updated about object deformation and appearance change. SiamRPN By integrating the Siamese network for tracking task and Region Proposal Network (RPN)
107
+ for detection, the SiamRPN tracker (Li et al., 2018) is developed to predict a list of object candidates per frame. The Siamese network contains two branches for feature extraction of the template φ(z) and search φ(x) regions. Then, the region proposal network with a pairwise correlation and supervision, split the feature vectors into the classification [φ(z)]cls, [φ(x)]cls and regression [φ(z)]reg, [φ(x)]reg vectors. The template feature maps [φ(z)]cls, [φ(z)]reg serve as the kernel for computing the correlation as follows:
108
+
109
+ $$\begin{array}{l}{{A^{c l s}=[\varphi(z)]_{\mathrm{cls}}\star[\varphi(x)]_{\mathrm{cls}}\,,}}\\ {{A^{r e g}=[\varphi(z)]_{\mathrm{cls}}\star[\varphi(z)]_{\mathrm{reg}}\,,}}\end{array}$$
110
+ reg = [φ(z)]cls ⋆ [φ(z)]reg , (9)
111
+ where ⋆ is the convolution operator. Using a softmax operation, the highest score for classification vector is chosen as the final target class and its corresponding coordinates in the regression vector determine the target bounding box. DaSiamRPN In DaSiamRPN tracker (Zhu et al., 2018), a distractor-aware strategy is taken to improve the SiamRPN tracker (Li et al., 2018) performance. In this strategy, the negative samples of object candidates(i.e., proposals) are considered distractor objects. Using Non Maximum Suppression (NMS), a set of distractors is selected to determine the target proposal as the candidate with the highest score. The rest of the set is considered as the distractors and re-ranked by a distractor-aware objective function to reduce their influence in the tracking process. DiMP The Discriminative Model Prediction (DiMP) tracker (Bhat et al., 2019) employs a discriminative approach that utilizes background information while also benefiting from the means of updating the target model. In contrast to Siamese-based trackers, which aim to identify the most correlated region as the target, DiMP processes background information within a discriminative framework. Initially, a classifier extracts input features, which are then utilized by a predictor to initialize the model and refine the target prediction.
112
+
113
+ The DiMP tracker comprises two essential components: Target Center Regression (TCR) and Bounding Box Regression (BBR). TCR produces a probabilistic map of the input image, wherein the target region is delineated more prominently compared to the background pixels. The BBR component generates the final bounding box prediction through a target conditional IoU-Net-based architecture (Jiang et al., 2018)
114
+ proposed in the ATOM tracker (Danelljan et al., 2019). PrDiMP The Probabilistic Regression DiMP (PrDiMP) (Danelljan et al., 2020) aims to forecast the conditional probability density p(y|*x, θ*) of the target given the input frame. Training the PrDiMP tracker involves utilizing the Kullback-Leibler (KL) divergence between the predicted density p(y|*x, θ*) and the conditional ground truth p(y|yi). The conditional ground truth p(y|yi) is formulated to account for label noise and the ambiguity inherent in the regression task, addressing the uncertainty associated with annotations.
115
+
116
+ PrDiMP builds upon the DiMP tracker (Bhat et al., 2019) by incorporating probabilistic elements into both the Target Center Regression (TCR) and Bounding Box Regression (BBR) components.
117
+
118
+ ## 3.2 Adversarial Attacks
119
+
120
+ In our study, we examined four attack approaches against object trackers, as follows.
121
+
122
+ CSA In attention-based Siamese trackers (Li et al., 2019), the loss function aims to locate the hottest region in the image where the correlation of the target and that location is the highest among all other regions.
123
+
124
+ Using two GANs, one for the template perturbation and the other for the search region perturbation, the CSA attack (Yan et al., 2020) is developed to firstly cool the hot regions in the end of the network and then, shrink the object bounding box predicted by the tracker. The perturbation generators, GANs, are trained to predict the adversarial template and search regions. For the cooling loss term, the object candidates with smaller scores are eliminated and then, the remaining candidates are divided into positive f+ and negative f− samples. The cooling term is computed as:
125
+
126
+ $$L_{\mathrm{cooling}}=\frac{1}{N}\max(f_{+}-f_{-},m_{c})\,,\tag{1}$$
127
+
128
+ where mc is the margin for the classification labels. For the shrinking term, the less likely candidate similar to the cooling term is removed and then, the shrinking loss is calculated as
129
+
130
+ $$L_{\mathrm{shrinking}}=\frac{1}{N}\operatorname*{max}(R_{w},m_{w})+\frac{1}{N}\operatorname*{max}(R_{h},m_{h})\,,$$
131
+
132
+ where mw and mh are the margins for the width and height regression factors, respectively. In our experiments, we used the template and search GANs trained with the cooling-shrinking loss L = Lcooling+Lshrinking to perturb the template and search regions for the trackers as a black-box attack.
133
+
134
+ IoU The IoU attack (Jia et al., 2021) proposes a black-box setting attack to generate the adversarial frames based on the object motion with the purpose of decreasing the IoU between the predicted bounding box and
135
+
136
+ $$(10)$$
137
+ $$(11)$$
138
+
139
+ the target. Two types of noises are added to achieve the final goal of the attack where the noise is bounded to a specific value for L1 norm. The IoU score SIoU is defined as
140
+
141
+ $$S_{\mathrm{IoU}}=\lambda S_{\mathrm{spatial}}+(1-\lambda)S_{\mathrm{temporal}}\,,$$
142
+ $$(12)$$
143
+ SIoU = λSspatial + (1 − λ)Stemporal , (12)
144
+ where Sspatial is the actual intersection over union between the predicted bounding box before and after adding the noise, while the Stemporal is the intersection over union between current and previous predicted bounding box. Using this IoU score, the final predicted bounding box is aimed at decreasing both in temporal and spatial domains. The level of noise is controlled depending on the IoU score and a final upper bound limits the algorithm in adding noise to the frame. SPARK In SPARK (Guo et al., 2020), the classification and regression labels are manipulated to create the white-box attack. We employed the untargeted SPARK attack in our experiments. The SPARK aims to minimize the perturbation in a way that the final intersection over union of the current prediction and previous prediction is minimum. Furthermore, a new regularization term is used in SPARK as
145
+
146
+ $$L_{\mathrm{reg}}=\lambda||\Gamma||_{1,2},\quad\Gamma=\left[\epsilon_{t-L},...,\epsilon_{t-1},\epsilon_{t}\right],$$
147
+ Lreg = λ||Γ||1,2, Γ = [ϵt−L, ..., ϵt−1, ϵt] , (13)
148
+ where Γ represents the incremental perturbations of the last L frames. The ϵt is computed as the ϵt =
149
+ Et − Et−1 in which Et is the current frame perturbation. The generated perturbation up to the last L = 30 frames is accumulated to create the adversarial search regions for the trackers. As a result, the computed perturbations are temporally and spatially sparse.
150
+
151
+ RTAA Using RTAA (Jia et al., 2020), the classification and regression of object candidates are manipulated to generate the adversarial search regions. The classification labels are simply reversed and the regression labels are manipulated as follows:
152
+
153
+ $$(13)$$
154
+ $$\begin{array}{l}{{x_{r}^{*}=x_{r}+\delta_{\mathrm{offset}}}}\\ {{y_{r}^{*}=y_{r}+\delta_{\mathrm{offset}}}}\\ {{w_{r}^{*}=w_{r}\times\delta_{\mathrm{scale}}}}\\ {{h_{r}^{*}=h_{r}\times\delta_{\mathrm{scale}}\;,}}\end{array}$$
155
+ $$(14)$$
156
+ r = hr × δscale , (14)
157
+ where δoffset and δscale are the random distance and scale variations, respectively. The adversarial loss is computed as the difference of the original loss for true labels and the manipulated labels. Only the last frame perturbation is used from the past in the current step of the attack.
158
+
159
+ ## 3.3 Attack Setups
160
+
161
+ In our study, we applied the attacks as they are proposed in the original works, with white-box attacks using the victim tracker's gradients, not transferring the attack from another tracker. For instance, SPARK (Guo et al., 2020) is attacking trackers in a not-transferred white-box setting. As SPARK uses both classification and regression labels to generate the perturbation, it cannot be applied in a white-box setting for some trackers that are not providing them both - e.g., Cui et al. (2022); Cai et al. (2023); Bhat et al. (2019);
162
+ Danelljan et al. (2020). Table 1 specifies the applicable attacks on visual trackers investigated in our study.
163
+
164
+ For instance, in MixFormer (Cui et al., 2022), the classification labels are fused by a score prediction module to infer one single score as the part of output, which is not compatible with SPARK. Also, in DiMP (Bhat et al., 2019) and PrDiMP (Danelljan et al., 2020), the regression labels for object candidates are not produced because these trackers predict the target bounding box directly. We have the same constraints with the RTAA attack (Jia et al., 2020), which also requires the classification and regression labels, and therefore, is not applicable as a white-box attack to the same trackers than SPARK. As for the CSA attack (Yan et al., 2020), since the template and search GANs are available, we are using these pre-trained networks and SiamRPN++ (Li et al., 2019) tracker to generate the corresponding perturbations. CSA is therefore applied in the black-box setting similar to the IoU attack (Jia et al., 2021). However, CSA attack is also not Table 1: Applicability of adversarial attacks (rows) to tracking methods (columns). In this work, we evaluate:
165
+ SPARK (Guo et al., 2020), RTAA (Jia et al., 2020), IoU (Jia et al., 2021), and CSA (Yan et al., 2020), against: ROMTrack (Cai et al., 2023), MixFormerM (Cui et al., 2022), TransT (Chen et al., 2021), DiMP (Bhat et al.,
166
+ 2019), PrDiMP (Danelljan et al., 2020), SiamRPN (Li et al., 2018), and DaSiamRPN (Zhu et al., 2018). We identify combinations as either applicable ("A") or not ("N/A").
167
+
168
+ | ROMTrack | MixFormerM | TransT | DiMP | PrDiMP | SiamRPN | DaSiamRPN | |
169
+ |------------------|--------------|----------|--------|----------|-----------|-------------|----|
170
+ | SPARK(white-box) | N/A | N/A | A | N/A | N/A | A | A |
171
+ | RTAA (white-box) | N/A | N/A | A | N/A | N/A | A | A |
172
+ | IoU (black-box) | A | A | A | A | A | A | A |
173
+ | CSA (black-box) | A | A | A | N/A | N/A | A | A |
174
+
175
+ applicable on DiMP and PrDiMP trackers, as they do not have template and search regions in the tracking process. This lead us to achieve our experiments over two white-box attacks, SPARK and RTAA, and two black-box attacks, CSA and IoU.
176
+
177
+ ## 4 Investigation
178
+
179
+ We conducted an analysis to determine how sensitive transformer trackers are to perturbations generated by existing attack methods under various conditions. We compared the difference in performance between the tracker's ability to provide accurate bounding boxes and binary masks by measuring the percentage difference from their original performance on clean data. We evaluated the impact of adversarial attacks on transformer trackers in predicting the object bounding boxes by varying the perturbation levels. Finally, we assessed the performance of the IoU attack when the generated perturbations were bounded at different noise levels. We then discussed the observations we drew from these sets of experiments.
180
+
181
+ ## 4.1 Adversarial Attacks Per Tracker Output
182
+
183
+ In this section, we have applied adversarial attack techniques against the TransT-SEG (Chen et al., 2023)
184
+ and MixFormerM (Cui et al., 2022) trackers and compared the results based on different tracking outputs.
185
+
186
+ The objective of this experiment is to determine the difference in each tracking metric before and after the attack when one of the tracker's outputs (bounding box or binary mask) is measured. Evaluation Protocol We selected the VOT2022 Short-term dataset and protocol (Kristan et al., 2023) because of these three reasons. Unlike other datasets that use the one-pass evaluation protocol, the VOT2022-
187
+ ST follows the anchor-based short-term protocol for the trackers evaluation. The metrics of VOT2022 baseline are obtained from the anchor-based short-term protocol which presents another view of the trackers' performances. Also, the VOT2022-ST provides the evaluation based on both object bounding boxes (STB) and binary masks (STS) which makes it a fair setup for our experiment. Therefore, the experiments for different attacks are achievable offline. Besides, the newest trackers are annually assessed by the VOT
188
+ community and the most robust trackers are listed and announced on different sub-challenges. For instance, the MixFormerM tracker (Cui et al., 2022) was among five top-ranked trackers for binary mask prediction and its other variant, MixFormerL won third place on the bounding box prediction sub-challenge. We have conducted a baseline experiment for the VOT2022 (Kristan et al., 2023) short-term sub-challenge in two cases: object bounding box (STB) vs. object masks (STS) for target annotation and tracking. The Expected Average Overlap (EAO), accuracy and the anchor-based robustness metrics are calculated in this experiment. The EAO computes the expected value of the prediction overlaps with the ground truth. The accuracy measures the average of overlaps between tracker prediction and the ground truth over a successful tracking period. The robustness is computed as the length of a successful tracking period over the length of the video sequence. The successful period is a period of tracking in which the overlap between the prediction and ground truth is always greater than the pre-defined threshold. The baseline metrics are computed based on the anchor-based protocol introduced in VOT2020 (Kristan et al., 2020). In every video sequence evaluation under this protocol, the evaluation toolkit will reinitialize the tracker from the next anchor of the Table 2: Evaluation results of the TransT-SEG (Chen et al., 2023) tracker attacked by different methods on the VOT2022 (Kristan et al., 2023) Short-Term (ST) dataset and protocol for two stacks: bounding box prediction via bounding box annotations (STB) and binary mask prediction via binary mask annotations
189
+ (STS). The "Clean" values are the original tracker performance without applying any attack.
190
+
191
+ | | EAO | | | Accuracy | Robustness | | | | | |
192
+ |-------|--------|-------|--------|------------|--------------|--------|--------|-------|--------|-------|
193
+ | Stack | Method | Clean | Attack | Drop | Clean | Attack | Drop | Clean | Attack | Drop |
194
+ | STB | CSA | 0.299 | 0.285 | 4.68% | 0.472 | 0.477 | -1.06% | 0.772 | 0.744 | 3.63% |
195
+ | IoU | 0.299 | 0.231 | 22.74% | 0.472 | 0.495 | -4.87% | 0.772 | 0.569 | 26.29% | |
196
+ | RTAA | 0.299 | 0.058 | 83.28% | 0.472 | 0.431 | 8.69% | 0.772 | 0.157 | 79.66% | |
197
+ | SPARK | 0.299 | 0.012 | 95.99% | 0.472 | 0.244 | 48.30% | 0.772 | 0.051 | 93.39% | |
198
+ | STS | CSA | 0.500 | 0.458 | 8.40% | 0.749 | 0.736 | 1.73% | 0.815 | 0.779 | 4.42% |
199
+ | IoU | 0.500 | 0.334 | 33.20% | 0.749 | 0.710 | 5.21% | 0.815 | 0.588 | 27.85% | |
200
+ | RTAA | 0.500 | 0.067 | 86.60% | 0.749 | 0.533 | 28.84% | 0.815 | 0.146 | 82.08% | |
201
+ | SPARK | 0.500 | 0.011 | 97.80% | 0.749 | 0.266 | 64.48% | 0.815 | 0.042 | 94.84% | |
202
+
203
+ data to compute the anchor-based metrics wherever the tracking failure happens. For visualization usage, we employed some video sequences from DAVIS2016 (Perazzi et al., 2016) dataset. Attacks Setting Four adversarial attacks are employed in this experiment, namely CSA (Yan et al., 2020), IoU (Jia et al., 2021), SPARK (Guo et al., 2020), and RTAA (Jia et al., 2020). However, not all of the attacks are applicable to both trackers. The SPARK and RTAA attacks manipulate the object candidates list, which includes classification labels and/or regression targets. If the tracker does not infer the candidates in the output, these attacks cannot be applied on, such as MixFormerM (Cui et al., 2022) that only outputs the predicted bounding box. In this experiment, we generated the perturbations using SPARK and RTAA attacks, i.e. white-box attacks, against trackers. However, the perturbation of CSA (Yan et al., 2020) is created by two GANs and passing the image into the SiamseRPN++ (Li et al., 2019) tracker to generate the adversarial loss depending on the SiamseRPN++ loss. Therefore, the CSA is a transferred black-box attack for both TransT-SEG (Chen et al., 2023) and MixFormer (Cui et al., 2022) trackers. The IoU method (Jia et al., 2021) is also a black-box approach that can perturb the whole frame using the tracker prediction for several times.
204
+
205
+ Results The following is a summary of the results obtained from an experiment conducted on the VOT2022 (Kristan et al., 2023) dataset using the TransT-SEG tracker (Chen et al., 2023) after adversarial attacks. The results are shown in Table 2 for both STB and STS cases. The most powerful attack against TransT-SEG (Chen et al., 2023) in all three metrics was found to be SPARK (Guo et al., 2020). According to the accuracy scores of Table 2, the object binary mask was more affected by the adversarial attacks than the object bounding box. However, the difference of drop percentages in assessing the predicted bounding boxes and binary masks are negligible for the EAO and Robustness metrics of TransT-SEG tracker.
206
+
207
+ It was observed that the CSA (Yan et al., 2020) attacks poorly degraded the outputs in evaluation metrics, except for the accuracy of the STB case. However, it was surprising to note that there was a negative difference in the accuracy metric after the CSA and IoU attack. Specifically, in the STB case, after the adversarial attacks, the accuracy of TransT-SEG (Chen et al., 2023) decreased by −1.06% for the CSA
208
+ attack and −4.87% for the IoU attack. This improvement was also observed for MixFormer (Cui et al.,
209
+ 2022) in Table 3 for EAO and robustness metrics after the CSA attack in the STB case and accuracy after the CSA in the STS case of the experiments. The binary masks predicted by the MixFormerM tracker after the IoU attack for all computed metrics dropped greater than the metrics computed for the predicted bounding boxes. Nonetheless, following the CSA attack, the drop percentages of metrics for MixFormerM
210
+ outputs were smaller and more consistent compared to those resulting from the IoU attack. Comparing the IoU attack results on both Tables 2 and 3 for bounding box evaluation (STB), the MixFormerM shows Table 3: Evaluation results of the MixFormerM (Cui et al., 2022) tracker attacked by different methods on the VOT2022 (Kristan et al., 2023) Short-Term (ST) dataset and protocol for two stacks: bounding box prediction via bounding box annotations (STB), binary mask prediction via binary mask annotations (STS).
211
+
212
+ The "Clean" values are the original tracker performance without applying any attack.
213
+
214
+ | | EAO | Accuracy | Robustness | | | | | | | |
215
+ |-------|--------|------------|--------------|--------|-------|--------|--------|-------|--------|--------|
216
+ | Stack | Method | Clean | Attack | Drop | Clean | Attack | Drop | Clean | Attack | Drop |
217
+ | STB | CSA | 0.303 | 0.308 | -1.65% | 0.479 | 0.478 | 0.21% | 0.780 | 0.791 | -1.41% |
218
+ | IoU | 0.303 | 0.246 | 18.81% | 0.479 | 0.458 | 4.38% | 0.780 | 0.665 | 14.74% | |
219
+ | STS | CSA | 0.589 | 0.562 | 4.58% | 0.798 | 0.803 | -0.63% | 0.880 | 0.857 | 2.61% |
220
+ | IoU | 0.589 | 0.359 | 39.05% | 0.798 | 0.660 | 17.30% | 0.880 | 0.677 | 23.07% | |
221
+
222
+ greater adversarial robustness with 14.74% drop percentage from original score whereas the robustness of TransT dropped by 26.29% after the same attack.
223
+
224
+ According to Table 3, the most powerful attack against MixFormerM (Cui et al., 2022) is the IoU attack (Jia et al., 2021). Even after the IoU attack (Jia et al., 2021), the adversarial perturbation slightly improves accuracy in assessing object binary masks (STS) and EAO in assessing predicted bounding boxes (STB). The EAO metric in the evaluation of the bounding box and binary mask is the most affected metric with 18.81% and 39.05% drop percentages after IoU attack. When compared to the corresponding metric for TransT-SEG (Chen et al., 2023) as shown in Table 2, the IoU attack had a greater damage on binary mask creation for MixFormerM than for TransT-SEG per EAO and accuracy metrics. However, this result was reversed for the same metrics when the object bounding box was evaluated in both trackers, Tables 2 and 3 (STB). For this point, it is important to mention that tracking and segmentation is performed by two different networks in MixFormerM (Cui et al., 2022). Therefore, the object bounding box evaluation is the assessment of the tracker's network while the binary mask evaluation is the assessment of the segmentation network (Yan et al., 2021). In contrast, the TransT-SEG (Chen et al., 2023) tracker performs both tracking and segmentation by a single transformer tracker.
225
+
226
+ Figure 1 demonstrates the results of different attacks on MixFormerM and TransT-SEG trackers in terms of the object bounding boxes and binary masks. The original outputs, bounding boxes and masks, are depicted by Green color, while the results after the attacks are exhibited with Red color. As the quantitative results indicated in Tables 2 and 3, the white-box attacks (SPARK and RTAA) have harmed the binary mask and bounding box more than the black-box attacks (IoU and CSA).
227
+
228
+ ## 4.2 Adversarial Attacks Per Perturbation Level
229
+
230
+ We test the effect of the perturbation levels on adversarial attack performance against transformer trackers. In white-box attacks such as SPARK (Guo et al., 2020) and RTAA (Jia et al., 2020), the generated perturbation is used to update the frame patch in each attack step. The overview of pseudocode of these two attacks is presented in Algorithms 1 and 2 where the ϕϵ indicates the operation of clipping the frame patch to the ϵ-ball. The α is the applied norm of gradients, and I is the search region. Although there are several differences in both settings, there is one similar step to generate the adversarial region from the input image gradient and previous perturbation(s). Line 4 of the RTAA pseudocode and line 6 of the SPARK pseudocode change the image pixels based on the computed gradients. By adjusting the ϵ-ball, the performance of attacks is evaluated to demonstrate the power of each adversarial idea for object trackers.
231
+
232
+ The plus or minus sign of the gradient sign corresponds to the decoupling direction and norm research in the gradient-based adversarial attack (Rony et al., 2019). For instance, SPARK (Guo et al., 2020) uses minus, while RTAA (Jia et al., 2020) sums up the sign of gradients with image values. Furthermore, note that in the original papers of SPARK (Guo et al., 2020) and RTAA (Jia et al., 2020), the attack parameters may have different names than those we used in this paper. Our goal was to unify their codes and approach into principle steps that make comparison more accessible for the audience. An important aspect of the
233
+
234
+ ![10_image_0.png](10_image_0.png)
235
+
236
+ MixFormerM TransT-SEG
237
+ Figure 1: Mask vs. bounding box predictions as the output of transformer trackers, MixFormerM (Cui et al., 2022) and TransT-SEG (Chen et al., 2023), while the adversarial attacks applied to perturb the input frame/search region. The TransT-SEG tracker's outputs harmed by the white-box methods, SPARK (Guo et al., 2020) and RTAA (Jia et al., 2020), more than black-box attacks, IoU (Jia et al., 2021) and CSA (Yan et al., 2020). The green mask/bounding box represents the object tracker's performance while the red mask/bound box belongs to the tracker's performance after each attack.
238
+
239
+ | Algorithm 1 RTAA (Jia et al., 2020) algorithm as the adversarial attack for object trackers 1: P ← P(t − 1) ▷ Initialize with perturbation map of previous fram | | e |
240
+ |-------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------|---------------------------------------------------------------|
241
+ | adv ← I | | ▷ Initialize with clean current frame |
242
+ | 2: I 3: for i = 1, . . . , imax do 4: I adv ← I adv + ϕ ϵ (P + α sign(∇I adv L)) | | ▷ Application of adversarial gradient descent |
243
+ | 5: | I adv ← max(0, min(I adv , 255)) | ▷ Clamp image values in [0, 255] |
244
+ | 6: | P ← I adv − I | ▷ Update perturbation map |
245
+ | 7: Return I adv ,P | | ▷ Return adversarial image and corresponding perturbation map |
246
+
247
+ SPARK algorithm, mentioned in (Guo et al., 2020), is its regularization term. This feature is convenient for maintaining sparse and imperceptible perturbations (Guo et al., 2020). The regularization term involves adding the L2,1 Norm of previous perturbations to the adversarial loss, which helps generate sparse and imperceptible noises. We generated examples of SPARK perturbation (Guo et al., 2020) versus RTAA (Jia et al., 2020) perturbations to verify this claim.
248
+
249
+ | Algorithm 2 SPARK (Guo et al., 2020) algorithm as the adversarial attack for object trackers 1: P ← P(t − 1) ▷ Initialize with perturbation map of previous fram | | e |
250
+ |--------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------------------------------------|---------------------------------------------------------------|
251
+ | 2: S ← PK i=1 P(t − i) | | ▷ Sum of perturbation maps of last K frames |
252
+ | 3: I adv ← I | | ▷ Initialize with clean current frame image |
253
+ | 4: for i = 1, . . . , imax do ′ ← I adv | | ▷ Get a copy of current adversarial image |
254
+ | 5: | I | |
255
+ | 6: | I adv ← I ′ + ϕ ϵ (P − α sign(∇I ′L)) + S | ▷ Application of adversarial gradient descent |
256
+ | 7: | I adv ← max(0, min(I adv , 255)) | ▷ Clamp image values in [0, 255] |
257
+ | 8: | P ← I adv − I ′ − S | ▷ Update perturbation map |
258
+ | 9: S ← PK i=1 P(t − i) | | ▷ Update the sum of perturbation maps |
259
+ | 10: I adv ← I + S | | ▷ Generate the current adversarial frame |
260
+ | 11: Return I adv ,P | | ▷ Return adversarial image and corresponding perturbation map |
261
+
262
+ Evaluation Protocol The test sets of the experiments on the perturbation level changes are the UAV123 dataset (Mueller et al., 2016) and VOTST2022 (Kristan et al., 2023). The UAV123 dataset comprises 123 video sequences with natural and synthetic frames in which an object appears and disappears from the frame captured by a moving camera. We calculate success and precision rates across various thresholds under the One Pass Evaluation (OPE) protocol. In this setup, the object tracker is initialized using the first frame and the corresponding bounding box. Subsequently, the tracker is evaluated for each frame's prediction for the rest of the video sequence. Precision is measured by calculating the distance between the center of the ground truth's bounding box and the predicted bounding box. The precision plot shows the percentage of bounding boxes that fall within a given threshold distance. The success rate is computed based on the Intersection over Union (IoU) between the ground truth and predicted bounding boxes. The success plot is generated by considering different thresholds over IoU and computing the percentage of bounding boxes that pass the given threshold. Furthermore, we computed the L1 norm and structural similarity (SSIM) (Wang et al., 2004) as the measurements of sparsity and imperceptibility of the generated perturbations per attack.
263
+
264
+ We chose some frames of the VOT2022ST (Kristan et al., 2023) dataset to visualize these metrics per frame.
265
+
266
+ Attacks Setting The SPARK (Guo et al., 2020) and RTAA (Jia et al., 2020) approaches applied on the TransT tracker (Chen et al., 2021) are assessed in this experiment using the OPE protocol. We chose the TransT tracker which is a pioneer on transformer trackers to observe the attack performance change on the perturbation levels. Both attacks generate the perturbed search region over a fixed number of iterations
267
+ (10). While the step size α for the gradient's update is 1 for RTAA and 0.3 for SPARK. We used five levels of perturbation ϵ ∈ {2.55, 5.1, 10.2, 20.4, 40.8} to compare its effects on the TransT (Chen et al., 2021)
268
+ performance on UAV123 (Mueller et al., 2016) and VOT2022ST (Kristan et al., 2023) datasets. The ϵ's are selected as a set of coefficients {0.01, 0.02, 0.04, 0.08, 0.16} of the maximum pixel value 255 in an RGB
269
+ image. It is worth mentioning that the ϵ for both attacks are set to 10 in their original settings. Therefore, the original performance of each attack is very close to the ϵ3 = 10.2 perturbation level.
270
+
271
+ Results Figure 2 shows the performance of TransT (Chen et al., 2021) under RTAA (Jia et al., 2020) and SPARK (Guo et al., 2020) attacks with different perturbation levels. The red curve indicates the clean performance of the tracker before applying any attacks. The other perturbation levels are demonstrated with different colors. Unlike classification networks with transformer backbones (Shao et al., 2022), the transformer tracker performances after the RTAA attack (Jia et al., 2020) using different ϵ's are minimally different but not after SPARK attacks (Guo et al., 2020). Adversarial perturbation methods against trackers use the previous perturbation to be added to the current frame. This setting may remove the sensitivity of the attack methods in the perturbation levels. In the RTAA attack (Jia et al., 2020), only one last perturbation is added to the current frame. In contrast, the SPARK (Guo et al., 2020) uses the previous perturbations in each time step for the last K = 30 frames, which reduces the sensitivity of the output to small changes in the inputs. For perturbation levels {ϵ3, ϵ4, ϵ5}, RTAA's performance (Jia et al., 2020)
272
+ remains the same, whereas using smaller levels affects its performance. It is noteworthy that RTAA (Jia et al., 2020) outperforms SPARK (Guo et al., 2020) on UAV datasets (Mueller et al., 2016) in almost every perturbation level except for the most minor level ϵ1 = 0.01 in which SPARK is the stronger attack.
273
+
274
+ In the main paper of SPARK (Guo et al., 2020), it has been mentioned that the technique generates a temporally sparse and imperceptible noise. Figure 3 displays various examples of perturbed search regions and perturbation maps produced by the TransT tracker after applying the SPARK attack. Upon applying the attack, we noted that some frames, like frame number "7" of the 'bubble' sequence and the first two rows of Figure 3, generated search regions and perturbation maps with fixed values for imperceptibility (SSIM
275
+ metric) and sparsity (L1 norm). Even by increasing the perturbation level, some frames retained the same level of imperceptibility and sparsity. However, there were also instances of super-perturbed search regions per video sequence, where the noise was noticeable, and the L1 norm had a high value, as shown in the last two rows of Figure 3. We consider a perturbed search region super-perturbed when the imperceptibility of the region is lower than 50%. The SPARK algorithm generates the most imperceptible noise with a constant and high SSIM value of 99.95% and sparse noise with L1 norm of 40.96 in all of the perturbation levels given the same frame. This fixed number of L1 norm is also the result of the regularization term discussed in the SPARK paper (Guo et al., 2020). This stability of SSIM and L1 norm have been repeated for many frames of the "bubble" sequence for SPARK attack (Guo et al., 2020) while in some frames, the imperceptibility and sparsity are not stable per perturbation levels.
276
+
277
+ In Figure 3, we indicate some super-perturbed regions with their perturbation maps per perturbation level. Interestingly, as we increase the perturbation levels, the number of super-perturbed regions also increases. In the attack settings, the perturbation of previous frames considered in the loss function is erased every 30
278
+
279
+ ![12_image_0.png](12_image_0.png)
280
+
281
+ Figure 2: The precision and success plots related to the TransT (Chen et al., 2021) performance after RTAA (Jia et al., 2020) (a, b) and SPARK (Guo et al., 2020) (c,d) attack under different levels of noise on UAV123 (Mueller et al., 2016) dataset. The average score for each metric is shown in the legend of the plots. The 'red' plot is the original TransT performance without any attack applied on the tracker. The e's are corresponded to ϵ's in our experiment, changing from e1 = 2.55 to e5 = 40.8 to assess the TransT performance after the white-box attacks under various perturbation levels. The SPARK performances per perturbation level shifts did not change on UAV123dataset as one can observe the SPARK curves are overlapped.
282
+ Table 4: The perturbation levels versus number of highly perturbed search regions generated by the SPARK
283
+ algorithm (Guo et al., 2020) applied on TransT (Chen et al., 2021) tracker. The SSIM and L1 norm are computed as the average number of highly perturbed regions on the "bubble" sequence of VOT2022 (Kristan et al., 2023) dataset. The "No. of frames" is the number of super-perturbed frames in which the SSIM value is below than 50%.
284
+
285
+ | ϵ | No. of frames | SSIM | L1 norm |
286
+ |------|-----------------|--------|-----------|
287
+ | 2.55 | 7 | 36.86 | 176.04 |
288
+ | 5.1 | 7 | 40.96 | 181.86 |
289
+ | 10.2 | 13 | 41.08 | 181.33 |
290
+ | 20.4 | 13 | 41.97 | 182.53 |
291
+ | 40.8 | 14 | 42.53 | 183.98 |
292
+
293
+ frames for RTAA (Jia et al., 2020) and SPARK (Guo et al., 2020) algorithms. Table 4 provides information about the number of highly perturbed frames during a video sequence and the average imperceptibility
294
+ (SSIM) and sparsity (L1 norm) scores. With higher levels of perturbations, more frames become highly perturbed, resulting in a greater L1 norm of the perturbations. Furthermore, in the lower perturbations levels, the highly perturbed search regions generate more perceptible noise, i.e. the imperceptibility of generated perturbations have grown by boosting the perturbation level. For the RTAA (Jia et al., 2020) attack applied on the TransT (Chen et al., 2021) tracker, whenever the perturbation level boosts the imperceptibility and sparsity declines. Figure 4 demonstrates the result of applying RTAA against TransT tracker for the same frame \#7 of the 'bubble' sequence. The RTAA attack perturbs search regions with higher SSIM values at the lowest ϵ level, i.e. the first level ϵ = 2.55. By increasing the perturbation levels, the perceptibly of the RTAA perturbation has been increased while the sparsity changes are small.
295
+
296
+ ## 4.3 Adversarial Attack Per Upper-Bound
297
+
298
+ We conducted an experiment to test the vulnerability of the IoU attack (Jia et al., 2021) to noise bounding, using different upper bounds. The IoU method (Jia et al., 2021) is a black-box attack that misleads trackers by adding various noises to the frame using object bounding boxes. The essential steps of the IoU attack (Jia et al., 2021) involve creating two levels of noise perturbations: orthogonal and normal direction noises. Our
299
+
300
+ ![13_image_0.png](13_image_0.png)
301
+
302
+ Figure 3: The search regions related to the "bubble" sequence in the VOT2022ST dataset (Kristan et al., 2023) after applying SPARK (Guo et al., 2020) attack on TransT (Chen et al., 2021) tracker. The perturbed search region is labeled with the SSIM (Wang et al., 2004) measured between search regions before and after the attack. The perturbation maps, following the work of (Yan et al., 2020), are created to demonstrate the added noise in colors. The L1 norm for perturbation maps are calculated to show the perturbation density/sparsity.
303
+
304
+ ![14_image_0.png](14_image_0.png)
305
+
306
+ Figure 4: The search regions related to the "bubble" sequence in the VOT2022ST dataset (Kristan et al.,
307
+ 2023) after applying RTAA (Jia et al., 2020) attack on TransT (Chen et al., 2021) tracker. The perturbed search region is labeled with the SSIM (Wang et al., 2004) measured between search regions before and after the attack. The perturbation maps, following the work of (Yan et al., 2020), are created to demonstrate the added noise in colors. The L1 norm for perturbation maps are calculated to show the perturbation density
308
+ (i.e. sparsity).
309
+ study aims to manipulate the attack settings in the second part of perturbation generation, which is in the normal direction. Dataset and Protocol The performances of the IoU attack are assessed against ROMTrack (Cai et al., 2023) on UAV123 (Mueller et al., 2016) dataset using the OPE protocol. The success rate, precision rate, and normalized precision rate are computed to compare the results. For this experiment, we report the average of the success rate called Area Under Curve (AUC), as well as the average precision and norm precision on the thresholds.
310
+
311
+ Attack Setting The IoU method (Jia et al., 2021) is a black-box attack on object trackers. It adds two types of noise to the frames: one in the tangential direction and the other in the normal direction. In the original setting, there was no limit on the number of times that noise could be added in the normal direction.
312
+
313
+ This noise was limited only by the upper bound ζ and the SIoU value in the original setting. The SIoU value is the weighted average of two computed IoU values: 1) the IoU of the noisy frame prediction and the first initialized frame I
314
+ adv 1in the attack algorithm, and 2) the IoU of the noisy frame prediction and the last frame bounding box in the tracking loop. In our experiment, we set a limit of 10 steps in the algorithm's last loop to reduce the processing time, especially for the larger upper bound ζ values. We tested the IoU attack under three upper bounds: ζ ∈ {8000, 10000, 12000}. The middle value of ζ = 10000 corresponds to the original setting of the IoU attack (Jia et al., 2021).
315
+
316
+ Results The images in Figure 5 display the results of the IoU attack (Jia et al., 2021) against ROMTrack (Cai et al., 2023) under various upper bounds for a single frame. The L1 norm of the perturbation has increased as the upper bounds were raised. Additionally, the imperceptibility, measured by the SSIM values, decreased as the perturbations became more severe. Since the IoU attack starts by generating some random noise, it is highly dependent on the initialization points. For some cases, the algorithm did not process a single video sequence even after 48 hours. One solution that worked for proceeding was to stop the processing without saving any results about the current sequence to restart the evaluation. After re-initialization, the attack began from another random point (noise) and it proceeded to the next sequence in less than 2 hours.
317
+
318
+ ![15_image_0.png](15_image_0.png)
319
+
320
+
321
+ Figure 5: The perturbed frames and perturbation maps generated by the IoU method (Jia et al., 2021)
322
+ against ROMTrack (Cai et al., 2023) using three upper bounds of ζ ∈ {8k, 10k, 12k}. The imperceptibility and L1 norm of the generated perturbations are shown in the frames representing the noise imperceptibility and sparsity of perturbation maps. Table 5: Evaluation results of the ROMTrack (Cai et al., 2023) attacked by the IoU approach (Jia et al.,
323
+ 2021) on the UAV123 (Mueller et al., 2016) dataset and protocol for three different upper bounds on the added noise in normal direction up to 10 processing steps.
324
+
325
+ | AUC | Precision | Norm Precision | | | | | | | |
326
+ |-------|-------------|------------------|-------|----------|--------|-------|----------|--------|-------|
327
+ | ζ | Original | Attack | Drop | Original | Attack | Drop | Original | Attack | Drop |
328
+ | 8k | 69.74 | 66.85 | 4.14% | 90.83 | 89.31 | 1.67% | 85.30 | 83.00 | 2.70% |
329
+ | 10k | 69.74 | 65.46 | 6.14% | 90.83 | 87.81 | 3.32% | 85.30 | 81.73 | 4.18% |
330
+ | 12k | 69.74 | 63.61 | 8.79% | 90.83 | 86.31 | 4.98% | 85.30 | 79.71 | 6.55% |
331
+
332
+ The results of the attack on ROMTrack (Cai et al., 2023) using the IoU method (Jia et al., 2021) with different upper bounds are presented in Table 5. It is clear that a higher upper bound leads to a more effective attack across all metrics. Despite the most substantial level of perturbation using the IoU method (Jia et al., 2021) resulting in an 8.79% decrease in the AUC metric, this outcome is insignificant. As shown in Figure 5, increasing ζ generates a perceptible perturbation with a lower SSIM to the original frame, resulting in a noisier frame that damages the tracking performance of ROMTrack (Cai et al., 2023) even more. However, the robust tracking performance is not affected more than 9% per metric, even in the highest perturbation level. In other words, ROMTrack (Cai et al., 2023) demonstrates good adversarial robustness against IoU
333
+ attack on UAV123 dataset.
334
+
335
+ ## 4.4 Transformer Versus Non-Transformer Trackers
336
+
337
+ In this experiment, we aim to study the adversarial robustness of trackers with different backbones especially transformer-based trackers compared to the non-transformer trackers. Evaluation Protocol This experiment is performed on the GOT10k dataset (Huang et al., 2019) which uses OPE protocol and the results are reported for three metrics: Average Overlap (AO), Success Rate SR0.5 with threshold 0.5 and Success Rate SR0.75 with threshold 0.75. The GOT10k test set contains 180 video sequences with axis aligned bounding box annotations. Attacks Setting We applied a set of attacks including two white-box attacks, RTAA (Jia et al., 2020), SPARK (Guo et al., 2020), and two black-box attacks, IoU (Jia et al., 2021) and CSA (Yan et al., 2020),
338
+ against a set of trackers including three transformer trackers (i.e., TransT (Chen et al., 2021), MixFormer (Cui et al., 2022) and ROMTrack (Cai et al., 2023)), two Siamese-based trackers (i.e., SiamRPN (Li et al., 2018)
339
+ and DaSiamRPN (Zhu et al., 2018)), and two discriminative trackers (i.e., DiMP (Bhat et al., 2019) and PrDiMP (Danelljan et al., 2020)). As it is discussed in Attack Setups Section 3.3 and Table 1, some attack methods are not applicable to some trackers due to the unavailability of the input for the attack algorithms.
340
+
341
+ The RTAA and SPARK are used as a form of white-box attack, while the IoU and CSA methods are black-box attacks. For MixFormer and ROMTrack trackers, the related pre-trained networks specialized for GOT10k dataset and their parameters are loaded and examined. For DaSiamesRPN tracker, the checkpoint released to test the OTB100 (Wu et al., 2015) dataset is used in this experiment. Results Table 6 indicates the vulnerability of the object trackers with different backbones, transformers vs.
342
+
343
+ non-transformers, against the adversarial attacks. The tracking performance before the attacks is indicated as the "No Attack" scores for each tracker, while the evaluation metrics after the attack are shown under the attack names. The "Drop(%)" columns in Table 6 represents the drop percentage of scores from the original tracker response per metrics, and it is calculated as 100 × [original − after attack]/original. The drop scores are calculated to demonstrate the attack performances in comparison to each other and to assess the tracker's performance after each attack compared to other trackers.
344
+
345
+ Beginning with the transformer trackers, the "No Attack" scores exhibit a relatively higher magnitude compared to non-transformer trackers. Within transformer trackers, the IoU and CSA methods are the mutual attacks, with the IoU method demonstrating a more significant damage compared to the CSA approach.
346
+
347
+ Specifically, among transformer trackers, TransT appears to be particularly susceptible to perturbations induced by the IoU attack. However, a greater decline in performance for MixFormer following the CSA
348
+ attack is observed in compared to other transformer trackers after the same attack, i.e. CSA attack.
349
+
350
+ All the test attack approaches could be applied to TransT, SiamRPN, and DaSiamRPN trackers due to the availability of the classification and regression labels in these trackers. Intriguingly, the RTAA method provides the most powerful attacking performance against SiamRPN and DaSiamRPN and TransT trackers.
351
+
352
+ Although the drop percentages of scores after SPARK attack is very close for Siamese-based trackers, this attack performance on TransT tracker is considerable. Repeated on for Siamese-based trackers, the IoU
353
+ attack is resulted in greater performance drop rather than the CSA attack.
354
+
355
+ Another intriguing observation is that MixFormer shows the smallest drop percentage after the IoU attack among transformer trackers. For example, the SR0.5 of MixFormer drops by 9.86%, whereas the percentage drop of the same metric is 18.31% for ROMTrack and 25.76% for TransT tracker. This indicates that
356
+
357
+ Table 6: The performance of transformer (Chen et al., 2023; Cui et al., 2022; Cai et al., 2023) and nontransformer trackers (Li et al., 2018; Zhu et al., 2018; Bhat et al., 2019; Danelljan et al., 2020) after white-box, SPARK (Guo et al., 2020) and RTAA (Jia et al., 2020), and black-box, IoU (Jia et al., 2021) and CSA (Yan et al., 2020), attacks on the GOT10k (Huang et al., 2019) dataset.
358
+
359
+ | Tracker | Attacker | Scores | Drop(%) | | | | |
360
+ |-----------|------------|----------|-----------|-------|--------|-------|-------|
361
+ | AO | SR0.5 | SR0.75 | AO | SR0.5 | SR0.75 | | |
362
+ | No Attack | 0.729 | 0.830 | 0.702 | - | - | - | |
363
+ | ROMTrack | CSA | 0.716 | 0.814 | 0.682 | 1.78 | 1.93 | 2.85 |
364
+ | IoU | 0.597 | 0.678 | 0.536 | 18.11 | 18.31 | 23.65 | |
365
+ | TransT | No Attack | 0.723 | 0.823 | 0.682 | - | - | - |
366
+ | CSA | 0.679 | 0.768 | 0.628 | 6.08 | 6.68 | 7.92 | |
367
+ | IoU | 0.532 | 0.611 | 0.432 | 26.42 | 25.76 | 36.66 | |
368
+ | SPARK | 0.137 | 0.085 | 0.032 | 81.05 | 89.67 | 95.31 | |
369
+ | RTAA | 0.048 | 0.019 | 0.011 | 93.36 | 97.69 | 98.39 | |
370
+ | No Attack | 0.696 | 0.791 | 0.656 | - | - | - | |
371
+ | MixFormer | CSA | 0.638 | 0.727 | 0.572 | 8.33 | 8.10 | 12.80 |
372
+ | IoU | 0.625 | 0.713 | 0.543 | 10.20 | 9.86 | 17.22 | |
373
+ | No Attack | 0.645 | 0.751 | 0.540 | - | - | - | |
374
+ | PrDiMP | IoU | 0.585 | 0.696 | 0.421 | 9.30 | 7.32 | 22.04 |
375
+ | No Attack | 0.602 | 0.717 | 0.463 | - | - | - | |
376
+ | DiMP | IoU | 0.549 | 0.653 | 0.372 | 8.80 | 8.93 | 19.65 |
377
+ | SiamRPN | No Attack | 0.406 | 0.499 | 0.101 | - | - | - |
378
+ | CSA | 0.382 | 0.460 | 0.083 | 5.91 | 7.81 | 17.82 | |
379
+ | IoU | 0.364 | 0.442 | 0.107 | 10.34 | 11.42 | -5.94 | |
380
+ | SPARK | 0.279 | 0.353 | 0.059 | 31.28 | 29.26 | 41.58 | |
381
+ | RTAA | 0.032 | 0.013 | 0.001 | 92.12 | 97.39 | 99.00 | |
382
+ | DaSiamRPN | No Attack | 0.389 | 0.465 | 0.090 | - | - | - |
383
+ | CSA | 0.314 | 0.353 | 0.061 | 19.28 | 24.09 | 32.22 | |
384
+ | IoU | 0.320 | 0.376 | 0.081 | 17.74 | 19.14 | 10 | |
385
+ | SPARK | 0.255 | 0.313 | 0.053 | 34.44 | 32.69 | 41.11 | |
386
+ | RTAA | 0.037 | 0.015 | 0.001 | 90.49 | 96.77 | 98.88 | |
387
+
388
+ although ROMTrack and TransT initially achieve better scores before the attacks, their performance is more greatly affected by IoU attack, resulting in larger drops in their evaluation metrics. Respectively, ROMTrack, TransT and MixFormer are ranked in terms of adversarial robustness after both IoU and CSA attack per AO and SR0.5 metrics among all trackers. However, for SR0.75 measurements, the ranking of adversarial robustness is MixFormer, ROMTrack and TransT after IoU attack. Conversely, the only applicable attack (from our attack methods) against discriminative trackers, DiMP and PrDiMP, is the black-box IoU method. A comparison between these two trackers reveals that PrDiMP exhibits greater robustness than DiMP. However, upon examining the evaluation scores, it's evident that PrDiMP experiences a more pronounced decline compared to DiMP in terms of percentage drop of SR0.75 metric. As the percentage drop related to SR0.75 for PrDiMP is 22.04%, while it is 19.65% for the same score of DiMP tracker. For the drop percentage of SR0.5 score, the PrDiMP, though, preserve its priority over DiMP tracker after the IoU attack.
389
+
390
+ The TransT, ROMTrack and DaSiamRPN are the top three in drop percentages of AO and SR0.5 scores after the IoU attack. The computed percentage drops indicates that although the transformer trackers demonstrate the highest scores before the attack, their scores after IoU attack fall more significantly than the non-transformer trackers in general. In some cases such as DaSiamRPN tracker, the percentage drop is also a big number per AO and SR0.5 metrics.
391
+
392
+ ## 5 Discussion
393
+
394
+ Our investigation began in Section 4.1 with an examination of the adversarial robustness of transformer trackers in producing object bounding boxes compared to object binary masks. Our findings indicated that the prediction of binary masks was more susceptible to adversarial perturbations than object bounding box predictions, particularly in terms of the accuracy metric. Additionally, we observed that white-box attacks, specifically SPARK and RTAA, exhibited greater efficacy compared to black-box attacks such as IoU and CSA, when targeting the TransT-SEG tracker. Notably, among the transformer-based trackers analyzed, MixFormerM, which employs deeper relation modeling than TransT-SEG, demonstrated superior adversarial robustness in terms of computed EAO, accuracy, and robustness on the VOT2022STS dataset against a single attack. Furthermore, we observed that MixFormerM is not susceptible to attacks like SPARK and RTAA with its own gradients due to the absence of necessary attack proxies, namely classification and regression labels.
395
+
396
+ In the subsequent experiment (c.f. Section 4.2), we demonstrated that the magnitude of perturbation shifts could influence or maintain the overall tracking outcomes, depending upon the attack strategy. An elevated level of perturbation consistently resulted in a higher count of super-perturbed search regions with increased L1 norm values for perturbations in both RTAA and SPARK. However, many perturbed search regions under the SPARK attack demonstrated the same values for SSIM and L1 norm metrics, indicating a stable imperceptibility and sparsity even with heightened levels of perturbation.
397
+
398
+ For the third experiment (c.f. Section 4.3), we evaluated the IoU attack across various upper bounds and discovered that, particularly in the context of black-box attacks involving random noises, the initialization point plays a crucial role. The IoU method becomes exceedingly time-consuming due to improper initialization. Higher upper bounds resulted in larger L1 norm values and smaller SSIM scores, suggesting less sparse and more perceptible noise in the perturbed frames.
399
+
400
+ In the concluding experiment (c.f. Section 4.4), we examined the impact of various backbones—transformerbased, discriminative-based and Siamese-based—on visual tracking before and after adversarial attacks.
401
+
402
+ Despite transformer trackers (ROMTrack, TransT, and MixFormer) showcasing the top-3 performance, their evaluation scores more notably decreased after applying the IoU method.
403
+
404
+ Another interesting finding is that the most effective attack may vary depending on the test set and tracker.
405
+
406
+ Indeed, for TransT-SEG, the strongest attack on the VOT2022 dataset was SPARK (Section 4.1), whereas RTAA outperformed SPARK on the GOT10k set on misleading TransT tracker's output (Section 4.4). For the SiamRPN tracker, the similar trend on the GOT10k set was observed: RTAA outperformed SPARK.
407
+
408
+ In our settings, the only applicable attacks against MixFormer and ROMTrack are black-box attacks, i.e.
409
+
410
+ IoU and CSA. Among black-box methods, the IoU attack outperformed CSA for TransT, MixFormer, MixFormerM and ROMTrack trackers, Sections 4.1 and 4.4. However, the effect of the IoU method against ROMTrack and MixFormer's are trivial, Section 4.4. The ROMTrack and MixFormer bounding box predictions were harmed by the IoU method up to 18.11 % and 18.81% on GOT10k and VOT2022 datasets for the average overlap metric, respectively. This indicates that these trackers were not being challenged enough with existing applicable attack methods. In addition to the aforementioned summary, our study also revealed the following observations:
411
+ - The generated perturbations for attacks that are applicable to transformer-based trackers have more impact on the object masks accuracy rather than on the accuracy of the bounding boxes on VOT2022ST
412
+ dataset (Section 4.1).
413
+
414
+ - Although it was demonstrated that adding previous perturbations to the current frame for perturbed search regions generation have an impact on the attacker performance (Guo et al., 2020), these previous perturbations result in more stable performance against changes in perturbation levels. For instance, such stability has been observed on SPARK (Guo et al., 2020) attack performance against TransT tracker on the UAV123 dataset (Mueller et al., 2016) (Section 4.2).
415
+
416
+ - The SPARK algorithm generates temporally sparse perturbations, meaning that the added perturbation to the search region is small for many frames. It results in imperceptible noise for those frames per video sequence, even though the perturbation level shifts to a higher value (Section 4.2).
417
+
418
+ - Increasing the perturbation level on SPARK results in more super-perturbed regions, i.e. regions with perceptible noise (Section 4.2).
419
+
420
+ - In IoU attack approach (Jia et al., 2020) and RTAA (Jia et al., 2021) attack, adding a higher perturbation level generates more perceptible noise for all frames, which damage more the overall tracking performance (Sections 4.2 and 4.3).
421
+
422
+ - The ranking of attack performance is sensitive to the experiment settings, dataset and protocol. For instance, SPARK method outperforms RTAA attack on VOT2022 for TransT-SEG tracker in Section 4.1, while RTAA scores are smaller than SPARK scores for TransT tracker on UAV123 dataset (Section 4.2).
423
+
424
+ Again we observed that RTAA outperforms SPARK approaches in attacking TransT, SiamRPN and DaSiamRPN trackers on GOT10k set (Section 4.4).
425
+
426
+ - The outcome of the IoU attack is sensitive to its initialization. The evaluation process may take a long time due to unsuitable initialization point of the attack (Section 4.3).
427
+
428
+ - Although transformer trackers, containing ROMTrack, MixFormer and TransT, exhibits more robust performance before and after the adversarial attacks, comparing the percentage drops from original scores reveals that these tracker's average overlap decreased from the original scores greater than almost all other trackers after the IoU attack (Section 4.4).
429
+
430
+ - Discriminative trackers also demonstrate a great adversarial robustness and ranked immediately after the transformer trackers on GOT10k dataset (Section 4.4).
431
+
432
+ ## 6 Conclusion
433
+
434
+ We conducted a study on adversarial attack methods for object trackers with the aim of testing their impact on transformer trackers and comparing their influences on visual trackers with different backbones. This paper includes several experiments on various tracking datasets, trackers, and attack settings. We evaluated three transformer trackers, ranging from light to deep relation modeling, and four non-transformer trackers with Siamese-based and discriminative backbones. Four attack methods, including two in white-box and two in black-box settings, were employed to assess the adversarial robustness of visual trackers. Our results showed that the accuracy of binary masks are more likely to harm by the adversarial attack in comparison to the accuracy of predicted bounding boxes. We also discovered that changes in the perturbation level do not necessarily affect the tracking performance over a tracking dataset. The sparsity and imperceptibility of the perturbations can be managed by advising a proper loss function. We also found that transformer trackers' performances after the adversarial attack drops from original performance greatly when compared to Siamese-based and discriminative trackers after the same attack. Our study indicates the need for further research on the adversarial robustness of transformer trackers since the existing attacks did not challenge these trackers significantly. The white-box attacks against Siamese-based trackers are not applicable to the transformer and discriminative trackers using their gradients due to the change of tracker backbone and architecture. One potential path for future work is the development of novel white-box attacks to target these kinds of trackers. The other direction can be on focusing more effective black-box attacks since they do not depend on the tracker backbones. The insightful findings in the adversarial robustness of classification networks with transformer backbones are also worthy to transfer on tracking networks with transformer backbones.
435
+
436
+ ## Acknowledgments
437
+
438
+ This work is supported by the DEEL Project CRDPJ 537462-18 funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) and the Consortium for Research and Innovation in Aerospace in Québec (CRIAQ), together with its industrial partners Thales Canada inc, Bell Textron Canada Limited, CAE inc and Bombardier inc. 1
439
+
440
+ ## References
441
+
442
+ Goutam Bhat, Martin Danelljan, Luc Van Gool, and Radu Timofte. Learning discriminative model prediction for tracking. In *IEEE Conf. on Comput. Vis.*, 2019.
443
+
444
+ Yidong Cai, Jie Liu, Jie Tang, and Gangshan Wu. Robust object modeling for visual tracking. In IEEE
445
+ Conf. on Comput. Vis., 2023.
446
+
447
+ Xin Chen, Bin Yan, Jiawen Zhu, Dong Wang, Xiaoyun Yang, and Huchuan Lu. Transformer tracking. In IEEE Conf. on Comput. Vis. and Pattern Recog., 2021.
448
+
449
+ Xin Chen, Bin Yan, Jiawen Zhu, Huchuan Lu, Xiang Ruan, and Dong Wang. High-performance transformer tracking. *IEEE Trans. on Pattern Analy. and Machine Intel.*, 45(7):8507–8523, 2023.
450
+
451
+ Yutao Cui, Cheng Jiang, Limin Wang, and Gangshan Wu. Mixformer: End-to-end tracking with iterative mixed attention. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2022.
452
+
453
+ Martin Danelljan, Goutam Bhat, Fahad Shahbaz Khan, and Michael Felsberg. ATOM: Accurate tracking by overlap maximization. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, pp. 4660–4669, 2019.
454
+
455
+ Martin Danelljan, Luc Van Gool, and Radu Timofte. Probabilistic regression for visual tracking. In IEEE
456
+ Conf. on Comput. Vis. and Pattern Recog., pp. 7183–7192, 2020.
457
+
458
+ Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In Int. Conf. Learn.
459
+
460
+ Represent., 2020.
461
+
462
+ Shenyuan Gao, Chunluan Zhou, Chao Ma, Xinggang Wang, and Junsong Yuan. Aiatrack: Attention in attention for transformer visual tracking. In *IEEE Conf. Euro. Conf. Comput. Vis.*, 2022.
463
+
464
+ Jindong Gu, Hengshuang Zhao, Volker Tresp, and Philip H. S. Torr. Segpgd: An effective and efficient adversarial attack for evaluating and boosting segmentation robustness. In *IEEE Conf. Euro. Conf.* Comput. Vis., 2022.
465
+
466
+ Qing Guo, Xiaofei Xie, Felix Juefei-Xu, Lei Ma, Zhongguo Li, Wanli Xue, Wei Feng, and Yang Liu. Spark:
467
+ Spatial-aware online incremental attack against visual tracking. In *IEEE Conf. Euro. Conf. Comput. Vis.*, 2020.
468
+
469
+ Lianghua Huang, Xin Zhao, and Kaiqi Huang. GOT-10k: A large high-diversity benchmark for generic object tracking in the wild. *IEEE Trans. on Pattern Analy. and Machine Intel.*, 43(5):1562–1577, 2019. ISSN 2835-8856.
470
+
471
+ 1https://deel.quebec Shuai Jia, Chao Ma, Yibing Song, and Xiaokang Yang. Robust tracking against adversarial attacks. In IEEE Conf. Euro. Conf. Comput. Vis., 2020.
472
+
473
+ Shuai Jia, Yibing Song, Chao Ma, and Xiaokang Yang. Iou attack: Towards temporally coherent black-box adversarial attack for visual object tracking. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2021.
474
+
475
+ Borui Jiang, Ruixuan Luo, Jiayuan Mao, Tete Xiao, and Yuning Jiang. Acquisition of localization confidence for accurate object detection. In *IEEE Conf. Euro. Conf. Comput. Vis.*, pp. 784–799, 2018.
476
+
477
+ Matej Kristan, Aleš Leonardis, and Jiří Matas. The eighth visual object tracking VOT2020 challenge results.
478
+
479
+ In *IEEE Conf. Euro. Conf. Comput. Vis. Worksh.*, 2020.
480
+
481
+ Matej Kristan, Aleš Leonardis, Jiří Matas, and Michael Felsberg. The tenth visual object tracking VOT2022 challenge results. In *IEEE Conf. Euro. Conf. Comput. Vis. Worksh.*, 2023.
482
+
483
+ Bo Li, Junjie Yan, Wei Wu, Zheng Zhu, and Xiaolin Hu. High performance visual tracking with siamese region proposal network. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2018.
484
+
485
+ Bo Li, Wei Wu, Qiang Wang, Fangyi Zhang, Junliang Xing, and Junjie Yan. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2019.
486
+
487
+ Tsung-Yi Lin, Piotr Dollar, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2017.
488
+
489
+ Kaleel Mahmood, Rigel Mahmood, and Marten van Dijk. On the robustness of vision transformers to adversarial examples. In *IEEE Conf. on Comput. Vis.*, 2021.
490
+
491
+ Matthias Mueller, Neil Smith, and Bernard Ghanem. A benchmark and simulator for uav tracking. In IEEE
492
+ Conf. Euro. Conf. Comput. Vis., 2016.
493
+
494
+ Federico Perazzi, Jordi Pont-Tuset, Brian McWilliams, Luc Van Gool, Markus Gross, and Alexander SorkineHornung. A benchmark dataset and evaluation methodology for video object segmentation. In *IEEE Conf.*
495
+ on Comput. Vis. and Pattern Recog., 2016.
496
+
497
+ Jerome Rony, Luiz G. Hafemann, Luiz S. Oliveira, Ismail Ben Ayed, Robert Sabourin, and Eric Granger.
498
+
499
+ Decoupling direction and norm for efficient gradient-based l2 adversarial attacks and defenses. In IEEE
500
+ Conf. on Comput. Vis. and Pattern Recog., 2019.
501
+
502
+ Rulin Shao, Zhouxing Shi, Jinfeng Yi, Pin-Yu Chen, and Cho-Jui Hsieh. On the adversarial robustness of vision transformers. *Trans. on Mach. Learn. Research*, 2022. ISSN 2835-8856.
503
+
504
+ Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In *Adv. Neur. Infor. Process. Syst.*, 2017.
505
+
506
+ Zhou Wang, A.C. Bovik, H.R. Sheikh, and E.P. Simoncelli. Image quality assessment: from error visibility to structural similarity. *IEEE Trans. on Image Process.*, 13(4):600–612, 2004.
507
+
508
+ Yi Wu, Jongwoo Lim, and Ming-Hsuan Yang. Object tracking benchmark. IEEE Trans. on Pattern Analy.
509
+
510
+ and Machine Intel., 37(9):1834–1848, 2015.
511
+
512
+ Bin Yan, Dong Wang, Huchuan Lu, and Xiaoyun Yang. Cooling-shrinking attack: Blinding the tracker with imperceptible noises. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2020.
513
+
514
+ Bin Yan, Xinyu Zhang, Dong Wang, Huchuan Lu, and Xiaoyun Yang. Alpha-refine: Boosting tracking performance by precise bounding box estimation. In *IEEE Conf. on Comput. Vis. and Pattern Recog.*, 2021.
515
+
516
+ Botao Ye, Hong Chang, Bingpeng Ma, Shiguang Shan, and Xilin Chen. Joint feature learning and relation modeling for tracking: A one-stream framework. In *IEEE Conf. Euro. Conf. Comput. Vis.*, 2022.
517
+
518
+ Xingyi Zhou, Dequan Wang, and Philipp Krähenbühl. Objects as points. *arXiv preprint arXiv:1904.07850*,
519
+ 2019.
520
+
521
+ Zheng Zhu, Qiang Wang, Bo Li, Wei Wu, Junjie Yan, and Weiming Hu. Distractor-aware siamese networks for visual object tracking. In *IEEE Conf. Euro. Conf. Comput. Vis.*, 2018.
FEEKR0Vl9s/FEEKR0Vl9s_meta.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "languages": null,
3
+ "filetype": "pdf",
4
+ "toc": [],
5
+ "pages": 23,
6
+ "ocr_stats": {
7
+ "ocr_pages": 0,
8
+ "ocr_failed": 0,
9
+ "ocr_success": 0,
10
+ "ocr_engine": "none"
11
+ },
12
+ "block_stats": {
13
+ "header_footer": 23,
14
+ "code": 0,
15
+ "table": 8,
16
+ "equations": {
17
+ "successful_ocr": 25,
18
+ "unsuccessful_ocr": 8,
19
+ "equations": 33
20
+ }
21
+ },
22
+ "postprocess_stats": {
23
+ "edit": {}
24
+ }
25
+ }