# mixformerv2_efficient_fully_transformer_tracking__d45bde63.pdf Mix Former V2: Efficient Fully Transformer Tracking Yutao Cui Tianhui Song Gangshan Wu Limin Wang State Key Laboratory for Novel Software Technology, Nanjing University, China https://github.com/MCG-NJU/Mix Former V2 Transformer-based trackers have achieved high accuracy on standard benchmarks. However, their efficiency remains an obstacle to practical deployment on both GPU and CPU platforms. In this paper, to mitigate this issue, we propose a fully transformer tracking framework based on the successful Mix Former tracker [14], coined as Mix Former V2, without any dense convolutional operation or complex score prediction module. We introduce four special prediction tokens and concatenate them with those from target template and search area. Then, we apply a unified transformer backbone on these mixed token sequence. These prediction tokens are able to capture the complex correlation between target template and search area via mixed attentions. Based on them, we can easily predict the tracking box and estimate its confidence score through simple MLP heads. To further improve the efficiency of Mix Former V2, we present a new distillation-based model reduction paradigm, including dense-to-sparse distillation and deep-to-shallow distillation. The former one aims to transfer knowledge from the dense-head based Mix Vi T to our fully transformer tracker, while the latter one is for pruning the backbone layers. We instantiate two Mix Foremr V2 trackers, where the Mix Former V2-B achieves an AUC of 70.6% on La SOT and AUC of 56.7% on TNL2k with a high GPU speed of 165 FPS, and the Mix Former V2-S surpasses FEAR-L by 2.7% AUC on La SOT with a real-time CPU speed. 1 Introduction Visual object tracking has been a fundamental and long-standing task in computer vision, which aims to locate the object in a video sequence, given its initial bounding box. It has a wide range of practical applications, which often require for low computational latency. So it is important to design a more efficient tracking architecture while maintaining high accuracy. Recently, the transformer-based one-stream trackers [7, 14, 55] attain excellent tracking accuracy than the previous Siamese-based ones [2, 10, 11], due to the unified modeling of feature extraction and target integration within a transformer block, which allows both components to benefit from the transformer development (e.g. Vi T [18], self-supervised pre-training [24] or contrastive pretraining [43]). However for these trackers, inference efficiency, especially on CPU, is still the main obstacle to practical deployment. Taking the state-of-the-art tracker Mix Vi T [15] as an instance, its pipeline contains i) transformer backbone on the token sequence from target template and search area, ii) dense corner head on the 2D search region for regression and iii) extra complex score prediction module for classification (i.e., estimating the box quality for reliable online samples selection). To achieve a high-efficiency tracker, there are still several issues on the design of Mix Vi T. First, the dense convolutional corner head still exhibits a time-consuming design, as implied in Tab 1. This is because it densely estimates the probability distribution of the box corners through a total of Equal contribution. Corresponding author (lmwang@nju.edu.cn). 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Layer Head Score GPU FPS GFLOPs 8 Pyram. Corner 90 27.2 8 Pyram. Corner - 120( 33.3%) 26.2 8 Token-based - 166( 84.4%) 22.5 Table 1: Efficiency analysis on Mix Vi T-B with different heads. Pyram. Corner represents for the pyramidal corner head [15]. Layer MLP Ratio Image Size CPU FPS GPU FPS 1 288 21 262 224 30 280 4 288 12 255 224 15 275 1 288 12 180 224 16 190 4 288 7 150 224 8 190 1 288 8 130 224 12 145 4 288 4 100 224 6 140 Table 2: Efficiency analysis on Mix Vi T-B with different backbone settings. The employed prediction head is plain corner head [14] for the analysis. 0 50 100 150 200 250 300 350 GPU Speed(fps) La SOT AUC(%) Mix Former V2-B Mix Former V2-S Sim Track-B Trans T STARK Light Track Figure 1: Comparison with state-of-the-art trackers in terms of AUC performance, model Flops and GPU Speed on La SOT. The circle diameter is in proportion to model flops. Mix Former V2-B surpasses existing trackers by a large margin in terms of both accuracy and inference speed. Mix Former V2-S achieves extremely high tracking speed of over 300 FPS while obtaining competitive accuracy compared with other efficient trackers [4, 5]. ten convolutional layers on the high-resolution 2D feature maps. Second, to deal with online template updating, an extra complex score prediction module composed of precise Ro I pooling layer, two attention blocks, and a three-layer MLP is required for improving online samples quality, which largely hinders its efficiency and simplicity of Mix Vi T. To avoid the dense corner head and complicated score prediction module, we propose a new fully transformer tracking framework Mix Former V2 without any dense convolutional operation. Our Mix Former V2 yields a very simple and efficient architecture, which is composed of a transformer backbone on the mixed token sequence and two simple MLP heads on the learnable prediction tokens. Specifically, we introduce four special learnable prediction tokens and concate them with the original tokens from target template and search area. Like the CLS token in standard Vi T, these prediction tokens are able to capture the complex relation between target template and search area, serving as a compact representation for subsequent regression and classification. Based on them, we can easily predict the target box and confidence score through simple MLP heads, which results in an efficient fully transformer tracker. Our MLP heads directly regress the probability distribution of four box coordinates, which improves the regression accuracy without increasing overhead. To further improve efficiency of Mix Former V2, we present a new model reduction paradigm based on distillation, including dense-to-sparse distillation and deep-to-shallow distillation. The denseto-sparse distillation aims to transfer knowledge from the dense-head based Mix Vi T, to our fully transformer tracker. Thanks to the distribution-based regression design in our MLP head, we can easily adopt logits mimicking strategy for distilling Mix Vi T trackers to our Mix Former V2. Based on the observation in Tab. 2, we also exploit the deep-to-shallow distillation to prune our Mix Former V2. We devise a new progressive depth pruning strategy by following a critical principle that constraining the initial distribution of student and teacher trackers to be as similar as possible, which can augment the capacity of transferring knowledge. Specifically, instructed by the frozen teacher model, some certain layers of a copied teacher model are progressively dropped and we use the pruned model as our student initialization. For CPU-realtime tracking, we further introduce an intermediate teacher model to bridge the gap between the large teacher and small student, and prune hidden dim of MLP based on the proposed distillation paradigm. Based on the proposed model reduction paradigm, we instantiate two types of Mix Former V2 trackers, Mix Former V2-B and Mix Former V2-S. As shown in Fig. 1, Mix Former V2 achieves better trade-off between tracking accuracy and inference speed than previous trackers. Especially, Mix Former V2-B achieves an AUC of 70.6% on La SOT with a high GPU speed of 165 FPS, and Mix Former V2-S outperforms FEAR-L by 2.7% AUC on La SOT with a real-time CPU speed. Our contributions are two-fold: 1) We propose the first fully transformer tracking framework without any convolution operation, dubbed as Mix Former V2, yielding a more unified and efficient tracker. 2) We present a new distillation-based model reduction paradigm to make Mix Former V2 more effective and efficient, which can achieve high-performance tracking on platforms with GPUs or CPUs. 2 Related Work Efficient Visual Object Tracking. In recent decades, the visual object tracking task has witnessed rapid development due to the emergence of new benchmark datasets[20, 28, 41, 42, 48] and better trackers [2, 10, 12 14, 32, 52, 55]. Researchers have tried to explore efficient and effective tracking architectures for practical applications, such as siamese-based trackers [2, 31, 32, 50], online trackers [3, 16] and transformer-based trackers [10, 37, 52]. Benefiting from transformer structure and attention mechanism, recent works [7, 14, 55] on visual tracking are gradually abandoning traditional three-stage model paradigm, i.e., feature extraction, information interaction and location head. They adopted a more unified one-stream model structure to jointly perform feature extraction and interaction, which turned out to be effective for modeling visual object tracking task. However, some modern tracking architectures are too heavy and computational expensive, making it hard to deploy in practical applications. Light Track [53] employed NAS to search the a light Siamese network, but its speed was not extremely fast on powerful GPUs. FEAR [5], HCAT [9], E.T.Track [4] designed more efficient framework, however were not suitable for one-stream trackers. We are the first to design efficient one-stream tracker so as to achieve good accuracy and speed trade-off. Knowledge Distillation. Knowledge Distillation [27] was proposed to learn more effective student models with teacher model s supervision. In the beginning, KD is applied in classification problem, where KL divergence is used for measuring the similarity of teacher s and student s predicted probability distribution. For regression problem like object detection, feature mimicking [1, 23, 33] is frequently employed. LD [57] operate logits distillation on bounding box location by converting Dirac delta distribution representation to probability distribution representation of bounding box, which well unifies logits distillation and location distillation. In this work, we exploit some customized strategies to make knowledge distillation more suitable for our tracking framework. Vision Transformer Compression. There exist many general techniques for the purpose of speeding up model inference, including model quantization [22, 47], knowledge distillation [27, 36], pruning [25], and neural architecture search [19]. Recently many works also focus on compressing vision transformer models. For example, Dynamic Vi T [44], Evo-Vi T[51] tried to prune tokens in attention mechanism. Auto Former [8], NASVi T [21], Slimming Vi T [6] employed NAS technique to explore delicate Vi T architecture. Vi TKD [54] provided several Vi T feature distillation guidelines but it focused on compressing the feature dimension instead of model depth. Mini Vi T [56] applied weights sharing and multiplexing to reduce model parameters. Since one-stream trackers highly rely on training-resource-consuming pre-training, we resort to directly prune the layers of our tracker. In this section, we first present the Mix Former V2, which is a more efficient and unified fully transformer tracking framework. Then we describe the proposed distillation-based model reduction, including dense-to-sparse distillation and deep-to-shallow distillation. 3.1 Fully Transformer Tracking: Mix Former V2 The proposed Mix Former V2 is a fully transformer tracking framework without any convolutional operation and complex score prediction module. Its backbone is a plain transformer on the mixed token sequence of three types: target template token, search area token, and learnable prediction token. Then, simple MLP heads are placed on top for predicting probability distribution of the box coordinates and corresponding target quality score. Compared with other transformer-based trackers (e.g. Trans T [10], STARK [52], Mix Former [14], OSTrack [55] and Sim Track [7]), our Mix Former V2 streamlines the tracking pipeline by effectively removing the customized convolutional classification and regression heads for the first time, which yields a more unified, efficient and flexible tracker. The overall architecture is depicted in Fig. 2. With inputting the template tokens, the search area tokens and learnable prediction tokens, Mix Former V2 predicts the target bounding boxes and quality score in an end-to-end manner. Search Area Special Learnable Prediction Tokens Distribution-based Localization Head Predicted Coordinates Distribution P-Mixed Attention Module 𝑵 Pred. Score Figure 2: Mix Former V2 Framework. Mix Former V2 is a fully transformer tracking framework, composed of a transformer backbone and two simple MLP heads on the learnable prediction tokens. Prediction-Token-Involved Mixed Attention. Compared to original slimming mixed attention [15] in Mix Vi T, the key difference lies in the introduction of the special learnable prediction tokens, which are used to capture the correlation between the target template and search area. These prediction tokens can progressively compress the target information and used as a compact representations for subsequent regression and classification. Specifically, given the concatenated tokens of multiple templates, search and four learnable prediction tokens, we pass them into N layers of predictiontoken-involved mixed attention modules (P-MAM). We use qt, kt and vt to represent template elements (i.e. query, key and value) of attention, qs, ks and vs to represent search region, qe, ke and ve to represent learnable prediction tokens. The P-MAM can be defined as: ktse = Concat(kt, ks, ke), vtse = Concat(vt, vs, ve), Attent = Softmax(qtk T t d )vt, Attens = Softmax(qsk T tse d )vtse, Attene = Softmax(qek T tse (1) where d represents the dimension of each elements, Attent, Attents and Attene are the attention output of the template, search and the learnable prediction tokens respectively. Similar to the original Mix Former, we use the asymmetric mixed attention scheme for efficient online inference. Like the CLS tokens in standard Vi T, the learnable prediction tokens automatically learn on the tracking dataset to compress both the template and search information. Direct Prediction Based on Tokens. After the transformer backbone, we directly use the prediction tokens to regress the target location and estimate its reliable score. Specifically, we exploit the distribution-based regression based on the four special learnable prediction tokens. In this sense, we regress the probability distribution of the four bounding box coordinates rather than their absolute positions. Experimental results in Section 4.2 also validate the effectiveness of this design. As the prediction tokens can compress target-aware information via the prediction-token-involved mixed attention modules, we can simply predict the four box coordinates with a same MLP head as follows: ˆPX(x) = MLP(token X), X {T , L, B, R}. (2) In implementation, we share the MLP weights among four prediction tokens. For predicted target quality assessment, the Score Head is a simple MLP composed of two linear layers. Specifically, firstly we average these four prediction tokens to gather the target information, and then feed the token into the MLP-based Score Head to directly predict the confidence score s which is a real number. Formally, we can represent it as: s = MLP (mean (token X)) , X {T , L, B, R}. These token-based heads largely reduces the complexity for both the box estimation and quality score estimation, which leads to a more simple and unified tracking architecture. Coordinates Logits Distillation Progressive Depth Pruning Process Search Area Eliminating Feature Mimicking KL Divergence Template Token Search Token Special Prediction Token Corner Probability Distribution Coordinates Probability Distribution Coordinates Probability Distribution Eliminating Block Figure 3: Distillation-Based Model Reduction for Mix Former V2. The Stage1 represents for the dense-to-sparse distillation, while the Stage2 is the deep-to-shallow distillation. The blocks with orange arrows are to be supervised and blocks with dotted line are to be eliminated. Feed Forward Eliminating Multi-Head Eliminating Feed Forward Input Tokens Figure 4: Progressive Depth Pruning Process for eliminating blocks. All weights in this block decay to zeros and finally only residual connection works, turning into an identity block. 3.2 Distillation-Based Model Reduction To further improve the efficiency and effectiveness of Mix Former V2, we present a distillation-based model reduction paradigm as shown in Fig. 3, which first perform dense-to-sparse distillation for better token-based prediction and then deep-to-shallow distillation for the model pruning. 3.2.1 Dense-to-Sparse Distillation In Mix Former V2, we directly regress the target bounding box based on the prediction tokens to the distribution of four random variables T , L, B, R R, which represents the box s top, left, bottom and right coordinate respectively. In detail, we predict the probability density function of each coordinate: X ˆPX(x), where X {T , L, B, R}. The final bounding box coordinates B can be derived from the expectation over the regressed probability distribution: BX = E ˆ PX[X] = Z R x ˆPX(x)dx. (3) Since the original Mix Vi T s dense convolutional corner heads predict two-dimensional probability maps, i.e. the joint distribution PT L(x, y) and PBR(x, y) for top-left and bottom-right corners, the one-dimensional version of box coordinates distribution can be deduced easily through marginal distribution: PT (x) = Z R PT L(x, y)dy, PL(y) = Z R PT L(x, y)dx R PBR(x, y)dy, PR(y) = Z R PBR(x, y)dx. (4) Herein, this modeling approach can bridge the gap between the dense corner prediction and our sparse token-based prediction, i.e., and the regression outputs of original Mix Vi T can be regarded as soft labels for dense-to-sparse distillation. Specifically, we use Mix Vi T s outputs PX as in Equation 4 for supervising the four coordinates estimation ˆPX of Mix Former V2, applying KL-Divergence loss as follows: Lloc = X X {T ,L,B,R} LKL( ˆPX, PX). (5) In this way, the localization knowledge is transferred from the dense corner head of Mix Vi T to the sparse token-based head of Mix Former V2. 3.2.2 Deep-to-Shallow Distillation For further improving efficiency, we focus on pruning the transformer backbone. However, designing a new light-weight backbone is not suitable for fast one-stream tracking. A new backbone of onestream trackers often highly relies on large-scale pre-training to achieve good performance, which requires for huge amounts of computation. Therefore, we resort to directly cut down some layers of Mix Former V2 backbone based on both the feature mimicking and logits distillation, as can be seen in Fig. 3:Stage2. Let F S i , F T j Rh w c denote the feature map from student and teacher, the subscript represents the index of layers. For logits distillation, we use KL-Divergence loss. For feature imitation, we apply L2 loss: (i,j) M L2(F S i , F T j ), (6) where M is the set of matched layer pairs need to be supervised. Specifically, we design a progressive model depth pruning strategy for distillation. Progressive Model Depth Pruning. Progressive Model Depth Pruning aims to compress Mix Former V2 backbone through reducing the number of transformer layers. Since directly removing some layers could lead to inconsistency and discontinuity, we explore a progressive method for model depth pruning based on the feature and logits distillation. Specifically, instead of letting teacher to supervise a smaller student model from scratch, we make the original student model a complete copy of the teacher model. Then, we will progressively eliminate certain layers of student and make the remaining layers to mimic teacher s representation during training with supervision of teacher. This design allows the initial representation of student and teacher to keep as consistent as possible, providing a smooth transition scheme and reducing the difficulty of feature mimicking. Formally, let xi denote output of the i-th layer of Mix Former V2 backbone, the calculation of attention block can be represented as below (Layer-Normalization operation is omitted in equation): x i = ATTN(xi 1) + xi 1, xi = FFN(x i) + x i = FFN(ATTN(xi 1) + xi 1) + ATTN(xi 1) + xi 1, (7) Let E be the set of layers to be eliminated in our student network, we apply a decay rate γ on the weights of these layers: xi = γ(FFN(ATTN(xi 1) + xi 1) + ATTN(xi 1)) + xi 1, i E. (8) During the first m epochs of student network training, γ will gradually decrease from 1 to 0 in the manner of cosine function: 0.5 1 + cos t 0, t > m. (9) This means these layers in student network are gradually eliminated and finally turn into identity transformation, as depicted in Figure 4. The pruned student model can be obtained by simply removing layers in E and keeping the remaining blocks. Intermediate Teacher. For distillation of an extremely shallow model (4-layers Mix Former V2), we introduce an intermediate teacher (8-layers Mix Former V2) for bridging the deep teacher (12-layers Mix Former V2) and the shallow one. Typically, the knowledge of teacher may be too complex for a small student model to learn. So we introduce an intermediate role serving as teaching assistant to relieve the difficulty of the extreme knowledge distillation. In this sense, we divide the problem of knowledge distillation between teacher and small student into several distillation sub-problems. MLP Reduction. As shown in Table 2, one key factor affecting the inference latency of tracker on CPU device is the hidden feature dim of MLP in Transformer block. In other words, it becomes the bottleneck that limits the real-time speed on CPU device. In order to leverage this issue, we further prune the hidden dim of MLP based on the proposed distillation paradigm, i.e., feature mimicking and logits distillation. Specifically, let the shape of linear weights in the original model is w Rd1 d2, and the corresponding shape in the pruning student model is w Rd 1 d 2, in which d 1 d1, d 2 d2, we will initialize weights for student model as: w = w[: d 1, : d 2]. Then we apply distillation supervision for training, letting the pruned MLP to simulate original heavy MLP. 3.3 Training of Mix Former V2 The overall training pipeline is demonstrated in Fig. 3, performing dense-to-sparse distillation and then deep-to-shallow distillation to yield our final efficient Mix Former V2 tracker. Then, we train the MLP based score head for 50 epochs. Particularly, for CPU real-time tracking, we employ the intermediate teacher to generate a shallower model (4-layer Mix Former V2) based on the proposed distillation. Besides, we also use the designed MLP reduction strategy for further pruning the CPU real-time tracker. The total loss of distillation training with student S and teacher T is calculated as: L = λ1L1(BS, Bgt) + λ2Lciou(BS, Bgt) + λ3Ldist(S, T), (10) where the first two terms are exactly the same as original Mix Former s location loss supervised by ground truth bounding box labels, and the rest term is for aforementioned distillation. 4 Experiments 4.1 Implemented Details Training and Inference. Our trackers are implemented using Python 3.6 and Py Torch 1.7. The distillation training is conducted on 8 NVidia Quadro RTX 8000 GPUs. The inference process runs on one NVidia Quadro RTX 8000 GPU and Intel(R) Xeon(R) Gold 6230R CPU @ 2.10GHz. The training datasets includes Tracking Net [42], La SOT [20], GOT-10k [28] and COCO [35] training splits., which are the same as Mix Former [14]. Each distillation training stage takes 500 epochs, where the first m = 40 epochs are for progressively eliminating layers. We train the score prediction MLP for additional 50 epochs. The batch size is 256, each GPU holding 32 samples. We use Adam W optimizer with weight decay of 10 4. The initial learning rate is 10 4 and will be decreased to 10 5 after 400 epochs. We use horizontal flip and brightness jittering for data augmentation. We instantiate two types of Mix Former V2, including Mix Former V2-B of 8 P-MAM layers for high-speed tracking on GPU platform and Mix Former V2-S of 4 P-MAM layers with MLP ratio of 1.0 for real-time tracking on CPU platform. Their numbers of parameters are 58.8M and 16.2M respectively. The resolutions of search and template images for Mix Former V2-B are 288 288 and 128 128 respectively. While for Mix Former V2-S, the resolutions of search and template images are 224 224 and 112 112 for real-time tracking on CPU platform. The inference pipeline is the same as Mix Former [14]. We use the first template together with the current search region as input of Mix Former V2. The dynamic templates are updated when the update interval of 200 is reached by default, where the template with the highest score is selected as an online sample. Distillation-Based Reduction. For dense-to-sparse distillation, we use Mix Vi T-L as teacher for training Mix Former V2-B by default. We also try to use Mix Vi T-B as the teacher in Tab 5. Particularly, we employ a customized Mix Vi T-B of plain corner head and with search input size of 224 224 as the teacher for Mix Former-S. For deep-to-shallow distillation, we use the progressive model depth pruning strategy to produce the 8-layer Mix Former V2-B from a 12-layer one. For Mix Former V2-S, we additionally employs intermediate teacher and MLP reduction strategies, and the process is 12-layers Mix Former V2 to 8-layers Mix Former V2, then 8-layers Mix Former V2 to 4-layers Mix Former V2, finally 4-layers MLP-ratio-4.0 Mix Former V2 to 4-layers MLP-ratio-1.0 Mix Former V2-S . 4.2 Exploration Studies To verify the effectiveness of our proposed framework and training paradigm, we analyze different components of Mix Former V2 and perform detailed exploration studies on La SOT [20] dataset. 4.2.1 Analysis on Mix Former V2 Framework Token-based Distribution Regression. The design of distribution-based regression with special learnable prediction tokens is the core of our Mix Former V2. We conduct experiments on different regression methods in Tab. 3a. All models employ Vi T-B as backbone and are deployed without distillation and online score prediction. Although the pyramidal corner head obtains the best performance, the running speed is largely decreased compared with our token-based regression head in Mix Former V2. Mix Former V2 with four prediction tokens achieves good trade-off between performance and inference latency. Besides, compared to the direct box prediction with one token on the Type Layer AUC FPS T1 12 63.1% 112 T4 12 67.5% 110 Py-Corner. 12 69.0% 92 (a) Different regression methods. T1 denotes direct box prediction based on one token, T4 is the proposed distribution-based prediction with 4 prediction tokens, and Py Conrer. is the pyramidal corner head as in Mix Vi T. Models are without distillation and score prediction. Method Score. AUC FPS Ours - 68.9% 166 Ours 70.6% 165 Mix Vi T-B - 69.0% 92 Mix Vi T-B 69.6% 80 (b) Quality score prediction. In Mix Former V2, we use token-based MLP head for sample quality score prediction. While in Mix Vi T, it use an extra SPM for score prediction. We employ Mix Former V2-B of 8 layers, with the Mix Vi T-L as the distillation teacher, for this analysis. Stu. Tea. Tea-AUC AUC base - - 67.5% base base 69.0% 68.9% base large 71.5% 69.6% (c) Dense-to-Sparse distillation. The base student denotes the 12-layers Mix Former V2 framework without score prediction. The base teacher is the Mix Vi T-B and the large teacher is the Mix Vi T-L. Tea AUC is the AUC of the teacher. Models are without score prediction. Log-dis. Feat-mim. AUC - - 60.7% - 62.4% 62.9% (d) Feature mimicking & logits distillation. For distillation analysis, we use the Mix Vi T-B of 12 layers with corner head as the teacher, and Mix Vi T of 4 layers as the student. Init. method AUC MAE-fir4 62.9% Tea-skip4 64.4% PMDP 64.8% (e) Progressive model depth pruning (PMDP). MAE-fir4 denotes using first 4 layers of MAE-B for student initialization. Tea-skip4 is using 4 skipped layers of the teacher. Inter. teacher AUC - 64.8% 65.5% (f) Intermediate teacher. For the analysis, we use the 12-layers Mix Vi T-B as the teacher, 8-layers Mix Vi T as the intermediate teacher and 4-layers Mix Vi T as the student. Epoch m AUC 30 68.3% 40 68.5% 50 68.5% (g) Eliminating Epochs. Epoch m indicates the number of epochs in progressive eliminating process. The model architecture is based on Mix Former V2-B. Models are without score prediction. blocks num. head AUC 12 Py-Corner. 69.0% 12 T4 68.9% 8 T4 68.5% (h) Model pruning route of Mix Former V2-B . T4 denotes the proposed distribution-based prediction with 4 prediction tokens. We use the Mix Vi T-B as the distillation teacher for this analysis. blocks num. head MLP-r AUC 12 Cor. 4 68.2% 12 T4 4 67.7% 8 T4 4 66.6% 4 T4 4 61.0% 4 T4 1 59.4% (i) Model pruning route of Mix Former V2-S. Cor. represents for the plain corner head, which is used in the initial teacher model. MLP-r denotes the MLP ratio in attention blocks. Table 3: Ablation studies on La SOT.The default choice for our model is colored in gray . first line of Tab. 3a, which estimates the absolute target position instead of the probability distribution of four coordinates, the proposed distribution-based regression obtains better accuracy. Besides, this design allows to perform dense-to-sparse distillation so as to further boost performance. Token-based Quality Score Prediction. The design of the prediction tokens also allows to perform more efficient quality score prediction via a simple MLP head. As shown in Tab. 3b, the token-based score prediction component improves the baseline Mix Former V2-B by 1.7% with increasing quite little inference latency. Compared to ours, the score prediction module in Mix Vi T-B further decreases the running speed by 13.0%, which is inefficient. Besides, the SPM in Mix Vi T requires precise Ro I pooling, which hinders the migration to various platforms. 4.2.2 Analysis on Dense-to-Sparse Distillation We verify the effectiveness of dense-to-sparse distillation in Tab. 3c. When use Mix Vi T-B without its SPM (69.0% AUC) as the teacher model, the Mix Former V2 of 12 P-MAM layers achieves an AUC score of 68.9%, increasing the baseline by 1.4%. This further demonstrate the significance of the design of four special prediction tokens, which allows to perform dense-to-sparse distillation. The setting of using Mix Vi T-L (71.5% AUC) as the teacher model increases the baseline by an AUC score of 2.2%, which implies the good distillation capacity of the large model. KCF Siam FC ATOM D3Sv2 Di MP To MP Trans T SBT Swin Track Ours-S Ours-B [26] [2] [16] [38] [3] [39] [10] [49] [34] EAO 0.239 0.255 0.386 0.356 0.430 0.511 0.512 0.522 0.524 0.431 0.556 Accuracy 0.542 0.562 0.668 0.521 0.689 0.752 0.781 0.791 0.788 0.715 0.795 Robustness 0.532 0.543 0.716 0.811 0.760 0.818 0.800 0.813 0.803 0.757 0.851 Table 4: State-of-the-art comparison on VOT2022 [30]. The best results are shown in bold font. 4.2.3 Analysis on Deep-to-Shallow Distillation In the following analysis on deep-to-shallow distillation, we use the Mix Vi T-B of 12 layers with plain corner head as the teacher, and Mix Vi T of 4 layers with the same corner head as the student. The models are deployed without score prediction module. Feature Mimicking & Logits Distillation. To give detailed analysis on different distillation methods for tracking, we conduct experiments in Tab. 3d. The models are all initialized with the first 4-layers MAE pre-trained Vi T-B weights. It can be seen that logits distillation can increase the baseline by 1.7% AUC, and adding feature mimicking further improves by 0.4% AUC, which indicates the effectiveness of both feature mimicking and logits distillation for tracking. Progressive Model Depth Pruning. We study the effectiveness of the progressive model depth pruning (PMDP) for the student initialization in Tab. 8b. It can be observed that the PMDP improves the traditional initialization method of using MAE pre-trained first 4-layers Vi T-B by 1.9%. This demonstrates that it is critical for constraining the initial distribution of student and teacher trackers to be as similar as possible, which can make the feature mimicking easier. Surprisingly, we find that even the initial weights of the four layers are not continuous, i.e., using the skipped layers (the 3,6,9,12-th) of the teacher for initialization, the performance is better than the baseline (62.9% vs. 64.4%), which further verifies the importance of representation similarity between the two ones. Determination of Eliminating Epochs. We conduct experiments as shown in the Table 3g to choose the best number of epochs m in the progressive eliminating period. We find that when the epoch m greater than 40, the choice of m seems hardly affect the performance. Accordingly we determine the epoch to be 40. Intermediate Teacher. Intermediate teacher is introduced to promote the transferring capacity from a deep model to a shallow one. We conduct experiment as in Table 3f. We can observe that the intermediate teacher can bring a gain of 0.7% AUC score which can verify that. 4.2.4 Model Pruning Route We present the model pruning route from the teacher model to Mix Former V2-B and Mix Former V2-S in Tab. 3h and Tab. 3i respectively. The models on the first line are corresponding teacher models. We can see that, through the dense-to-sparse distillation, our token-based Mix Former V2-B obtains comparable accuracy with the dense-corner-based Mix Vi T-B with higher running speed. Through the progressive model depth pruning based on the feature and logits distillation, Mix Former V2-B with 8 layers only decreases little accuracy compared to the 12-layers one. 4.3 Comparison with the Previous Methods Comparison with State-of-the-art Trackers. We evaluate the performance of our proposed trackers on 6 benchmark datasets: including the large-scale La SOT [20], La SOText [20], Tracking Net [42], UAV123 [41], TNL2K [48] and VOT2022 [30]. La SOT is a large-scale dataset with 1400 long videos in total and its test set contains 280 sequences. Tracking Net provides over 30K videos with more than 14 million dense bounding box annotations. UAV123 is a large dataset containing 123 aerial videos which is captured from low-altitude UAVs. VOT2022 benchmark has 60 sequences, which measures the Expected Average Overlap (EAO), Accuracy (A) and Robustness (R) metrics. Among them, La SOText and TNL2K are two relatively recent benchmarks. La SOText is a released extension of La SOT, which consists of 150 extra videos from 15 object classes. TNL2K consists of 2000 sequences, with natural language description for each. We evaluate our Mix Former V2 on the test set with 700 videos. The results are presented in Tab. 4 and Tab. 5. More results on other datasets will be present in supplementary materials. Only the trackers of similar complexity are Method La SOT La SOText TNL2K Tracking Net UAV123 Speed AUC PNorm P AUC P AUC P AUC PNorm P AUC P GPU Mix Former V2-B 70.6 80.8 76.2 50.6 56.9 57.4 58.4 83.4 88.1 81.6 69.9 92.1 165 Mix Former V2-B 69.5 79.1 75.0 - - 56.6 57.1 82.9 87.6 81.0 70.5 91.9 165 Mix Former [14] 69.2 78.7 74.7 - - - - 83.1 88.1 81.6 70.4 91.8 25 CTTrack-B [45] 67.8 77.8 74.0 - - - - 82.5 87.1 80.3 68.8 89.5 40 OSTrack-256 [55] 69.1 78.7 75.2 47.4 53.3 54.3 - 83.1 87.8 82.0 68.3 - 105 Sim Track-B [7] 69.3 78.5 - - - 54.8 53.8 82.3 86.5 - 69.8 89.6 40 CSWin TT [46] 66.2 75.2 70.9 - - - - 81.9 86.7 79.5 70.5 90.3 12 SBT-Base [49] 65.9 - 70.0 - - - - - - - - - 37 Swin Track-T [34] 67.2 - 70.8 47.6 53.9 53.0 53.2 81.1 - 78.4 - - 98 To MP101 [39] 68.5 79.2 68.5 - - - - 81.5 86.4 78.9 66.9 - 20 STARK-ST50 [52] 66.4 - - - - - - 81.3 86.1 - - - 42 Keep Track [40] 67.1 77.2 70.2 48.2 - - - - - - 69.7 - 19 Trans T [10] 64.9 73.8 69.0 - - 50.7 51.7 81.4 86.7 80.3 69.1 - 50 Pr Di MP [17] 59.8 68.8 60.8 - - - - 75.8 81.6 70.4 68.0 - 47 ATOM [16] 51.5 57.6 50.5 - - - - 70.3 77.1 64.8 64.3 - 83 Table 5: State-of-the-art comparison on Tracking Net [42], La SOT [20], La SOText [20], UAV123 [41] and TNL2K [48]. The best two results are shown in bold and underline fonts. * denotes tracker with Mix Vi T-B as the teacher during the dense-to-sparse distillation process. The default teacher is Mix Vi T-L. Only trackers of similar complexity are included. Method La SOT La SOText TNL2K Tracking Net UAV123 Speed AUC PNorm P AUC P AUC P AUC PNorm P AUC P GPU CPU Mix Former V2-S 60.6 69.9 60.4 43.6 46.2 48.3 43 75.8 81.1 70.4 65.8 86.8 325 30 FEAR-L [5] 57.9 68.6 60.9 - - - - - - - - - - - FEAR-XS [5] 53.5 64.1 54.5 - - - - - - - - - 80 26 HCAT[9] 59.0 68.3 60.5 - - - - 76.6 82.6 72.9 63.6 - 195 45 E.T.Track [4] 59.1 - - - - - - 74.5 80.3 70.6 62.3 - 150 42 Light Track-Large A [53] 55.5 - 56.1 - - - - 73.6 78.8 70.0 - - - - Light Track-Mobile [53] 53.8 - 53.7 - - - - 72.5 77.9 69.5 - - 120 36 STARK-Lightning 58.6 69.0 57.9 - - - - - - - - - 200 42 Di MP [3] 56.9 65.0 56.7 - - - - 74.0 80.1 68.7 65.4 - 77 15 Siam FC++ [50] 54.4 62.3 54.7 - - - - 75.4 80.0 70.5 - - 90 20 Table 6: Comparison with CPU-realtime trackers on Tracking Net [42], La SOT [20], La SOText [20], UAV123 [41] and TNL2k [48]. The best results are shown in bold fonts. included, i.e., the trackers with large-scale backbone or large input resolution are excluded. Our Mix Former V2-B achieves state-of-the-art performance among these trackers with a very fast speed, especially compared to transformer-based one-stream tracker. For example, Mix Former V2-B without post-processing strategies surpasses OSTrack by 1.5% AUC on La SOT and 2.4% AUC on TNL2k, running with quite faster speed (165 FPS vs. 105 FPS). Even the Mix Former V2-B with Mix Vi T-B as the teacher model obtains better performance than existing SOTA trackers, such as Mix Former, OSTrack, To MP101 and Sim Track, with much faster running speed on GPU. Comparison with Efficient Trackers. For real-time running requirements on limited computing resources such as CPU, we explore a lightweight model, i.e. Mix Former V2-S, which still reaches strong performance. And it is worth noting that this is the first time that transformer-based onestream tracker is able to run on CPU device with a real-time speed. As demonstrated in Figure 6, Mix Former V2-S surpasses all other architectures of CPU-real-time trackers by a large margin. We take a comparison with other prevailing efficient trackers on multiple datasets, including La SOT, Tracking Net, UAV123 and TNL2k, in Tab 6. We can see that our Mix Former V2-S outperforms FEAR-L by a an AUC score of 2.7% and STARK-Lightning by an AUC score of 2.0% on La SOT. 5 Conclusion In this paper, we have proposed a fully transformer tracking framework Mix Former V2, composed of standard Vi T backbones on the mixed token sequence and simple MLP heads for box regression and quality score estimation. Our Mix Former V2 streamlines the tracking pipeline by removing the dense convolutional head and the complex score prediction modules. We also present a distillation based model reduction paradigm for Mix Former V2 to further improve its efficiency. Our Mix Former V2 obtains a good trade-off between tracking accuracy and speed on both GPU and CPU platforms. We hope our Mix Former V2 can facilitate the development of efficient transformer trackers in the future. Acknowledgement This work is supported by National Key R&D Program of China (No. 2022ZD0160900), National Natural Science Foundation of China (No. 62076119, No. 61921006), Fundamental Research Funds for the Central Universities (No. 020214380099), and Collaborative Innovation Center of Novel Software Technology and Industrialization. [1] Romero Adriana, Ballas Nicolas, K Samira Ebrahimi, Chassang Antoine, Gatta Carlo, and B Yoshua. Fitnets: Hints for thin deep nets. Proc. ICLR, 2, 2015. [2] Luca Bertinetto, Jack Valmadre, Joao F Henriques, Andrea Vedaldi, and Philip HS Torr. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, ECCV Workshops, 2016. [3] Goutam Bhat, Martin Danelljan, Luc Van Gool, and Radu Timofte. Learning discriminative model prediction for tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV, pages 6182 6191, 2019. [4] Philippe Blatter, Menelaos Kanakis, Martin Danelljan, and Luc Van Gool. Efficient visual tracking with exemplar transformers. ar Xiv preprint ar Xiv:2112.09686, 2021. [5] Vasyl Borsuk, Roman Vei, Orest Kupyn, Tetiana Martyniuk, Igor Krashenyi, and Jiˇri Matas. Fear: Fast, efficient, accurate and robust visual tracker. ar Xiv preprint ar Xiv:2112.07957, 2021. [6] Arnav Chavan, Zhiqiang Shen, Zhuang Liu, Zechun Liu, Kwang-Ting Cheng, and Eric P Xing. Vision transformer slimming: Multi-dimension searching in continuous optimization space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4931 4941, 2022. [7] Boyu Chen, Peixia Li, Lei Bai, Lei Qiao, Qiuhong Shen, Bo Li, Weihao Gan, Wei Wu, and Wanli Ouyang. Backbone is all your need: A simplified architecture for visual object tracking. In Proceedings of the European Conference on Computer Vision, ECCV, 2022. [8] Minghao Chen, Houwen Peng, Jianlong Fu, and Haibin Ling. Autoformer: Searching transformers for visual recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 12270 12280, 2021. [9] Xin Chen, Dong Wang, Dongdong Li, and Huchuan Lu. Efficient visual tracking via hierarchical crossattention transformer. ar Xiv preprint ar Xiv:2203.13537, 2022. [10] Xin Chen, Bin Yan, Jiawen Zhu, Dong Wang, Xiaoyun Yang, and Huchuan Lu. Transformer tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2021. [11] Zedu Chen, Bineng Zhong, Guorong Li, Shengping Zhang, and Rongrong Ji. Siamese box adaptive network for visual tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2020. [12] Yutao Cui, Cheng Jiang, Limin Wang, and Gangshan Wu. Target transformed regression for accurate tracking. Co RR, 2021. [13] Yutao Cui, Cheng Jiang, Limin Wang, and Gangshan Wu. Fully convolutional online tracking. Computer Vision and Image Understanding, 224:103547, 2022. [14] Yutao Cui, Cheng Jiang, Limin Wang, and Gangshan Wu. Mixformer: End-to-end tracking with iterative mixed attention. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2022. [15] Yutao Cui, Cheng Jiang, Gangshan Wu, and Limin Wang. Mixformer: End-to-end tracking with iterative mixed attention, 2023. [16] Martin Danelljan, Goutam Bhat, Fahad Shahbaz Khan, and Michael Felsberg. ATOM: accurate tracking by overlap maximization. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2019. [17] Martin Danelljan, Luc Van Gool, and Radu Timofte. Probabilistic regression for visual tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2020. [18] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, ICLR, 2021. [19] Thomas Elsken, Jan Hendrik Metzen, and Frank Hutter. Neural architecture search: A survey. The Journal of Machine Learning Research, 20(1):1997 2017, 2019. [20] Heng Fan, Liting Lin, Fan Yang, Peng Chu, Ge Deng, Sijia Yu, Hexin Bai, Yong Xu, Chunyuan Liao, and Haibin Ling. Lasot: A high-quality benchmark for large-scale single object tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2019. [21] Chengyue Gong, Dilin Wang, Meng Li, Xinlei Chen, Zhicheng Yan, Yuandong Tian, Vikas Chandra, et al. Nasvit: Neural architecture search for efficient vision transformers with gradient conflict aware supernet training. In International Conference on Learning Representations, 2021. [22] Yunchao Gong, Liu Liu, Ming Yang, and Lubomir Bourdev. Compressing deep convolutional networks using vector quantization. ar Xiv preprint ar Xiv:1412.6115, 2014. [23] Jianyuan Guo, Kai Han, Yunhe Wang, Han Wu, Xinghao Chen, Chunjing Xu, and Chang Xu. Distilling object detectors via decoupled features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2154 2164, 2021. [24] Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2022. [25] Yihui He, Xiangyu Zhang, and Jian Sun. Channel pruning for accelerating very deep neural networks. In Proceedings of the IEEE international conference on computer vision, pages 1389 1397, 2017. [26] João F. Henriques, Rui Caseiro, Pedro Martins, and Jorge Batista. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell., 37(3):583 596, 2015. [27] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network (2015). ar Xiv preprint ar Xiv:1503.02531, 2, 2015. [28] Lianghua Huang, Xin Zhao, and Kaiqi Huang. Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell., 43(5):1562 1577, 2021. [29] Matej Kristan, Ales Leonardis, and et. al. The eighth visual object tracking VOT2020 challenge results. In Adrien Bartoli and Andrea Fusiello, editors, Proceedings of the European Conference on Computer Vision, ECCV Workshops, 2020. [30] Matej Kristan, Aleš Leonardis, Jiˇrí Matas, Michael Felsberg, Roman Pflugfelder, Joni-Kristian Kämäräinen, Hyung Jin Chang, Martin Danelljan, Luka ˇCehovin Zajc, Alan Lukežiˇc, et al. The tenth visual object tracking vot2022 challenge results. In ECCV 2022 Workshops, pages 431 460, 2023. [31] Bo Li, Wei Wu, Qiang Wang, Fangyi Zhang, Junliang Xing, and Junjie Yan. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2019. [32] Bo Li, Junjie Yan, Wei Wu, Zheng Zhu, and Xiaolin Hu. High performance visual tracking with siamese region proposal network. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2018. [33] Quanquan Li, Shengying Jin, and Junjie Yan. Mimicking very efficient network for object detection. In Proceedings of the ieee conference on computer vision and pattern recognition, pages 6356 6364, 2017. [34] Liting Lin, Heng Fan, Yong Xu, and Haibin Ling. Swintrack: A simple and strong baseline for transformer tracking. Neural Information Processing Systems, NIPS, 2022. [35] Tsung-Yi Lin, Michael Maire, Serge J. Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. Microsoft COCO: common objects in context. In Proceedings of the European Conference on Computer Vision, ECCV, 2014. [36] Benlin Liu, Yongming Rao, Jiwen Lu, Jie Zhou, and Cho-Jui Hsieh. Metadistiller: Network self-boosting via meta-learned top-down distillation. In European Conference on Computer Vision, pages 694 709. Springer, 2020. [37] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV, 2021. [38] Alan Lukezic, Jiri Matas, and Matej Kristan. D3S - A discriminative single shot segmentation tracker. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2020. [39] Christoph Mayer, Martin Danelljan, Goutam Bhat, Matthieu Paul, Danda Pani Paudel, Fisher Yu, and Luc Van Gool. Transforming model prediction for tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2022. [40] Christoph Mayer, Martin Danelljan, Danda Pani Paudel, and Luc Van Gool. Learning target candidate association to keep track of what not to track. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV, 2021. [41] Matthias Mueller, Neil Smith, and Bernard Ghanem. A benchmark and simulator for UAV tracking. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Proceedings of the European Conference on Computer Vision, ECCV, 2016. [42] Matthias Müller, Adel Bibi, Silvio Giancola, Salman Al-Subaihi, and Bernard Ghanem. Trackingnet: A large-scale dataset and benchmark for object tracking in the wild. In Proceedings of the European Conference on Computer Vision, ECCV, 2018. [43] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748 8763. PMLR, 2021. [44] Yongming Rao, Wenliang Zhao, Benlin Liu, Jiwen Lu, Jie Zhou, and Cho-Jui Hsieh. Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems, 34:13937 13949, 2021. [45] Zikai Song, Run Luo, Junqing Yu, Yi-Ping Phoebe Chen, and Wei Yang. Compact transformer tracker with correlative masked modeling. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), February 2023. [46] Zikai Song, Junqing Yu, Yi-Ping Phoebe Chen, and Wei Yang. Transformer tracking with cyclic shifting window attention. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2022. [47] Kuan Wang, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. Haq: Hardware-aware automated quantization with mixed precision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8612 8620, 2019. [48] Xiao Wang, Xiujun Shu, Zhipeng Zhang, Bo Jiang, Yaowei Wang, Yonghong Tian, and Feng Wu. Towards more flexible and accurate object tracking with natural language: Algorithms and benchmark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13763 13773, 2021. [49] Fei Xie, Chunyu Wang, Guangting Wang, Yue Cao, Wankou Yang, and Wenjun Zeng. Correlation-aware deep tracking. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2022. [50] Yinda Xu, Zeyu Wang, Zuoxin Li, Ye Yuan, and Gang Yu. Siamfc++: Towards robust and accurate visual tracking with target estimation guidelines. In Proceedings of the AAAI Conference on Artificial Intelligence, AAAI, 2020. [51] Yifan Xu, Zhijie Zhang, Mengdan Zhang, Kekai Sheng, Ke Li, Weiming Dong, Liqing Zhang, Changsheng Xu, and Xing Sun. Evo-vit: Slow-fast token evolution for dynamic vision transformer. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 2964 2972, 2022. [52] Bin Yan, Houwen Peng, Jianlong Fu, Dong Wang, and Huchuan Lu. Learning spatio-temporal transformer for visual tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV, 2021. [53] Bin Yan, Houwen Peng, Kan Wu, Dong Wang, Jianlong Fu, and Huchuan Lu. Lighttrack: Finding lightweight neural networks for object tracking via one-shot architecture search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15180 15189, 2021. [54] Zhendong Yang, Zhe Li, Ailing Zeng, Zexian Li, Chun Yuan, and Yu Li. Vitkd: Practical guidelines for vit feature knowledge distillation. ar Xiv preprint ar Xiv:2209.02432, 2022. [55] Botao Ye, Hong Chang, Bingpeng Ma, and Shiguang Shan. Joint feature learning and relation modeling for tracking: A one-stream framework. Proceedings of the European Conference on Computer Vision, ECCV, 2022. [56] Jinnian Zhang, Houwen Peng, Kan Wu, Mengchen Liu, Bin Xiao, Jianlong Fu, and Lu Yuan. Minivit: Compressing vision transformers with weight multiplexing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12145 12154, 2022. [57] Zhaohui Zheng, Rongguang Ye, Qibin Hou, Dongwei Ren, Ping Wang, Wangmeng Zuo, and Ming-Ming Cheng. Localization distillation for object detection. ar Xiv preprint ar Xiv:2204.05957, 2022. Broader Impact In this paper, we introduce Mix Former V2, a fully transformer tracking approach for efficiently and effectively estimating the state of an arbitrary target in a video. Generic object tracking is one of the fundamental computer vision problems with numerous applications. For example, object tracking (and hence Mix Former V2) could be applied to human-machine interaction, visual surveillance and unmanned vehicles. Our research could be used to improve the tracking performance while maintaining a high running speed. Of particular concern is the use of the tracker by those wishing to position and surveil others illegally. Besides, if the tracker is used in unmanned vehicles, it may be a challenge when facing the complex real-world scenarios. To mitigate the risks associated with using Mix Former V2, we encourage researchers to understand the impacts of using the trackers in particular real-world scenarios. Limitations The main limitation lies in the training overhead of Mix Former V2-S, which performs multiple model pruning based on the dense-to-sparse distillation and deep-to-shallow distillation. In detail, we first perform distillation from Mix Vi T with 12 layers and plain corner head to Mix Former V2 of 12 layers. The 12-layers Mix Former V2 is pruned to 8-layers and then to 4-layers Mix Former V2 based on the deep-to-shallow distillation. Finally, the MLP-ratio-4.0 4-layers Mix Former V2 is pruned to the MLP-ratio-4.0 4-layers Mix Former V2-S for real-time tracking on CPU. For each step, it requires training for 500 epochs which is time-consuming. S.1 Details of Training Time The models are trained on 8 Nvidia RTX8000 GPUs. The dense-to-sparse stage takes about 43 hours. The deep-to-shallow stage1 (12-to-8 layers) takes about 42 hours, and stage2 (8-to-4 layers) takes about 35 hours. S.2 More Results on VOT2020 and GOT10k VOT2020. We evaluate our tracker on VOT2020 [29] benchmark, which consists of 60 videos with several challenges including fast motion, occlusion, etc. The results is reported in Table 7, with metrics Expected Average Overlap(EAO) considering both Accuracy(A) and Robustness. Our Mix Former V2B obtains an EAO score of 0.322 surpassing CSWin TT by 1.8%. Besides, our Mix Former V2-S achieves an EAO score of 0.258, which is higher than the efficient tracker Light Track, with a real-time running speed on CPU. GOT10k. GOT10k [28] is a large-scale dataset with over 10000 video segments and has 180 segments for the test set. Apart from generic classes of moving objects and motion patterns, the object classes in the train and test set are zero-overlapped. We evaluate Mix Former V2 trained with the four datasets of La SOT, Tracking Net, COCO and GOT10k-train on GOT10k-test. We compare it with Mix Former and Trans T with the same training datasets for fair comparison. Mix Former V2-B improves Mix Former and Trans T by 0.7% and 1.6% on AO respectively with a high running speed of 165 FPS. KCF Siam FC ATOM Light Track Di MP STARK Trans T CSWin TT Mix Former Ours-S Ours-B [26] [2] [16] [53] [3] [52] [10] [46] [14] VOT20EAO 0.154 0.179 0.271 0.242 0.274 0.280 - 0.304 - 0.258 0.322 GOT10k AO 0.203 0.348 0.556 - 0.611 0.688 0.723 0.694 0.726 0.621 0.739 Table 7: State-of-the-art comparison on VOT2020 [29] and GOT10k [28]. denotes training with four datasets including La SOT [20], Tracking Net [42], GOT10k [28] and COCO [35]. The best results are shown in bold font. S.3 More Ablation Studies Design of Prediction Tokens. We practice three different designs of prediction tokens for the target localization in Tab. 8a. All the three methods use the formulation of estimating the probability token num. MLP num. AUC 1 4 67.1% 4 4 67.3 4 1 67.5% (a) Different prediction designs. token num. indicates the number of the learnable prediction tokens, MLP num. denotes the number of employed MLP layers for localization. Init. method La SOT La SOT_ext UAV123 Tea-fir4 62.9% 45.2% 65.7% Tea-skip4 64.4% 46.1% 66.6% PMDP 64.8% 47.1% 67.5% (b) Progressive Model Depth Pruning (PMDP). Tea-fir4 denotes using first 4 layers of the teacher for student initialization. Teaskip4 is using 4 skipped layers of the teacher. Table 8: More ablation studies. The default choice for our model is colored in gray . distribution of the four coordinates of the bounding box. The model on the first line denotes using one prediction token and then predicting coordinates distribution with four independent MLP heads. It can be observed that adopting separate prediction tokens for the four coordinates and a same MLP head retains the best accuracy. More Exploration of PMDP Tea-skip4 is a special initialization method, which chooses the skiped four layers (layer-3/6/9/12) of the teacher (Mix Vi T-B) for initialization. In other words, Tea-skip4 is an extreme case of ours PMDP when the eliminating epoch m equal to 0. So it is reasonable that Tea-skip4 performs better than the baseline Tea-fir4, which employs the first four layers of the teacher (Mix Vi T-B) to initialize the student backbone. In Table 8b, we further evaluate the performance on more benchmarks. It can be seen that ours PMDP surpasses Tea-skip4 by 1.0% on La SOT_ext, which demonstrate its effectiveness. Computation Loads of Different Localization Head We showcase the FLOPs of different heads as follows. Formally, we denote Cin as input feature dimension, Cout as output feature dimension, Hin, Win as input feature map shape of convolution layer, Hout, Wout as output feature map shape, and K as the convolution kernel size. The computational complexity of one linear layer is O(Cin Cout), and that of one convolutional layer is O(Cin Cout Hout Wout K2). In our situation, for T4, the Localization Head contains four MLP to predict four coordinates. Each MLP contains two linear layer, whose input and output dimensions are all 768. The loads can be calculated as: Load T 4 = 4 (768 768 + 768 72) = 2580480 2.5M For Py-Corner, totally 24 convolution layers are used. The loads can be calculated as: Load P y Corner = 2 (768 384 18 18 3 3+ 384 192 18 18 3 3+ 384 192 18 18 3 3+ 192 96 36 36 3 3+ 384 96 18 18 3 3+ 96 48 72 72 3 3+ 48 1 72 72 3 3+ 192 96 18 18 3 3+ 96 48 18 18 3 3+ 48 1 18 18 3 3+ 96 48 36 36 3 3+ 48 1 36 36 3 3) =3902587776 3.9B For simplicity, we do not include some operations such as bias terms and Layer/Batch-Normalization, which does not affect the overall calculation load level. Besides, the Pyramid Corner Head utilize additional ten interpolation operations. Obviously the calculation load of Py-Corner is still hundreds of times of T4. S.4 Visualization Results Visualization of Attention Map To explore how the introduced learnable prediction tokens work within the P-MAM, we visualize the attention maps of prediction-token-to-search and predictiontoken-to-template in Fig. 5 and Fig. 6, where the prediction tokens are served as query and the others as key/val of the attention operation. From the visualization results, we can arrive that the four prediction tokens are sensitive to corresponding part of the targets and thus yielding a compact object bounding box. We suspect that the performance gap between the dense corner head based Mix Vi T-B and our fully transformer Mix Former V2-B without distillation lies in the lack of holistic target modeling capability. Besides, the prediction tokens tend to extract partial target information in both the template and the search so as to relate the two ones. Visualization of Predicted Probability Distribution We show two good cases and bad cases in Figure 7. In Figure 7a Mix Former V2 deals with occlusion well and locate the bottom edge correctly. As show in Figure 7b, the probability distribution of box representation can effectively alleviate issue of ambiguous boundaries. There still exist problems like strong occlusion and similar objects which will lead distribution shift, as demonstrated in Figure 7c and 7d. Search Top Left Bottom Right Figure 5: Visualization of prediction-token-tosearch attention maps, where the prediction tokens are served as query of attention operation. Template Top Left Bottom Right Figure 6: Visualization of prediction-token-totemplate attention maps, where the prediction tokens are served as query of attention operation. (d) Figure 7: In each figure, the left one is plot of the probability distribution of predicted box (red), which demonstrates how our algorithm works. The right one is heatmap of attention weights in the backbone. The examples are from La SOT test subset and the green boxes are ground truths.