Explicit Visual Prompts for Visual Object Tracking
Abstract
How to effectively exploit spatio-temporal information is crucial to capture target appearance changes in visual tracking. However, most deep learning-based trackers mainly focus on designing a complicated appearance model or template updating strategy, while lacking the exploitation of context between consecutive frames and thus entailing the when-and-how-to-update dilemma. To address these issues, we propose a novel explicit visual prompts framework for visual tracking, dubbed EVPTrack. Specifically, we utilize spatio-temporal tokens to propagate information between consecutive frames without focusing on updating templates. As a result, we cannot only alleviate the challenge of when-to-update, but also avoid the hyper-parameters associated with updating strategies. Then, we utilize the spatio-temporal tokens to generate explicit visual prompts that facilitate inference in the current frame. The prompts are fed into a transformer encoder together with the image tokens without additional processing. Consequently, the efficiency of our model is improved by avoiding how-to-update. In addition, we consider multi-scale information as explicit visual prompts, providing multiscale template features to enhance the EVPTrack’s ability to handle target scale changes. Extensive experimental results on six benchmarks (i.e., LaSOT, LaSOText, GOT-10k, UAV123, TrackingNet, and TNL2K.) validate that our EVPTrack can achieve competitive performance at a real-time speed by effectively exploiting both spatio-temporal and multi-scale information. Code and models are available at https://github.com/GXNU-ZhongLab/EVPTrack.
Introduction

Visual object tracking is a fundamental task in computer vision that aims to locate targets in subsequent frames of a video given an arbitrary target in the first frame. It plays an important role in traffic flow monitoring, human-computer interaction, surveillance, security, etc. Although various top-performing trackers have been proposed over the years, it is still challenging due to target appearance changes, deformation, occlusions, background clutter, and other factors.
In recent years, deep learning-based trackers (Bertinetto et al. 2016; Li et al. 2018, 2019; Chen et al. 2020, 2021; Ye et al. 2022; Chen et al. 2023) have achieved excellent performance and efficiency with their powerful representation capabilities, leading the rapid development of the tracking field. Generally speaking, as shown in Fig. 1, they can be divided into two categories: the trackers with an initial fixed template and the trackers with a dynamic template. The representative trackers with an initial fixed template include siamese-based trackers (Bertinetto et al. 2016; Li et al. 2018, 2019; Chen et al. 2020) and transformer-based trackers (Chen et al. 2021; Ye et al. 2022; Chen et al. 2023). The siamese-based trackers formulate the tracking task as a similarity learning problem, matching the initial state of the target with the search frame image to estimate a new target state. The transformer-based trackers introduce attention mechanisms(Vaswani et al. 2017) that are utilized for feature extraction or feature fusion. As visualized in Fig. 1(a), these methods typically use image pairs as input and utilize only the information from an initial template to predict the target positions in the search frames. As a result, they ignore frame-to-frame associations in a video and thus make it challenging to cope with significant target appearance changes. Therefore, some online updated trackers (Zhang et al. 2019; Yan et al. 2021; Cui et al. 2022) introduce template updating mechanisms to enrich the target information and perceive target appearance changes by selecting the search frame as a dynamic template, as shown in Fig. 1(b). To ensure timely updating and avoid introducing redundant information, the above mentioned methods usually require careful design of the selection strategy and relevant hyperparameters (e.g., score thresholds that control when to update). In addition, some other methods(Fu et al. 2021; Wang et al. 2021a) use spatio-temporal information by introducing memory networks and online update modules, which usually harms efficiency. Despite their success, the above methods face difficulties in deciding when and how to update, and they lack the exploitation of contextual information between consecutive frames.
To address the aforementioned issues, inspired by the philosophy of prompt learning, we propose a novel explicit visual prompts (i.e., spatio-temporal prompts and multi-scale prompts.) tracking framework (EVPTrack) for visual tracking. As illustrated in Fig. 1(c), EVPTrack utilizes tokens to propagate information between consecutive frames, and generate prompts with spatio-temporal information via the tokens. Specifically, we design a simple Spatio-Temporal Encoder for obtaining new spatio-temporal tokens to propagate the spatio-temporal information to the next frame. And, we design a Prompt Generator for obtaining sptaio-temporal prompts to refine the spatio-temnporal tokens from the previous frame. In addition, we divide the template image into multiple scale patches to obtain different fine-grained template features to generate multi-scale prompts that enable the tracker to model multi-scale targets efficiently. These explicit visual prompts are fed into Image-Prompt Encoder along with image tokens for relational modeling. This is a novel approach that we propose that can use both spatio-temporal and multi-scale information. Although inspiration comes from prompt learning, it has essential differences from prompt learning: (i) Tasks are different. Prompt learning is to adapt large models to downstream tasks, while we are exploiting temporal information. (ii) Different technical solutions. Prompt learning is by freezing the large model and adding learnable parameters, which is an implicit prompt, whereas we generate prompts using templates and spatio-temporal tokens, which is an explicit visual prompt. In summary, our main contributions are as follows:
-
•
We present a novel explicit visual prompts tracking framework (EVPTrack) that effectively utilizes both spatio-temporal and multi-scale information, and our explicit visual prompts are made from these two kinds of information.
-
•
To avoid the when-and-how-to-update dilemma associated with template update mechanisms, we introduce a novel propagation mechanism that utilizes tokens to propagate spatio-temporal information between consecutive frames.
-
•
Our tracker achieves state-of-the-art results on six benchmarks. In particular, our method achieved 72.7% success score (AUC) on the LaSOT test set.
Related Work
Tracking methods using an initial template. The initial template is indispensable, whether it is the previous siamese-based trackers or the mainstream transformer-based trackers nowadays. Generally, the initial template is the target state given in the first frame with reliable apparent information. Therefore, benefiting from the powerful representation capability of deep learning, the template-search based matching paradigm is widely popular due to its simplicity and effectiveness. For example, SiamFC(Bertinetto et al. 2016) is a pioneering work in the siamese series of trackers that uses a siamese-network of shared weights to extract search features and template features. Subsequently, it uses the features to perform cross-correlation operations to localize a target in the current frame. To utilize richer semantic information and avoid falling into local optimization due to cross-correlation, transformer-based trackers(Chen et al. 2021; Yan et al. 2021) utilize the transformer to fusion search features and template features. Recent one-stream trackers(Ye et al. 2022; Chen et al. 2022), which use a transformer-based network to perform both feature extraction and feature fusion, allow for a more full fusion of search features with template features. The above methods can reach a competitive performance by relying only on the initial template frame. However, due to the lack of sensing target changes, they are not robust when facing complex scenes (scale variation, aspect ratio change, etc.) and long-term tracking. Therefore, it is necessary to enable trackers to perceive target changes by exploiting spatio-temporal information.

Tracking methods using a dynamic template. To overcome the above challenges, on the one hand, the DCF-based trackers(Danelljan et al. 2019; Danelljan, Gool, and Timofte 2020) try to exploit the temporal information by updating the model online, but they suffer from complex manual optimization. STMTrack(Fu et al. 2021) introduces a memory network, which stores the historical information of the target to guide the tracker to focus on the most informative regions in the current frame. TCTrack(Cao et al. 2022) exploits consecutive contextual information through online temporal adaptive convolution and adaptive temporal transformers. On the other hand, template update mechanisms are usually introduced in the siamese-based trackers(Zhang et al. 2019) and the transformer-based trackers(Yan et al. 2021; Cui et al. 2022). UpdateNet(Zhang et al. 2019) designs a CNN network to generate the optimal template available for the next frame to realize the template update function. Stark(Yan et al. 2021) designs a target quality branch that evaluates the quality of the target in the current frame to update the dynamic template selectively. Mixformer(Cui et al. 2022) selects high-quality templates by presenting a valid target prediction score module, similar to STARK. Despite the success of the above methods, there are the following limitations: (1) It requires a complex design of update modules and update strategies to ensure timely updates and avoid introducing redundant information. (2) It cannot fully utilize the context information between consecutive frames. Therefore, in this work, EVPTrack uses tokens to propagate spatio-temporal information that avoids complex update strategies.
Method
In this section, we will present the proposed EVPTrack method in detail. First, we will briefly describe an overview tracking framework of EVPTrack. Then, we will give an account of each module of the whole framework one by one. Finally, we will introduce the training and inference pipelines.
Overview
As shown in Fig.2, EVPTrack is a simple end-to-end tracker with Image-Prompt Encoder, Spatio-Temporal Encoder, and Prompt Generator. The Image-Prompt Encoder is used for the fusion of explicit visual prompts with image features to efficiently utilize spatio-temporal and multi-scale information. The Spatio-Temporal Encoder is utilized for the interaction among template tokens, search tokens, and spatio-temporal tokens to propagate spatio-temporal information. The Prompt Generator extracts information from template and spatio-temporal tokens to generate explicit visual prompts (i.e., multi-scale prompts and spatio-temporal prompts). EVPTrack employs image pairs and spatio-temporal tokens as input. Specifically, here the tokens come from the previous frame. One of the images is a template image . The image is from the target region in the first frame of the video and contains information about the initial appearance of the target and some of the background information. The other image is a search image . It comes from the region where the target may appear in the subsequent sequences of the video. To summarize, we utilize spatio-temporal and multi-scale information by providing explicit visual prompts to improve the robustness of the tracker in complex scenarios such as target appearance changes and deformation.
Image-Prompt Encoder
Image-Prompt Encoder inputs include search tokens, template tokens, and explicit visual prompts. We use explicit visual prompts to guide feature extraction and fusion without other complex operations. This way of utilizing spatio-temporal information is simple and avoids additional elaborate updated modules. Specifically, we first perform linear projection on the image pairs. Here, we utilize hierarchical patch embedding with a total downsampling stride of 16, and obtain search tokens and template tokens denoted as and , respectively. Here, , , . And we add learnable position encoding (, ) to preserve spatial information. Then, we concatenate , and feed them into N-layer encoder. Finally, the search features are utilized to locate the target. This process can be described by the following equations:
(1) |
here, , denote explicit visual prompts and layer normalization, respectively. For a more detailed design of encoder and patch embedding refer to HiViT(Zhang et al. 2023).

Spatio-Temporal Encoder
Spatio-Temporal Encoder is a critical component of our framework for maintaining spatio-temporal tokens and thus exploiting the rich contextual information in successive frames. Its structure is shown in Fig.3, which is also based on the transformer encoder implementation. Specifically, it uses template, search, and spatio-temporal tokens as input, and obtains new spatio-temporal tokens after feature fusion by transformer encoder. This process can be described by the following equations:
(2) |
here, denotes the spatio-temporal tokens obtained at time t-1, which will be used at moment t. When t = 1, we initialize it using template tokens, i.e., , which ensures that the initialized spatio-temporal tokens can be adapted to any video sequence and can enhance the target features. Then, we split the template tokens into as new spatio-temporal tokens. This is to avoid heavy accumulation of mistakes in the spatio-temporal tokens propagated during long-term tracking.
We propagate information between successive frames via tokens, thus effectively exploiting spatio-temporal information to capture the target appearance changes. In contrast to intermittent template updating mechanisms, we utilize a mechanism that allows for continuous propagation of spatio-temporal information. Therefore, avoiding the difficult problem of when-to-update that appears with intermittent template updates.
Prompt Generator
EVPTrack is equipped with prompt generators to obtain explicit visual prompts (i.e., spatio-temporal prompts and multi-scale prompts). On the one hand, the role of spatio-temporal prompts is to enable the tracker to capture target appearance changes. On the other hand, the multi-scale prompts provide different fine-grained features, thus maintaining excellent tracking even when variations in target scale. For multi-scale prompt generator, whose structure is shown in Fig.4(a). We first crop the template image into patches with resolution (). Then, each patch is mapped to a 1D vector of size D using a learnable linear projection . In order to obtain features at different scales, here we use three sizes of resolution, P is set to 14,16,18 respectively. The features are obtained corresponding to three different scales. We perform avgpooling on each feature separately and then concatenate to get . Finally, are fed into a fully connected network to get the final multi-scale prompts . The following equations can describe this process:
(3) |
As shown in Fig.4(b), the spatio-temporal prompt generator uses spatio-temporal tokens as input. First, the input is average pooled by performing avgpooling operation. Then, the extracted features are fed into a fully connected network to obtain the final prompts. Although, these prompt generators are simple in design, it is shown in the experimental section that the generated explicit visual prompts are effective. Therefore, EVPTrack using a small number of explicit visual prompts reduces the computational burden and avoids introducing too much redundant information.
Training and Inference

Offline Training. EVPTrack uses a conventional center point prediction(Ye et al. 2022) that contains a regression branch and a classification branch for estimating the target bounding box’s and determining the target box location, respectively. During training optimization, For the classification branch we use Gaussian weights focal loss(Lin et al. 2017), For the regression branch, a combined L1 loss and GIoU loss (Rezatofighi et al. 2019) is employed. The total loss function is denoted as:
(4) |
here, and are regularization parameters.
Online Inference. During the inference process, we used post-processing on the prediction results in order to smooth the target movement and changes. Specifically, similar to trackers(Ye et al. 2022; Chen et al. 2021), a penalty is applied to the classification scores using the Hanning window. After performing the penalty, the box corresponding to the highest score point is selected thereby obtaining the prediction bounding box.
Experiments
Implementation Details
Our methods are implemented based on python3.8 and pytorch1.10 framework. Our trackers were trained on 4 NVIDIA A10 GPUs. During the inference phase, the trackers were tested at speed on a single NVIDIA RTX2080Ti.
Models. We train two variants of EVPTrack with different input image pair resolutions as follows:
-
•
EVPTrack-224. Template size: 112x112 pixels. Search region size: 224x224 pixels.
-
•
EVPTrack-384. Template size: 192x192 pixels. Search region size: 384x384 pixels.
Training strategy. We use HiViT-Base(Zhang et al. 2023) model as the Image-Prompt Encoder and its parameters are initialized with MAE(He et al. 2022). EVPTrack is trained on the same datasets as mainstream trackers(Ye et al. 2022), including LaSOT(Fan et al. 2019), GOT-10k (Huang, Zhao, and Huang 2021), TrackingNet(Müller et al. 2018), COCO (Lin et al. 2014). During training, data augmentation employed horizontal flip and brightness jittering, and the optimizer utilized AdamW(Loshchilov and Hutter 2019). The learning rate of backbone is set to , the learning rate decay is set to , and the learning rate of other parameters is set to . A total of 150 epochs of training, and each epoch uses 60k search images. The learning rate decreases by factor after 120 epochs. In addition, for a fair comparison of got-10k, we follow its standard protocol and train out the other two models using only the got10k dataset. The models were trained for 50 epochs and started decaying in the 40th epoch as it was trained with only one dataset.
In order to allow EVPTrack to learn the spatio-temporal information between consecutive frames, the sampling strategy during training is therefore different from the traditional sampling strategy. Specifically, we sample M videos in one iteration, each containing N search frames and one template frame. Therefore the batch size is . Then, the N search frames are predicted sequentially in order, enabling the tracker to utilize tokens to propagate spatio-temporal information between consecutive frames. For EVPTrack-224, we set N, M to 4 and 8, respectively, with a batch size of 32. EVPTrack-224 is trained on 4 GPUs, so the total batch size is 128. Due to GPU memory constraints, for EVPTrack-384 we set N, M to 4 and 8, respectively.
Method | Source | LaSOT | GOT-10K∗ | UAV123 | |||||||
AUC | P | AUC | P | AO | AUC | ||||||
SiamFC (Bertinetto et al. 2016) | ECCVW16 | 33.6 | 42.0 | 33.9 | 23.0 | 31.1 | 26.9 | 34.8 | 35.3 | 9.8 | 46.8 |
MDNet (Nam and Han 2016) | CVPR16 | 39.7 | 46.0 | 37.3 | 27.9 | 34.9 | 31.8 | 29.9 | 30.3 | 9.9 | 52.8 |
ECO (Danelljan et al. 2017) | ICCV 17 | 32.4 | 33.8 | 30.1 | 22.0 | 25.2 | 24.0 | 31.6 | 30.9 | 11.1 | 53.5 |
SiamPRN++ (Li et al. 2019) | CVPR19 | 49.6 | 56.9 | 49.1 | 34.0 | 41.6 | 39.6 | 51.7 | 61.6 | 32.5 | 61.0 |
Ocean (Zhang et al. 2020) | ECCV 20 | 56.0 | 65.1 | 56.6 | - | - | - | 61.1 | 72.1 | 47.3 | - |
TrDiMP(Wang et al. 2021b) | CVPR21 | 63.9 | - | 61.4 | - | - | - | 67.1 | 77.7 | 58.3 | 67.5 |
TransT (Chen et al. 2021) | CVPR21 | 64.9 | 73.8 | 69.0 | - | - | - | 67.1 | 76.8 | 60.9 | 69.1 |
AutoMatch(Zhang et al. 2021) | ICCV21 | 58.3 | - | 59.9 | - | - | - | 65.2 | 76.6 | 54.3 | - |
STARK(Yan et al. 2021) | ICCV21 | 67.1 | 77.0 | - | - | - | - | 68.8 | 78.1 | 64.1 | - |
GTELT(Zhou et al. 2022) | CVPR22 | 67.7 | - | - | 45.0 | 54.2 | 52.2 | - | - | - | - |
AiATrack(Gao et al. 2022) | ECCV22 | 69.0 | 79.4 | 73.8 | - | - | - | 69.6 | 80.0 | 63.2 | 70.6 |
MixFormer-22k (Cui et al. 2022) | CVPR22 | 69.2 | 78.7 | 74.7 | - | - | - | 70.7 | 80.0 | 67.8 | 70.4 |
SimTrack-B/16(Chen et al. 2022) | ECCV22 | 69.3 | 78.5 | - | - | - | - | 68.6 | 78.9 | 62.4 | 69.8 |
OSTrack-384(Ye et al. 2022) | ECCV22 | 71.1 | 81.1 | 77.6 | 50.5 | 61.3 | 57.6 | 73.7 | 83.2 | 70.8 | 70.7 |
VideoTrack(Xie et al. 2023) | CVPR23 | 70.2 | - | 76.4 | - | - | - | 72.9 | 81.9 | 69.8 | 69.7 |
SeqTrack-B384(Chen et al. 2023) | CVPR23 | 71.5 | 81.1 | 77.8 | 50.5 | 61.6 | 57.5 | 74.5 | 84.3 | 71.8 | 68.6 |
ARTrack-384(Xing et al. 2023) | CVPR23 | 72.6 | 81.7 | 79.1 | 51.9 | 62.0 | 58.5 | 75.5 | 84.3 | 74.3 | 70.5 |
EVPTrack-224 | Ours | 70.4 | 80.9 | 77.2 | 48.7 | 59.5 | 55.1 | 73.3 | 83.6 | 70.7 | 70.2 |
EVPTrack-384 | Ours | 72.7 | 82.9 | 80.3 | 53.7 | 65.5 | 61.9 | 76.6 | 86.7 | 73.9 | 70.9 |
Comparison with State-of-the-art Trackers

LaSOT. LaSOT (Fan et al. 2019) test dataset is a challenging long-term tracking benchmark dataset containing 280 video sequences to effectively evaluate the long-term tracking performance of trackers. EVPTrack’s evaluation results are shown in Tab.1, EVPTrack outperforms existing trackers by obtaining 72.7%, 82.9% and 80.3% for AUC, P and respectively. In addition, Fig.5 demonstrates the performance of EVPTrack-384 in different challenge scenarios (e.g., scale variation, aspect ratio change). Illustrating that our SeqTrack-B384 performs better than other competing trackers on almost all challenge scenarios.
LaSOText. (Fan et al. 2021) is an expansion of LaSOT with the addition of 150 videos from 15 new categories. There are many similar interfering objects and fast-moving small objects in the added videos, so it is a challenging dataset. As shown in Tab.1, although EVPTrack-224 has only 48.7% AUC, EVPTrack-384 obtains 53.7% AUC and 61.9% P at larger input resolutions. EVPTrack-384 outperforms EVPTrack-224 by 5% while outperforming all comparative trackers. The reason for this disparity, we consider, is that the low resolution lacks enough background context to distinguish between similar objects.
TrackingNet. TrackingNet (Müller et al. 2018) is a large-scale tracking benchmark, that contains 511 video sequences, which covers diverse object classes and scenes. It does not provide ground-truth and needs to submit the prediction results to its server to get the Success(AUC) and Precision(P and ). The returned results are shown in Tab. 2, our EVPTrack-384 obtains 84.4% , 89.1% and 84.2% in terms of AUC, P, and , respectively. Our EVPTrack-384 can still achieve an advanced performance.
GOT-10k. GOT-10k (Huang, Zhao, and Huang 2021) is a large-scale challenging tracking benchmark with over 10k video segments and has 180 segments for the test set. One of the challenges of this dataset is that there is no overlap between the object classes in the training and test sets, except for the general classes of motion objects and motion patterns. We follow the defined protocol proposed in (Huang, Zhao, and Huang 2021) and train our model only using the GOT-10k training set and submit the tracking output to the official evaluation server. As reported in Tab.1, EVPTrack-384 obtains 76.6% and 86.7% in terms of AO and , respectively. Furthermore, EVPTrack-384 achieved the top-ranked performance on AUC, reaching 76.6%, surpassing all other trackers. This strongly verifies that our method has strong generalization and is not sensitive to categories.
TNL2K. TNL2K is a recently publicly released large-scale tracking dataset, which contains 3000 challenging video sequences: 700 for testing and 2300 for training. We evaluated our tracker with a test set of 700 video sequences. From Tab. 3, we can observe that our tracker achieves a state-of-the-art performance on this dataset. In particular, EVPTrack-224 obtained 57.5% AUC, even more than SeqTrack-384 using a higher resolution.
UAV123. The UAV123 (Mueller, Smith, and Ghanem 2016) contains 123 video sequences that are captured from the UAV platform. It is a challenging aerial video dataset. The benchmarks can be used to assess whether the tracker is suitable for deployment to a UAV123 in real-time scenarios. From Tab. 1, we can observe that our tracker performs the best among all compared trackers.
SiamFC | ECO | SiamPRN++ | ATOM | TransT | STARK | AutoMatch | OSTrack | SimTrack | SeqTrack-B | ARTrack | EVPTrack-224 | EVPTrack-384 | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AUC | 57.1 | 55.4 | 73.3 | 70.3 | 81.4 | 82.0 | 76.0 | 83.9 | 82.3 | 83.9 | 85.1 | 83.5 | 84.4 |
66.3 | 61.8 | 80.0 | 77.1 | 86.7 | 86.9 | - | 88.5 | 86.5 | 88.8 | 89.1 | 88.3 | 89.1 |
SiamFC | ECO | Ocean | SiamBAN | ATOM | TransT | AutoMatch | OSTrack | SimTrack | SeqTrack-B | ARTrack | EVPTrack-224 | EVPTrack-384 | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AUC | 29.5 | 32.6 | 38.4 | 41.0 | 40.1 | 50.7 | 47.2 | 55.9 | 54.8 | 56.4 | 59.8 | 57.5 | 59.1 |
P | 28.6 | - | 37.7 | 41.7 | 39.2 | 51.7 | 43.5 | - | 53.8 | - | - | 58.8 | 62.0 |
# | Method | LaSOT | UAV123 | |
AUC | P | AUC | ||
1 | baseline | 69.4 | 75.9 | 69.0 |
2 | + explicit visual prompts | 70.4 | 77.2 | 70.2 |
3 | + learnable tokens | 69.6 | 76.2 | 69.1 |
4 | + multi-scale prompts | 70.0 | 76.7 | 69.6 |
5 | + spatio-temporal prompts | 69.7 | 76.5 | 69.3 |
6 | + using template token | 69.2 | 76.0 | 69.5 |
7 | + single scale (14x14) | 69.6 | 76.3 | 69.7 |
8 | + single scale (16x16) | 68.9 | 75.5 | 69.4 |
9 | + single scale (18x18) | 69.7 | 76.0 | 69.7 |
# | batchsize | M | N | LaSOT | UAV123 | |
---|---|---|---|---|---|---|
AUC | P | AUC | ||||
1 | 32 | 8 | 4 | 69.9 | 76.6 | 69.6 |
2 | 32 | 4 | 8 | 70.4 | 77.2 | 70.2 |
3 | 32 | 2 | 16 | 69.2 | 76.1 | 69.1 |
Ablation Study and Analysis
We use EVPTrack-224 to validate the effectiveness of our method on the LaSOT and UAV123 benchmarks.
Baseline EVPTrack. In Tab.4, we compare the performance of the baseline tracker #1 using the initial template and ours EVPTrack #2 using explicit visual prompts. The results show that EVPTrack performance improves with the addition of spatio-temporal and multi-scale explicit visual prompts. While the initial template contained reliable targets, it was unable to cope with target target appearance changes during tracking. In contrast, EVPTrack provides multi-scale target information to enhance the target’s features and spatio-temporal information to perceive target appearance changes, which helps the tracker accurately predict the bounding box.
Study on explicit visual prompts. As shown in Tab.4, #3 adds the same number of learnable tokens as explicit visual prompts, which is a type of implicit prompt. Although implicit prompts can improve performance, they are slightly lower than EVPTrack. This is because the implicit prompts does not change during the inference phase, and the tracker relies solely on initial template information and lacks exploitation of spatio-temporal information. In contrast, our display prompt is adaptively generated from templates and spatio-temporal tokens.
Moreover, we conducted experiments #4,#5,#6,#7,#8 and #9 to explore the effectiveness of the multi-scale prompts and the spatio-temporal prompts. We can observe that there is a significant improvement on the performance of multi-scale prompts relative to single-scale prompts. In experiment #4 compared to baseline #1, this improves the performance of the tracker by providing different fine-grained template features. In experiment #5, we trained EVPTrack to use only spatio-temporal prompts. This improves the performance of the tracker by exploiting spatio-temporal information between consecutive frames compared to baseline #1. Comparison of experiments #5 and #6 demonstrates the effectiveness of the spatio-temporal encoder. In Tab.4, the best performance is achieved when both prompts are employed.
Video sequence length during training. As shown in Tab.5, to explore the spatio-temporal information between consecutive frames, we performed experiments related to the length of the video sequence during training. We keep the batchsize, the learning rate, and the total amount of data constant when training the trackers. Comparing experiments #1 and #2, when the sequence length was changed from 4 to 8, the AUC increased by 0.5% on LaSOT. This demonstrates that the tracker can obtain spatio-temporal information from consecutive video frames. Furthermore, comparing experiments #2 and #3, although the sequence length changed from 8 to 16, the AUC decreased by 1.2% on LaSOT, which may be because only two videos were used in each iteration during training, leading to a lack of generalization of the trained tracker. So we can notice a limitation of our approach: the limited GPUs memory restricts the length of the training video which hinders our tracker from fully exploiting the spatio-temporal information.
Speed, FLOPs, and Params. As demonstrated in Tab.6, our EVPTrack-224 can run in real-time at over 71fps. Furthermore, the FLOPs of EVPTrack-384 are 2x lower than SeqTrack-B384, and the AUC of EVPTrack-384 on LaSOT is 1.2% higher than that of SeqTrack-B384. This demonstrates that our method can have the effective benefit of temporal and multi-scale information, as well as a good balance between efficiency and performance.
Model | Params(M) | FLOPs(G) | Speed(FPS) | AUC |
---|---|---|---|---|
SeqTrack-b384 | 89 | 148 | 11 | 71.5 |
EVPTrack-224 | 73 | 21 | 71 | 70.4 |
EVPTrack-384 | 73 | 65 | 28 | 72.7 |

Qualitative comparison. In addition, to demonstrate the effectiveness of our method more intuitively, we provide some qualitative comparison results in Fig.6. For example, in the first sequence, compared to the other three SOTA trackers, our EVPTrack can still obtain accurate tracking results in the situation of having similar objects by utilizing explicit visual prompts.
Conclusions
In this work, we propose a novel explicit visual prompts framework for visual tracking, dubbed EVPTrack. We introduce a novel mechanism of spatio-temporal information propagation which eliminates the need to focus on when-to-update. Furthermore, we abstract spatio-temporal and multi-scale information into explicit visual prompts. These prompts are simply fused with the image tokens via transformer encoders without the need to customize additional modules. Therefore, EVPTrack avoids complicated updating strategies. Extensive experiments conducted on six visual tracking benchmarks have shown that the proposed EVPTrack achieves competitive performance at real-time speed, confirming its effectiveness and efficiency.
Acknowledgements
This work is supported by the National Natural Science Foundation of China (No.U23A20383, 61972167 and U21A20474), the Project of Guangxi Science and Technology (No.2022GXNSFDA035079 and 2023GXNSFDA026003), the Guangxi ”Bagui Scholar” Teams for Innovation and Research Project, the Guangxi Collaborative Innovation Center of Multi-source Information Integration and Intelligent Processing, the Guangxi Talent Highland Project of Big Data Intelligence and Application, and the Research Project of Guangxi Normal University (No.2022TD002).
References
- Bertinetto et al. (2016) Bertinetto, L.; Valmadre, J.; Henriques, J. F.; Vedaldi, A.; and Torr, P. H. S. 2016. Fully-Convolutional Siamese Networks for Object Tracking. In ECCV Workshops, 850–865.
- Cao et al. (2022) Cao, Z.; Huang, Z.; Pan, L.; Zhang, S.; Liu, Z.; and Fu, C. 2022. TCTrack: Temporal Contexts for Aerial Tracking. In CVPR, 14778–14788.
- Chen et al. (2022) Chen, B.; Li, P.; Bai, L.; Qiao, L.; Shen, Q.; Li, B.; Gan, W.; Wu, W.; and Ouyang, W. 2022. Backbone is All Your Need: A Simplified Architecture for Visual Object Tracking. In ECCV (22), 375–392.
- Chen et al. (2023) Chen, X.; Peng, H.; Wang, D.; Lu, H.; and Hu, H. 2023. SeqTrack: Sequence to Sequence Learning for Visual Object Tracking. CVPR, abs/2304.14394.
- Chen et al. (2021) Chen, X.; Yan, B.; Zhu, J.; Wang, D.; Yang, X.; and Lu, H. 2021. Transformer Tracking. In CVPR, 8126–8135.
- Chen et al. (2020) Chen, Z.; Zhong, B.; Li, G.; Zhang, S.; and Ji, R. 2020. Siamese Box Adaptive Network for Visual Tracking. In CVPR, 6667–6676.
- Cui et al. (2022) Cui, Y.; Jiang, C.; Wang, L.; and Wu, G. 2022. MixFormer: End-to-End Tracking with Iterative Mixed Attention. In CVPR, 13598–13608.
- Danelljan et al. (2017) Danelljan, M.; Bhat, G.; Khan, F. S.; and Felsberg, M. 2017. ECO: Efficient Convolution Operators for Tracking. In CVPR, 6931–6939.
- Danelljan et al. (2019) Danelljan, M.; Bhat, G.; Khan, F. S.; and Felsberg, M. 2019. ATOM: Accurate Tracking by Overlap Maximization. In CVPR, 4660–4669.
- Danelljan, Gool, and Timofte (2020) Danelljan, M.; Gool, L. V.; and Timofte, R. 2020. Probabilistic Regression for Visual Tracking. In CVPR, 7181–7190.
- Fan et al. (2021) Fan, H.; Bai, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Harshit; Huang, M.; Liu, J.; Xu, Y.; Liao, C.; Yuan, L.; and Ling, H. 2021. LaSOT: A High-quality Large-scale Single Object Tracking Benchmark. Int. J. Comput. Vis., 439–461.
- Fan et al. (2019) Fan, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Bai, H.; Xu, Y.; Liao, C.; and Ling, H. 2019. LaSOT: A High-Quality Benchmark for Large-Scale Single Object Tracking. In CVPR, 5374–5383.
- Fu et al. (2021) Fu, Z.; Liu, Q.; Fu, Z.; and Wang, Y. 2021. STMTrack: Template-Free Visual Tracking With Space-Time Memory Networks. In CVPR, 13774–13783.
- Gao et al. (2022) Gao, S.; Zhou, C.; Ma, C.; Wang, X.; and Yuan, J. 2022. AiATrack: Attention in Attention for Transformer Visual Tracking. In ECCV (22), 146–164.
- He et al. (2022) He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; and Girshick, R. B. 2022. Masked Autoencoders Are Scalable Vision Learners. In CVPR, 15979–15988.
- Huang, Zhao, and Huang (2021) Huang, L.; Zhao, X.; and Huang, K. 2021. GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild. IEEE Trans. Pattern Anal. Mach. Intell., 43(5): 1562–1577.
- Li et al. (2019) Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; and Yan, J. 2019. SiamRPN++: Evolution of Siamese Visual Tracking With Very Deep Networks. In CVPR, 4282–4291.
- Li et al. (2018) Li, B.; Yan, J.; Wu, W.; Zhu, Z.; and Hu, X. 2018. High Performance Visual Tracking With Siamese Region Proposal Network. In CVPR, 8971–8980.
- Lin et al. (2017) Lin, T.; Goyal, P.; Girshick, R. B.; He, K.; and Dollár, P. 2017. Focal Loss for Dense Object Detection. In ICCV, 2999–3007.
- Lin et al. (2014) Lin, T.; Maire, M.; Belongie, S. J.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; and Zitnick, C. L. 2014. Microsoft COCO: Common Objects in Context. In ECCV, 740–755.
- Loshchilov and Hutter (2019) Loshchilov, I.; and Hutter, F. 2019. Decoupled Weight Decay Regularization. In ICLR.
- Mueller, Smith, and Ghanem (2016) Mueller, M.; Smith, N.; and Ghanem, B. 2016. A Benchmark and Simulator for UAV Tracking. In ECCV, 445–461.
- Müller et al. (2018) Müller, M.; Bibi, A.; Giancola, S.; Al-Subaihi, S.; and Ghanem, B. 2018. TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking in the Wild. In ECCV, 310–327.
- Nam and Han (2016) Nam, H.; and Han, B. 2016. Learning Multi-domain Convolutional Neural Networks for Visual Tracking. In CVPR, 4293–4302.
- Rezatofighi et al. (2019) Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I. D.; and Savarese, S. 2019. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In CVPR, 658–666.
- Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; and Polosukhin, I. 2017. Attention is All you Need. In NIPS, 5998–6008.
- Wang et al. (2021a) Wang, N.; Zhou, W.; Wang, J.; and Li, H. 2021a. Transformer Meets Tracker: Exploiting Temporal Context for Robust Visual Tracking. In CVPR, 1571–1580.
- Wang et al. (2021b) Wang, N.; Zhou, W.; Wang, J.; and Li, H. 2021b. Transformer meets tracker: Exploiting temporal context for robust visual tracking. In CVPR, 1571–1580.
- Xie et al. (2023) Xie, F.; Chu, L.; Li, J.; Lu, Y.; and Ma, C. 2023. VideoTrack: Learning to Track Objects via Video Transformer. In CVPR, 22826–22835.
- Xing et al. (2023) Xing, W.; Yifan, B.; Yongchao, Z.; Dahu, S.; and Yihong, G. 2023. Autoregressive Visual Tracking. In CVPR, 9697–9706.
- Yan et al. (2021) Yan, B.; Peng, H.; Fu, J.; Wang, D.; and Lu, H. 2021. Learning Spatio-Temporal Transformer for Visual Tracking. In ICCV, 10428–10437.
- Ye et al. (2022) Ye, B.; Chang, H.; Ma, B.; Shan, S.; and Chen, X. 2022. Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework. In ECCV (22), 341–357.
- Zhang et al. (2019) Zhang, L.; Gonzalez-Garcia, A.; van de Weijer, J.; Danelljan, M.; and Khan, F. S. 2019. Learning the Model Update for Siamese Trackers. In ICCV, 4009–4018.
- Zhang et al. (2023) Zhang, X.; Tian, Y.; Xie, L.; Huang, W.; Dai, Q.; Ye, Q.; and Tian, Q. 2023. HiViT: A Simpler and More Efficient Design of Hierarchical Vision Transformer. In ICLR.
- Zhang et al. (2021) Zhang, Z.; Liu, Y.; Wang, X.; Li, B.; and Hu, W. 2021. Learn To Match: Automatic Matching Network Design for Visual Tracking. In ICCV, 13319–13328.
- Zhang et al. (2020) Zhang, Z.; Peng, H.; Fu, J.; Li, B.; and Hu, W. 2020. Ocean: Object-Aware Anchor-Free Tracking. In ECCV, 771–787.
- Zhou et al. (2022) Zhou, Z.; Chen, J.; Pei, W.; Mao, K.; Wang, H.; and He, Z. 2022. Global Tracking via Ensemble of Local Trackers. In CVPR, 8751–8760.