This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Department of Computer Science and Engineering,
Southern University of Science and Technology, Shenzhen, China
22institutetext: University of Birmingham, Birmingham, United Kingdom 33institutetext: University of Electronic Science and Technology of China , Chengdu, China

Learning Dual-Fused Modality-Aware Representations for RGBD Tracking

Shang Gao 11    Jinyu Yang 1122    Zhe Li 11    Feng Zheng Corresponding author.11    Aleš Leonardis 22    Jingkuan Song 33
Abstract

With the development of depth sensors in recent years, RGBD object tracking has received significant attention. Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference. However, some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored. On the other hand, some methods attempt to fuse the two modalities by treating them equally, resulting in the missing of modality-specific features. To tackle these limitations, we propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking. The first fusion module focuses on extracting the shared information between modalities based on cross-modal attention. The second aims at integrating the RGB-specific and depth-specific information to enhance the fused features. By fusing both the modality-shared and modality-specific information in a modality-aware scheme, our DMTracker can learn discriminative representations in complex tracking scenes. Experiments show that our proposed tracker achieves very promising results on challenging RGBD benchmarks.

Keywords:
RGBD tracking, object tracking, multi-modal learning

1 Introduction

Object tracking is to localize an arbitrary object in a video sequence, given only the object description in the first frame. It can be applied in lots of applications in video surveillance, autonomous driving [18, 35, 23], and robotics [19]. Recent years witness the development of RGBD (RGB+Depth) object tracking thanks to the affordable and accurate depth cameras. RGBD tracking aims to track the objects more robustly and accurately with the help of depth information, even in color-failed scenarios, e.g., target occlusion and dark scenes.

Refer to caption
Figure 1: Visual comparison of our proposed DMTracker and the state-of-the-art DeT [31]. Our tracker is more robust and accurate on handling the tracking challenges, e.g., (a) background clutter, (b) dark scenes.

Compared to conventional RGB-based tracking, the major difficulty of RGBD tracking is the modality discrepancy resulting from intrinsically distinct information obtained from different modalities. For accurate and robust tracking, existing RGBD trackers have many attempts on either RGB and depth fusion [16], or RGB and depth feature extraction [25]. For example, early RGBD trackers fuse RGB and depth HoG features by using different weights [29]. Camplani et al. [3] applied the Kernel Correlation Filter (KCF) in RGB and depth maps, respectively. With the boosting of deep learning, researchers combine depth information with pre-trained RGB trackers to improve tracking performance. DAL [24] utilizes depth information to enhance RGB features via depth-aware convolutions. Very recently, Yan et al. proposed DeT [31], which utilized the depth colormaps for feature extraction and fused color and depth features via a simple operation like element-wise maximum.

However, on the one hand, a simple modality fusion strategy generally treats the two modalities equally. Thus, some modality-specific features which are strongly complementary cannot be discovered. For example, such fusing operations have side effects on the color modality. As RGB frames are informative, preserving the color and texture information is important for tracking. On the other hand, separately using RGB and depth information may lose the shared information between different modalities. Thus, with simply-fused features only, or using color and depth features separately, the upper bound of the discrimination ability of the feature representation is limited. Up to now, few methods explicitly exploit both intra- and inter-modality characteristics for RGBD tracking. Such deficiencies impede the tracking performance, and thus designing a high-performance RGBD tracker is still challenging.

Existing works in multi-modal learning demonstrate that models can benefit from exploring both the shared information and modality-specific characteristics [11]. Inspired by this, in this paper, we aim to tackle the above deficiencies by proposing Dual-fused Modality-aware Tracker for RGBD object tracking, namely DMTracker. The proposed framework consists of dual fusions, which apply the modality similarity and discrepancy for a robust feature representation. The first fusion module is the Cross-Modal Integration Module (CMIM), which is based on cross-attention. The second fusion module is Specificity Preserving Module (SPM). Specifically, CMIM explores the affinities between RGB and depth features to obtain shared information between different modalities. The obtained modality-shared features represent the correlation between RGB and depth channels through cross-modal attention. SPM compensate the modality-shared features by fusing the weighted modality-specific features and the shared ones. Thus, the second fusion tries to make up the missing specific information from the individual modalities. In DMTracker, intra-modality and inter-modality characteristics are both preserved through modality-aware fusions for similarity matching between tracking templates and search regions. This scheme can compensate for the lack of modality-specific information and enhance the modality-shared features, which finally improves the overall representation ability. Experiments on popular RGBD tracking benchmarks demonstrate the effectiveness of the proposed method.

Our contributions are three-fold:

  • We propose a new tracking method dedicated to RGBD object tracking. It is capable of end-to-end offline training and real-time applications.

  • We develop two novel modality-aware fusion modules that incorporate both modality-shared and modality-specific information for robust tracking.

  • The proposed DMTracker achieves state-of-the-art performance on existing challenging RGBD tracking benchmarks with running at real-time speed.

2 Related Work

RGBD Object Tracking. Early RGBD trackers devote to applying RGBD fusion on handcrafted features between color and depth channels. PTB [25] presents a hybrid RGBD tracker composed of HOG features, optical flow, and 3D point clouds. An et al. [1] extends KCF by adding the depth channel. Another idea is to utilize the depth information on specific scenarios, e.g., occlusion handling and target re-detection [16, 13], with hand-crafted depth features. In late fused trackers, the decisions from different channels are combined by using weighted summation [29], or multi-step localization [28]. Recently, deep learning-based RGBD trackers appear [24, 34, 31]. Among them, DAL [24] reformulates a deep discriminative correlation filter (DCF) to embed the depth information into deep features. DeT [31] transfers the depth maps to depth colormaps and fuses the color and depth features via a mixing operation, e.g., maximum. However, these methods ignore the modality discrepancy, which decreases the tracking accuracy. Compared with existing methods, our DMTracker eliminates the bias that heterogeneous features learned from different modalities, and exploits the correlation between them for robust tracking.

Multi-modal Learning. Multi-modal learning has attracted more and more attention since a large amount of data can be collected from various sources or sensors. As events or actions can be described by information from multiple modalities, multi-modal learning aims to understand the correlation between different modalities. Early methods directly concatenate the features from the multiple modalities into a feature vector. However, simple concatenations often fail to exploit the complex correlations among different modalities. Therefore, several multi-modal learning methods have been developed to explicitly fuse the complementary information from different modalities to improve model performance [14, 22]. Hu et al. [11] present a shareable and individual multi-view learning strategy to explore more properties of multi-modal data. Besides, Zhou et al. [36] propose SPNet which improves saliency detection performance by exploring both the shared information and modality-specific properties. Chen et al. [7] propose a multi-modal framework with an inter-modal module that learns cross-modal features and an intra-modal module to self-learn feature representations across videos.

Transformer in Tracking. The transformer is originally proposed for natural language processing. With attention-based encoders and decoders, transformers have shown their great potential for vision tasks. Transformers have been applied to generic object tracking with considerable success [6, 30, 4, 33, 15, 27]. Existing transformers for tracking focus on the correlation between target templates and search regions. Typically, the transformer decoder fuses feature from the template and search region by cross-attention layers to obtain discriminative features. For example, TransT [6] is proposed to use cross-attention instead of the traditional correlation operation. Later, STARK is proposed based on DETR [5], utilizing self-attention to learn the relationship between different inputs. Existing works demonstrate the effectiveness of the transformer on track, while they mostly focus on the RGB tracking area. To the best of our knowledge, there are no transformer architectures focusing on cross-modal RGBD tracking. Our proposed method utilizes an attention mechanism between color and depth modalities and shows promising performance.

Refer to caption
Figure 2: Overview of the DMTracker pipeline. Our DMTracker consists of three main components: feature extraction, dual-fused modality-aware network, and prediction head. The dual-fused modality-aware network consists of two novel modules, CMIM (cross-modal integration module) and SPM (specificity preserving module), which are introduced in Section 3.2 and 3.3.

3 Methodology

In this section, we present the cross-modal correlation tracker, i.e. DMTracker. We first describe our DMTracker pipeline in Section 3.1. Next, we present the key modules in the dual-fused modality-aware network in Section 3.2 and 3.3, in which the two main modules will be introduced. Training loss is present in Section 3.4. Finally, we detail our training and inference procedures in Section 3.5.

3.1 Overview

As shown in Figure 2, the proposed DMTracker contains three main components: feature extractor, dual-fused modality-aware network, and prediction head. The feature extractor separately extracts the features from the template and the search region of both RGB and depth modalities. Then, the features from different modalities are enhanced by the proposed dual-fused modality-aware network to be a robust representation. Finally, the prediction head performs the binary classification and bounding box regression on the fused features to predict the bounding boxes.

Feature extractor. Firstly, the feature extractor takes template patches from RGB and depth frames Zrgb,Zdepth3×Hz0×Wz0Z_{rgb},~{}Z_{depth}\in\mathbb{R}^{3\times H_{z_{0}}\times W_{z_{0}}} and corresponding search region patches Xrgb,Xdepth3×Hx0×Wx0X_{rgb},~{}X_{depth}\in\mathbb{R}^{3\times H_{x_{0}}\times W_{x_{0}}}. With the ResNet-50 backbone, the feature maps of the template and the search patch from the two modalities, i.e., Iz,DzC×Hz×WzI_{z},~{}D_{z}\in\mathbb{R}^{C\times H_{z}\times W_{z}} and Ix,DxC×Hx×WxI_{x},D_{x}\in\mathbb{R}^{C\times H_{x}\times W_{x}}, can be obtained. Hz=Hz0nH_{z}=\frac{H_{z_{0}}}{n}, Wz=Wz0nW_{z}=\frac{W_{z_{0}}}{n}, Hx=Hx0nH_{x}=\frac{H_{x_{0}}}{n}, Wx=Wx0nW_{x}=\frac{W_{x_{0}}}{n}, nn is the downsampling factor and CC is the dimension of the feature maps.

Dual-fused modality-aware network. Then, the dual-fused modality-aware network takes the feature maps of the two modalities of RGB and depth under the same branch as input, i.e., Ix,Dx,IzI_{x},D_{x},I_{z}, and DzD_{z}. The whole network consists of two fusion modules, i.e., the Cross-Modal Integration Module (CMIM) and Specificity Preserving Module (SPM). Features are firstly fed into the preliminary fusion in CMIM to obtain the initial fusion features through the modality-aware cross-modal attention. And then the features, and both the RGB and depth features, are fed into the SPM to get the final fused features with preserving the modality-aware specificity. The details of the two modules in the network are introduced in Section 3.2 and 3.3.

Prediction head. After getting the final modality-aware feature representations of templates and search regions, a prediction head [2] is used to obtain the bounding box estimation. For the classification head, the fused template feature maps are input to the model predictor that fully exploits the information from the target and background. The model predictor outputs the weights of the convolution layer, which then performs target classification for the search branch. For the localization head, using the template features and the initial target bounding box, IoU Modulation component computes the modulation vectors carrying the target-specific appearance information. Finally, bounding box estimation is performed by maximizing the IoU prediction based on the modulation vectors and proposal bounding boxed in the search region.

3.2 Cross-Modal Integration Module (CMIM)

The cross-modal integration module is built as shown in Figure 5. In this module, we fuse the cross-modal features from the RGB and depth modalities to learn their shared representation.

Attention. In the attention mechanism, given queries QQ, keys KK and values VV, the attention function is defined by the scale dot-product attention:

Attention(Q,K,V)=Softmax(QKdk)V,Attention(Q,K,V)=Softmax(\frac{QK^{\top}}{\sqrt{d_{k}}})V, (1)

where dkd_{k} is the dimension of the key KK.

It can also be extended to multi-head attention (MHA) as:

MHA(Q,K,V)=Concat(Head1,Headh)WO,Headi=Attention(QWiQ,KWiK,VWiV),\begin{split}MHA(Q,K,V)=Concat(Head_{1},Head_{h})W^{O},\\ Head_{i}=Attention(QW^{Q}_{i},KW^{K}_{i},VW^{V}_{i}),\end{split} (2)

where the projections matrices are WiQ,WiKdmodel×dkW^{Q}_{i},W^{K}_{i}\in\mathbb{R}^{d_{model}\times d_{k}}, WiVdmodel×dvW^{V}_{i}\in\mathbb{R}^{d_{model}\times d_{v}} and WOhdv×dmodelW^{O}\in\mathbb{R}^{hd_{v}\times d_{model}}. By projecting the inputs to different spaces through different linear transformations, the model can learn various data distributions and pay attention to varying levels of information, for more details [26]. In this work, we employ hh = 8, dmodeld_{model} = 256 and dk=dv=dm=32d_{k}=d_{v}=d_{m}=32 as default values.

Cross-Modal Attention (CMA) Block. To model the global information of the dual modalities and minimize the computation of our real-time tracker, we use the lightweight attention architecture named cross-modal attention (CMA) block as shown in Figure 5. We simplify the original transformer, remove the self-attention part, but retain the cross-attention, and feed-forward network because we pay more attention to the interactive fusion between the dual modalities, and it can significantly reduce the amount of computation and improve the speed of the tracker. A feed-forward network (FFN) is a fully connected network that consists of two linear functions with a Relu activate function to enhance the fitting ability of the attention. The attention mechanism can not distinguish the position information of the input feature sequence. Thus, we introduce the spatial positional encoding attracting to input following the setting [5, 6]. Cross-Modal Attention Block can enhance depth features by finding the most related visual clues from RGB features, and the CMA can be formulated as equation(33).

ft=Norm(D+MHA(D,I,I)),F=Norm(ft+FFN(ft)),\begin{split}f_{t}&=Norm(D+MHA(D,I,I)),\\ F&=Norm(f_{t}+FFN(f_{t})),\end{split} (3)

where F,DC×H×WF,D\in\mathbb{R}^{C\times H\times W}.

Cross-Modal Integration. The cross-modal integration module is designed based on the CMA block. The basic idea of our proposed CMIM is to use the attention mechanism to globally model the correlation information between the dual modalities learning the shared representation. Formally, we use I(0),D(0)I^{(0)},D^{(0)} to denote the initial RGB feature and depth feature obtained from the backbone network. First, a 1x1 convolution reduces the channel dimension obtaining two lower dimension feature I(1)I^{(1)} and D(1)D^{(1)}:

I(1),D(1)=(I(0),D(0)),I^{(1)},D^{(1)}=\mathcal{F}(I^{(0)},D^{(0)}), (4)

where I(0),D(0)C×H×W,I(1),D(1)Ci×H×WI^{(0)},D^{(0)}\in\mathbb{R}^{C\times H\times W},I^{(1)},D^{(1)}\in\mathbb{R}^{C_{i}\times H\times W}. We employ Ci=256C_{i}=256 in our implementation. Since the cross-modal attention (CMA) takes a set of feature vectors as input, we flatten I(1)I^{(1)} and D(1)D^{(1)} in spatial dimension, obtaining I(2)Ci×HWI^{(2)}\in\mathbb{R}^{C_{i}\times HW} and D(2)Ci×HWD^{(2)}\in\mathbb{R}^{C_{i}\times HW}. In our scenario, we have dual modality features which can be used to attend to each other using the Dual-fused Modality-aware Network i.e., using RGB feature to enhance depth feature, or using depth feature to enhance RGB feature. In our method, we adopt the former: DD as query, II as key and value, as this essentially puts more emphasis on the RGB feature, if we want to make depth feature to integrate more color or texture features from the RGB feature. After cross-modal attention (CMA) block and a 1x1 convolution, the initial fused RGBD feature F(0)F^{(0)} can be obtained as follows:

F(0)=(CMA(I(2),D(2))),F(0)C×H×W.\begin{split}F^{(0)}&=\mathcal{F}(CMA(I^{(2)},D^{(2)})),\quad F^{(0)}\in\mathbb{R}^{C\times H\times W}.\end{split} (5)
Refer to caption
Figure 3: Cross-Modal Integration Module (CMIM).
Refer to caption
Figure 4: Specificity Preserving Module (SPM).
Refer to caption
Figure 5: Cross-Modal Attention (CMA).

3.3 Specificity Preserving Module (SPM)

The specificity preserving module is built as shown in Figure 5. As we have the shared representations from the fused modality, and the features from the modality-specific feature extractors, we propose SPM that make full use of modality-specific information to further enhance the fusion features. Due to the considerable gap between the depth space information from the depth feature and the visual information from the RGB feature, the initial fused features obtained by the CMIM not only contain the correlation information between the two modalities but are also mixed with a lot of irrelevant interference features. Thus, the initial fused feature first passes through an information filter to weaken the chaotic interference information and enhance the shared characteristics of the dual modalities. Then the specificity of each modality is attached to the filtered features, and the final fused features can be obtained which contain the modality-shared and modality-specific cues. The detailed process is described as follows.

We design a learnable vector 𝒱\mathcal{V} to filter the interference and enhance the shared cues of the initial fused features. \odot means element-wise multiplication. It is then updated with depth-specific features through residual connections:

F(1)=D(0)+𝒱F(0),𝒱C,F(0)C×H×W.F^{(1)}=D^{(0)}+\mathcal{V}\odot F^{(0)},\mathcal{V}\in\mathbb{R}^{C},F^{(0)}\in\mathbb{R}^{C\times H\times W}. (6)

Then the obtained features F(1)F^{(1)} and initial RGB-specific features I(0)I^{(0)} are fused by learnable weights to get the final RGBD feature(77).

F(final)=αI(0)+βF(1),F(final)C×H×W.F^{(final)}=\alpha I^{(0)}+\beta F^{(1)},\quad F^{(final)}\in\mathbb{R}^{C\times H\times W}. (7)

3.4 Training Loss

We train our network with the following function,

total=λcls+bbox.\mathcal{L}_{total}=\lambda\mathcal{L}_{cls}+\mathcal{L}_{bbox}. (8)

Classification loss cls\mathcal{L}_{cls} provides the network the ability to robustly discriminate the target object and background distractors, which is defined as [2].

l(s,z)={szz>Tmax(0,s)zTl(s,z)=\left\{\begin{array}[]{rcl}s-z&&{z>T}\\ max(0,s)&&{z\leq T}\end{array}\right. (9)
cls=1Niteri=0Niter(x,c)Mtestl(xf,zc)2.\mathcal{L}_{cls}=\frac{1}{N_{iter}}\sum_{i=0}^{N_{iter}}\sum_{(x,c)\in M_{test}}\|l(x\ast f,z_{c})\|^{2}. (10)

Here, according to the label confidence zz, the threshold TT defines the target and background regions. NiterN_{iter} refers to the number of iterations. zcz_{c} is the target center label that is set from a Gaussian function. We only penalize positive confidence values for the background if zTz\leq T. If z>Tz>T, we take the difference between them. Here, ff is the wights obtained by the model predictor network.

For bounding box estimation, following [8], it extends the training procedure to image sets by computing the modulation vector on the first frame in MtrainM_{train} and sampling candidate boxes from images in MtestM_{test}. The bounding box loss bbox\mathcal{L}_{bbox} is computed as the mean squared error between the predicted BBox defined by BB and the groundtruth BBox defined by B¯\bar{B} in MtestM_{test},

bbox=MSEiMtest(Bi,B¯i).\mathcal{L}_{bbox}=\mathop{MSE}\limits_{i\in M_{test}}(B_{i},\bar{B}_{i}). (11)

3.5 Implementation Details

Training. Our tracker is trained in an end-to-end fashion. 1) Template and Search Area. Different from DeT [31], we use RGB images and raw depth images as inputs. To adapt the depth image to the pre-trained feature extraction model, we copy the raw depth to stack three channels. To incorporate background information, the template and search region is obtained by adding random translation and cropping a square patch centered on the target, which is 525^{2} of the target area. Then these regions are resized to 288×288288\times 288 pixels. Horizontal flip and brightness jittering are used for data augmentations. 2) Initialization. The backbone ResNet50 is initialized by the pre-trained wights on ImageNet. The weights of IoU modulation and IoU predictor in the regression head are initialized using [8]. The model predictor in the classification head is initialized using [2]. Niter=5,T=0.05.N_{iter}=5,T=0.05. We use 𝒱=0.01\mathcal{V}=0.01 to initialize the information filter. The learnable weights between the two modalities:α\alpha and β\beta are initialized to 0.5. The network is optimized using the AdamW optimizer with a learning rate decay of 0.2 every 15th epoch and weight decay 1e41e-4. The target classification loss weight is set to λ\lambda = 1e21e-2.

Inference. Different from the training phase, we track the target in a series of subsequent frames giving the target position in the first frame. Once the annotated first frame is given, we use the data enhancement strategy to build an initial set of 15 samples to refine the model predictor. With the subsequent frames with high confidence, the model predictor can continuously refine the classification head and enhance the ability to distinguish the target and background. Bounding box estimation is performed as in the training phase.

3.6 Visualization

Refer to caption
Figure 6: Visualization of the attention maps for a representative frame. From left to right, they are input depth and RGB frames, feature maps in the search region after CMIM and SPM, and the output prediction, respectively.

To explore how the dual-fused modality-aware network works in our framework, we visualize the attention maps of the fusion modules in a representative frame, as shown in Figure 6. We visualize the attention maps of the modality shared and specific features from the output of CMIM and SPM, respectively. As shown, we can obtain a very informative feature map after the proposed CMIM. Based on cross-modal attention, the modality-shared feature is generated to make the focus on candidates, i.e., the box and the face. Then the SPM is to use the modality-specific information to suppress unreliable candidates. The combination of two kinds of features in SPM produces high-quality representations. This proves that the shared and specific features can be well fused by our CMIM and SPM, leading to more discriminative feature representations for tracking.

4 Experiments

4.1 Experiment Settings

4.1.1 Datasets.

To validate the effectiveness of the proposed tracker, we evaluate it on three public RGBD tracking datasets, including STC [29], CDTB [17] and DepthTrack [31]. Among them, STC is a compact short-term benchmark containing 36 sequences, while CDTB and DepthTrack are long-term ones containing 80 sequences and 50 sequences, respectively. Furthermore, we carry out an attribute-based evaluation to study the performance under different challenges.

4.1.2 Evaluation metrics.

For STC, a representative short-term evaluation protocol, success rate, is used. Given the tracked bounding box btb_{t} and the groundtruth bounding box bab_{a}, the overlap score is defined as S=btbabtbaS=\frac{b_{t}\cap b_{a}}{b_{t}\cup b_{a}}, which is indeed the intersection and union of two regions. With a series of thresholds from 0 to 1, the area-under-curve (AUC) is calculated as the success rate.

For CDTB and DepthTrack, long-term evaluation, i.e., tracking precision and recall are used [32]. They can be obtained by:

Pr(τθ)=1NpAt(τθ)Ω(At(τθ),Gt),Re(τθ)=1NgGtΩ(At(τθ),Gt),Pr(\tau_{\theta})=\frac{1}{N_{p}}\sum_{A_{t}(\tau_{\theta})\neq\varnothing}\Omega(A_{t}(\tau_{\theta}),G_{t}),~{}~{}Re(\tau_{\theta})=\frac{1}{N_{g}}\sum_{G_{t}\neq\varnothing}\Omega(A_{t}(\tau_{\theta}),G_{t}), (12)

where Pr(τθ)Pr(\tau_{\theta}) and Re(τθ)Re(\tau_{\theta}) denote the precision (Pr) and the recall (Re) over all frames, respectively. F-score then can be obtained by F(τθ)=2Re(τθ)Pr(τθ)Re(τθ)+Pr(τθ)F(\tau_{\theta})=\frac{2Re(\tau_{\theta})Pr(\tau_{\theta})}{Re(\tau_{\theta})+Pr(\tau_{\theta})}. NpN_{p} denotes the number of frames in which the target is predicted visible, and NgN_{g} denotes the number of frames in which the target is indeed visible.

4.1.3 Compared trackers.

We compare the proposed tracker with the following benchmarking RGBD trackers.

  • Traditional RGBD trackers: PT [25], OAPF [21], DS-KCF [3], DS-KCF-shape [10], CA3DMS [16], CSR_RGBD++ [12], STC [29] and OTR [13];

  • Deep RGBD trackers: DAL [24], TSDM [34], and DeT111For a fair comparison, we use DeT-DiMP50-Max checkpoint in all experiments. [31];

  • State-of-the-art RGB trackers: ATOM [8], DiMP [2], PrDiMP [9], KeepTrack [20] and TransT [6].

Table 1: Overall performance on DepthTrack test set [31]. The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF [3] RGBD 0.075 0.077 0.076
DS-KCF-Shape [10] RGBD 0.070 0.071 0.071
CSR_RGBD++ [12] RGBD 0.113 0.115 0.114
CA3DMS [16] RGBD 0.212 0.216 0.214
DAL [24] RGBD 0.478 0.390 0.421
TSDM [34] RGBD 0.393 0.376 0.384
DeT [31] RGBD 0.560 0.506 0.532
DiMP50 [2] RGB 0.387 0.403 0.395
KeepTrack [20] RGB 0.504 0.514 0.508
TransT [6] RGB 0.484 0.494 0.489
DMTracker RGBD 0.619 0.597 0.608
Table 2: Overall performance on CDTB dataset [17]. The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF [3] RGBD 0.036 0.039 0.038
DS-KCF-Shape [10] RGBD 0.042 0.044 0.043
CA3DMS [16] RGBD 0.271 0.284 0.259
CSR_RGBD++ [12] RGBD 0.187 0.201 0.194
OTR [13] RGBD 0.336 0.364 0.312
DAL [24] RGBD 0.661 0.565 0.592
TSDM [34] RGBD 0.578 0.541 0.559
DeT [31] RGBD 0.651 0.633 0.642
ATOM [8] RGB 0.541 0.537 0.539
DiMP50 [2] RGB 0.546 0.549 0.547
DMTracker RGBD 0.662 0.658 0.660
Table 3: Overall performance and attribute-based results on STC [29]. The top 3 results are shown in red, blue, and green.
Attributes
Method Type Success IV DV SV CDV DDV SDC SCC BCC BSC PO
OAPF [21] RGBD 0.26 0.15 0.21 0.15 0.15 0.18 0.24 0.29 0.18 0.23 0.28
DS-KCF [3] RGBD 0.34 0.26 0.34 0.16 0.07 0.20 0.38 0.39 0.23 0.25 0.29
PT [25] RGBD 0.35 0.20 0.32 0.13 0.02 0.17 0.32 0.39 0.27 0.27 0.30
DS-KCF-Shape [10] RGBD 0.39 0.29 0.38 0.21 0.04 0.25 0.38 0.47 0.27 0.31 0.37
STC [29] RGBD 0.40 0.28 0.36 0.24 0.24 0.36 0.38 0.45 0.32 0.34 0.37
CA3DMS [16] RGBD 0.43 0.25 0.39 0.29 0.17 0.33 0.41 0.48 0.35 0.39 0.44
CSR_RGBD++ [12] RGBD 0.45 0.35 0.43 0.30 0.14 0.39 0.40 0.43 0.38 0.40 0.46
OTR [13] RGBD 0.49 0.39 0.48 0.31 0.19 0.45 0.44 0.46 0.42 0.42 0.50
DAL [24] RGBD 0.62 0.52 0.60 0.51 0.62 0.46 0.63 0.63 0.55 0.58 0.57
TSDM [34] RGBD 0.57 0.53 0.42 0.43 0.54 0.37 0.56 0.55 0.41 0.45 0.53
DeT [31] RGBD 0.60 0.39 0.56 0.48 0.58 0.40 0.59 0.62 0.48 0.46 0.54
DiMP [2] RGB 0.60 0.49 0.61 0.47 0.56 0.57 0.60 0.64 0.51 0.54 0.57
PrDiMP [9] RGB 0.60 0.47 0.62 0.47 0.51 0.58 0.64 0.62 0.53 0.54 0.58
KeepTrack [20] RGB 0.62 0.51 0.63 0.47 0.57 0.56 0.61 0.62 0.53 0.52 0.58
DMTracker RGBD 0.63 0.52 0.59 0.46 0.67 0.63 0.61 0.65 0.52 0.52 0.60
Refer to caption
Figure 7: Overall tracking performance is presented as tracking Precision-Recall and F-score Plots on DepthTrack [31].
Refer to caption
Figure 8: Overall tracking performance is presented as tracking Precision-Recall and F-score Plots on CDTB [17].

4.2 Main Results

4.2.1 Quantitative results.

We give the quantitative results of the compared models on DepthTrack, CDTB, and STC in Table 10, 10, and 11, respectively. For DepthTrack and CDTB, corresponding tracking Precision-Recall and F-score Plots are also given in Figure 8 and Figure 8. As shown in Table 10, our method outperforms all of the compared state-of-the-art methods and obtains the best F-score of 0.608 on the DepthTrack dataset. On CDTB, we also obtain the top F-score of 0.660. Compared with the state-of-the-art RGBD tacker DeT, our DMTracker performs 7.6% and 1.8% higher on DepthTrack and CDTB, respectively, according to F-score. DMTracker also outperforms TransT by 11.9% higher on DepthTrack. Overall, the proposed DMTracker provide better accuracy than RGB trackers or advanced RGBD trackers on existing RGBD datasets.

4.2.2 Qualitative results.

Figure 9 shows several representative example results of representative trackers and the proposed tracker on various benchmark datasets. Compared with other RGBD models, we can see that our DMTracker can achieve better tracking performance under multiple difficulties. Sequence 1 shows a scene with background clutter. Our method can accurately distinguish the target object, while other trackers are confused by the clutters. In sequence 2, we show an example with multiple similar objects with fast motion, where it is challenging to accurately track the object. While our DMTracker can track the small-sized target ball robustly. In sequence 3 and 5, we show two examples when there are background clutter or occlusion in outdoor scenarios. Our method locates the objects more accurately compared to other approaches. In sequence 4, under a very dark scene with fast motion of the object, our tracker can robustly track the target object. Finally, It can be seen that some approaches fail to track the rotated object in a dark room, while our DMTracker can produce reliable results. Overall, our tracker can handle various tracking challenges and produce promising results.

Refer to caption
Figure 9: Qualitative results in different challenging scenarios. Seq 1: background clutter; Seq 2: similar objects and fast motion; Seq 3: outdoor scenario and full occlusion; Seq 4: dark scenes and fast motion; Seq 5: outdoor scenario and camera motion; Seq 6: dark scene and target rotation.
Refer to caption
Figure 10: Attribute-based F-score on DepthTrack [31].
Refer to caption
Figure 11: Attribute-based F-score on CDTB [17].

4.2.3 Attribute-based analysis.

We report attribute-based evaluations of representative state-of-the-art algorithms on STC, DepthTrack, and CDTB for in-depth analysis. Success rates according to different attributes on STC are given in Table 11. We can see that our proposed outperforms other compared models on Color Distribution Variation (CDV), Depth Distribution Variation (DDV), Surrounding Color Clutter (SCC) , and Partial Occlusion (PO). Attribute-specific F-scores for the tested trackers on DepthTrack and CDTB are shown in Figure 11 and Figure 11, respectively. As shown, our DMTracker outperforms other trackers on 12 attributes on DepthTrack. Our DMTracker outperforms other trackers on 7 attributes on CDTB. It is obvious that DMTracker can handle the depth-related challenges better, e.g., background clutter (BC), dark scenes (DS), depth change (DC), and similar objects (SO).

4.3 Ablation Study

To verify the relative importance of different key components of our tracker, we conduct ablation studies on the most recent DepthTrack dataset [31].

Table 4: Ablation study on Cross-Modal Attention (CMA) layer numbers.
Method CMA Layer Pr Re F-score
DMTracker 1 0.585 0.571 0.578
DMTracker 2 0.619 0.597 0.608
DMTracker 3 0.570 0.552 0.561

4.3.1 Effects of CMA layers.

To investigate the effects of different numbers of Cross-Modal Attention (CMA) layers, we compare tracking performance with different settings. Table 4 shows the comparison results using different numbers of CMA layers, i.e., 1, 2, and 3. From the results, we can see that our model with 2 CMA layers obtains better performance. This can be explained that only one CMA layer leads to inadequate learning of the target knowledge and cross-modal correlation, while deep layers face the risks of overfitting.

Table 5: Ablation study on key components in the network.
Base CMIM SPM Pr Re F-score
0.517 0.492 0.504
0.522 0.426 0.469
0.584 0.563 0.574
0.619 0.597 0.608
Table 6: Ablation study on different SPM settings.
Method αI0+βF1\alpha I^{0}+\beta F^{1} αD0+βF1\alpha D^{0}+\beta F^{1} Pr Re F-score
DMTracker 0.576 0.560 0.568
DMTracker 0.619 0.597 0.608
Table 7: Overall performance on DepthTrack test set. The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF RGBD 0.075 0.077 0.076
DS-KCF-Shape RGBD 0.070 0.071 0.071
CSR_RGBD++ RGBD 0.113 0.115 0.114
CA3DMS RGBD 0.212 0.216 0.214
DAL RGBD 0.478 0.390 0.421
TSDM RGBD 0.393 0.376 0.384
DeT RGBD 0.560 0.506 0.532
DiMP50 RGB 0.387 0.403 0.395
KeepTrack RGB 0.504 0.514 0.508
TransT RGB 0.484 0.494 0.489
DMTracker RGBD 0.619 0.597 0.608
Table 8: Overall performance on CDTB dataset . The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF RGBD 0.036 0.039 0.038
DS-KCF-Shape RGBD 0.042 0.044 0.043
CA3DMS RGBD 0.271 0.284 0.259
CSR_RGBD++ RGBD 0.187 0.201 0.194
OTR RGBD 0.336 0.364 0.312
DAL RGBD 0.661 0.565 0.592
TSDM RGBD 0.578 0.541 0.559
DeT RGBD 0.651 0.633 0.642
ATOM RGB 0.541 0.537 0.539
DiMP50 RGB 0.546 0.549 0.547
DMTracker RGBD 0.662 0.658 0.660

4.3.2 Component-wise analysis.

Since the proposed DMTracker is used to fuse cross-modal features and learn a unified representation, we ablate each key component for analysis. The base model is constructed by removing the whole dual-fused modality-aware network. The fused features are obtained by simply obtaining the element-wise addition of RGB and depth feature maps. As shown in Table 6, the tracking performance first degrades with only CMIM or SPM and then highly improves with adding the CMIM and SPM. As we analyzed in Section 3.6, the CMIM will bring very informative and even redundant features which leads to the sub-optimal performance. With only SPM, the interaction of cross-modal information is lacking, so the model does not fully exploit the modality-aware correlation information. Then, by adding the CMIM and SPM, the redundant information is effectively suppressed, shared and specificity information is enhanced and thus the tracking performance improves.

4.3.3 Effects on modalities fusion order in SPM.

In the specificity preserving module (SPM) of our DMTracker, We first add depth-specific features to RGBD features and then fuse them with weighted RGB-specific features. When given different orders of specific modalities, the performance of the tracker degrades. This can be explained by the fact that the RGB features are more informative than depth features, and this part of the information is dominant in the tracking process. Assume the spatial information is added in high weight at the end fusion. In that case, the proportion of visual information in the fusion feature is weakened. Mixing excess spatial information confuses the tracker.

5 Conclusions

In this paper, we proposed a novel DMTracker to learn dual-fused modality-aware representations for RGBD tracking. On the one hand, the proposed cross-modal integration module can learn the shared information between RGB and depth channels through cross-modal attention. On the other hand, the following specificity preserving module can adaptively fuse the shared features and RGB and depth features to get discriminative representations. Extensive experiments validate the superior performance of the proposed DMTracker, as well as the effectiveness of each component.

Acknowledgements

This work is supported by the National Natural Science Foundation of China under Grant No. 61972188 and 62122035.

Table 9: Overall performance on DepthTrack test set. The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF RGBD 0.075 0.077 0.076
DS-KCF-Shape RGBD 0.070 0.071 0.071
CSR_RGBD++ RGBD 0.113 0.115 0.114
CA3DMS RGBD 0.212 0.216 0.214
DAL RGBD 0.478 0.390 0.421
TSDM RGBD 0.393 0.376 0.384
DeT RGBD 0.560 0.506 0.532
DiMP50 RGB 0.387 0.403 0.395
KeepTrack RGB 0.504 0.514 0.508
TransT RGB 0.484 0.494 0.489
DMTracker RGBD 0.619 0.597 0.608
Table 10: Overall performance on CDTB dataset . The top 3 results are shown in red, blue, and green.
Method Type Pr Re F-score
DS-KCF RGBD 0.036 0.039 0.038
DS-KCF-Shape RGBD 0.042 0.044 0.043
CA3DMS RGBD 0.271 0.284 0.259
CSR_RGBD++ RGBD 0.187 0.201 0.194
OTR RGBD 0.336 0.364 0.312
DAL RGBD 0.661 0.565 0.592
TSDM RGBD 0.578 0.541 0.559
DeT RGBD 0.651 0.633 0.642
ATOM RGB 0.541 0.537 0.539
DiMP50 RGB 0.546 0.549 0.547
DMTracker RGBD 0.662 0.658 0.660
Table 11: Overall performance and attribute-based results on STC . The top 3 results are shown in red, blue, and green.
Attributes
Method Type Success IV DV SV CDV DDV SDC SCC BCC BSC PO
OAPF RGBD 0.26 0.15 0.21 0.15 0.15 0.18 0.24 0.29 0.18 0.23 0.28
DS-KCF RGBD 0.34 0.26 0.34 0.16 0.07 0.20 0.38 0.39 0.23 0.25 0.29
PT RGBD 0.35 0.20 0.32 0.13 0.02 0.17 0.32 0.39 0.27 0.27 0.30
DS-KCF-Shape RGBD 0.39 0.29 0.38 0.21 0.04 0.25 0.38 0.47 0.27 0.31 0.37
STC RGBD 0.40 0.28 0.36 0.24 0.24 0.36 0.38 0.45 0.32 0.34 0.37
CA3DMS RGBD 0.43 0.25 0.39 0.29 0.17 0.33 0.41 0.48 0.35 0.39 0.44
CSR_RGBD++ RGBD 0.45 0.35 0.43 0.30 0.14 0.39 0.40 0.43 0.38 0.40 0.46
OTR RGBD 0.49 0.39 0.48 0.31 0.19 0.45 0.44 0.46 0.42 0.42 0.50
DAL RGBD 0.62 0.52 0.60 0.51 0.62 0.46 0.63 0.63 0.55 0.58 0.57
TSDM RGBD 0.57 0.53 0.42 0.43 0.54 0.37 0.56 0.55 0.41 0.45 0.53
DeT RGBD 0.60 0.39 0.56 0.48 0.58 0.40 0.59 0.62 0.48 0.46 0.54
DiMP RGB 0.60 0.49 0.61 0.47 0.56 0.57 0.60 0.64 0.51 0.54 0.57
PrDiMP RGB 0.60 0.47 0.62 0.47 0.51 0.58 0.64 0.62 0.53 0.54 0.58
KeepTrack RGB 0.62 0.51 0.63 0.47 0.57 0.56 0.61 0.62 0.53 0.52 0.58
DMTracker RGBD 0.63 0.52 0.59 0.46 0.67 0.63 0.61 0.65 0.52 0.52 0.60

References

  • [1] An, N., Zhao, X.G., Hou, Z.G.: Online rgb-d tracking via detection-learning-segmentation. In: 2016 23rd International Conference on Pattern Recognition. pp. 1231–1236. IEEE (2016)
  • [2] Bhat, G., Danelljan, M., Gool, L.V., Timofte, R.: Learning discriminative model prediction for tracking. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 6182–6191 (2019)
  • [3] Camplani, M., Hannuna, S.L., Mirmehdi, M., Damen, D., Paiement, A., Tao, L., Burghardt, T.: Real-time rgb-d tracking with depth scaling kernelised correlation filters and occlusion handling. In: BMVC. vol. 4, p. 5 (2015)
  • [4] Cao, Z., Huang, Z., Pan, L., Zhang, S., Liu, Z., Fu, C.: Tctrack: Temporal contexts for aerial tracking. arXiv preprint arXiv:2203.01885 (2022)
  • [5] Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European conference on computer vision. pp. 213–229. Springer (2020)
  • [6] Chen, X., Yan, B., Zhu, J., Wang, D., Yang, X., Lu, H.: Transformer tracking. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 8126–8135 (2021)
  • [7] Chen, Y.W., Tsai, Y.H., Yang, M.H.: End-to-end multi-modal video temporal grounding. Advances in Neural Information Processing Systems 34 (2021)
  • [8] Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: Atom: Accurate tracking by overlap maximization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 4660–4669 (2019)
  • [9] Danelljan, M., Gool, L.V., Timofte, R.: Probabilistic regression for visual tracking. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7183–7192 (2020)
  • [10] Hannuna, S., Camplani, M., Hall, J., Mirmehdi, M., Damen, D., Burghardt, T., Paiement, A., Tao, L.: Ds-kcf: a real-time tracker for rgb-d data. Journal of Real-Time Image Processing 16(5), 1–20 (2016)
  • [11] Hu, J., Lu, J., Tan, Y.P.: Sharable and individual multi-view metric learning. IEEE Transactions on Pattern Analysis and Machine Intelligence pp. 1–1 (2017)
  • [12] Kart, U., Kämäräinen, J.K., Matas, J.: How to make an rgbd tracker? In: ECCVW (2018)
  • [13] Kart, U., Lukežič, A., Kristan, M., Kämäräinen, J.K., Matas, J.: Object tracking by reconstruction with view-specific discriminative correlation filters. In: IEEE Conference on Computer Vision and Pattern Recognition (2019)
  • [14] Kim, J., Ma, M., Pham, T., Kim, K., Yoo, C.D.: Modality shifting attention network for multi-modal video question answering. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 10106–10115 (2020)
  • [15] Lin, L., Fan, H., Xu, Y., Ling, H.: Swintrack: A simple and strong baseline for transformer tracking. arXiv preprint arXiv:2112.00995 (2021)
  • [16] Liu, Y., Jing, X.Y., Nie, J., Gao, H., Liu, J., Jiang, G.P.: Context-aware three-dimensional mean-shift with occlusion handling for robust object tracking in rgb-d videos. IEEE Transactions on Multimedia 21(3), 664–677 (2018)
  • [17] Lukezic, A., Kart, U., Kapyla, J., Durmush, A., Kamarainen, J.K., Matas, J., Kristan, M.: Cdtb: A color and depth visual object tracking dataset and benchmark. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 10013–10022 (2019)
  • [18] Luo, W., Yang, B., Urtasun, R.: Fast and furious: Real time end-to-end 3d detection, tracking and motion forecasting with a single convolutional net. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition. pp. 3569–3577 (2018)
  • [19] Machida, E., Cao, M., Murao, T., Hashimoto, H.: Human motion tracking of mobile robot with kinect 3d sensor. In: 2012 Proceedings of SICE Annual Conference (SICE). pp. 2207–2211. IEEE (2012)
  • [20] Mayer, C., Danelljan, M., Paudel, D.P., Van Gool, L.: Learning target candidate association to keep track of what not to track. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 13444–13454 (2021)
  • [21] Meshgi, K., ichi Maeda, S., Oba, S., Skibbe, H., zhe Li, Y., Ishii, S.: An occlusion-aware particle filter tracker to handle complex and persistent occlusions. Computer Vision and Image Understanding 150, 81–94 (2016). https://doi.org/https://doi.org/10.1016/j.cviu.2016.05.011, https://www.sciencedirect.com/science/article/pii/S1077314216300649
  • [22] Pan, B., Cai, H., Huang, D.A., Lee, K.H., Gaidon, A., Adeli, E., Niebles, J.C.: Spatio-temporal graph for video captioning with knowledge distillation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 10870–10879 (2020)
  • [23] Qi, H., Feng, C., Cao, Z., Zhao, F., Xiao, Y.: P2b: Point-to-box network for 3d object tracking in point clouds. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6329–6338 (2020)
  • [24] Qian, Y., Yan, S., Lukežič, A., Kristan, M., Kämäräinen, J.K., Matas, J.: DAL : A deep depth-aware long-term tracker. In: International Conference on Pattern Recognition (2020)
  • [25] Song, S., Xiao, J.: Tracking revisited using rgbd camera: Unified benchmark and baselines. In: Proceedings of the IEEE international conference on computer vision. pp. 233–240 (2013)
  • [26] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., Polosukhin, I.: Attention is all you need. Advances in neural information processing systems 30 (2017)
  • [27] Wang, N., Zhou, W., Wang, J., Li, H.: Transformer meets tracker: Exploiting temporal context for robust visual tracking. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1571–1580 (2021)
  • [28] Wang, Q., Fang, J., Yuan, Y.: Multi-cue based tracking (2014)
  • [29] Xiao, J., Stolkin, R., Gao, Y., Leonardis, A.: Robust fusion of color and depth data for rgb-d target tracking using adaptive range-invariant depth models and spatio-temporal consistency constraints. IEEE Transactions on Cybernetics 48(8), 2485–2499 (2017)
  • [30] Yan, B., Peng, H., Fu, J., Wang, D., Lu, H.: Learning spatio-temporal transformer for visual tracking. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 10448–10457 (2021)
  • [31] Yan, S., Yang, J., Käpylä, J., Zheng, F., Leonardis, A., Kämäräinen, J.K.: Depthtrack: Unveiling the power of rgbd tracking. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 10725–10733 (2021)
  • [32] Yang, J., Li, Z., Yan, S., Zheng, F., Leonardis, A., Kämäräinen, J.K., Shao, L.: Rgbd object tracking: An in-depth review. arXiv preprint arXiv:2203.14134 (2022)
  • [33] Yu, B., Tang, M., Zheng, L., Zhu, G., Wang, J., Feng, H., Feng, X., Lu, H.: High-performance discriminative tracking with transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 9856–9865 (2021)
  • [34] Zhao, P., Liu, Q., Wang, W., Guo, Q.: Tsdm: Tracking by siamrpn++ with a depth-refiner and a mask-generator. In: 2020 25th International Conference on Pattern Recognition. pp. 670–676. IEEE (2021)
  • [35] Zheng, C., Yan, X., Gao, J., Zhao, W., Zhang, W., Li, Z., Cui, S.: Box-aware feature enhancement for single object tracking on point clouds. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 13199–13208 (2021)
  • [36] Zhou, T., Fu, H., Chen, G., Zhou, Y., Fan, D.P., Shao, L.: Specificity-preserving rgb-d saliency detection. In: International Conference on Computer Vision (ICCV) (2021)