This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Superb-AI, 3D vision team
11email: [email protected]

Spb3DTracker: A Robust LiDAR-Based Person Tracker for Noisy Environment

Eunsoo Im    Changhyun Jee    Jung Kwon Lee
Abstract

Person detection and tracking (PDT) has seen significant advancements with 2D camera-based systems in the autonomous vehicle field, leading to widespread adoption of these algorithms. However, growing privacy concerns have recently emerged as a major issue, prompting a shift towards LiDAR-based PDT as a viable alternative. Within this domain, "Tracking-by-Detection" (TBD) has become a prominent methodology. Despite its effectiveness, LiDAR-based PDT has not yet achieved the same level of performance as camera-based PDT. This paper examines key components of the LiDAR-based PDT framework, including detection post-processing, data association, motion modeling, and lifecycle management. Building upon these insights, we introduce SpbTrack, a robust person tracker designed for diverse environments. Our method achieves superior performance on noisy datasets and state-of-the-art results on KITTI Dataset benchmarks and custom office indoor dataset among LiDAR-based trackers.

Keywords:
Lidar-based Tracker Robust 3D Tracker Person Detection and Tracking autonomous vehicle.

1 Introduction

Lidar-based person detection and tracking (PDT) is crucial for ensuring safety and reliability in various domains, including autonomous driving, industrial safety, and crowd management [35]. While image-based trackers have been widely used, recent concerns about personal information protection have highlighted their limitations [14]. As a result, LiDAR technology has emerged as a promising alternative, as LiDAR data does not contain personally identifiable information, making it impossible to distinguish specific individuals solely from this data [15]. However, LiDAR-based person tracking presents its own set of challenges. To address these, we delve into the tracking module and propose improved algorithms. LiDAR-based PDT generally fall into two categories: neural network-based approaches and tracking-by-detection (TBD) methods. Neural network-based approaches, such as those utilizing Graph Neural Networks (GNNs), offer attractive end-to-end solutions [23]. However, they often require ground truth (GT) labels for object IDs, which are scarce in real-world datasets and challenging to obtain [18]. TBD has emerged as the predominant and most effective paradigm for LiDAR-based tracking in recent years [4]. This paper focuses on enhancing the TBD mechanism due to its practicality and widespread applicability. Traditional TBD tracking systems comprise several modules, including detection, association, motion modeling, and object management [5]. Each module requires careful analysis to adapt to person tracking scenarios. To address these challenges, we propose a comprehensive analysis of these modules, considering various conditions and scenarios. Our research aims to improve the overall performance and reliability of LiDAR-based person tracking systems by optimizing each component of the TBD framework. This work contributes to the advancement of person tracking technology, which has significant implications for safety and efficiency in autonomous driving, industrial environments, and crowded public spaces [36].

Refer to caption
(a) tracklet by associating high score 3D boxes
Refer to caption
(b) Tracklet by associating every 3D boxes

In this paper, we propose several enhancements to the LiDAR-based PDT framework to improve trajectory robustness and continuity. Our approach addresses key limitations in current tracking-by-detection (TBD) methods for LiDAR data. Firstly, we argue that relying solely on distance-based Intersection over Union (IoU) for association is insufficient [3]. We introduce a more robust approach that incorporates object similarity metrics and object dimensions [26]. This enhancement improves the accuracy of object association across frames. Secondly, we address the limitations of standard Kalman filters in handling non-linear motion and poor observation scenarios. person motion is often non-linear, and LiDAR detectors typically operate at 5Hz despite LiDAR sensors publishing raw data at 10Hz [20]. The conventional Kalman filter, with its fixed system covariance, struggles to support non-linear motion and cope with inadequate observations [6]. To overcome this, we propose an adaptive system covariance based on prediction confidence, allowing for more accurate state estimation in challenging scenarios [7]. Thirdly, we redesign the life-cycle management system. Many existing LiDAR-based tracker methods using TBD discard detection 3D boxes with low confidence scores [10]. We challenge this approach, asserting that low-confidence 3D boxes may represent occluded objects. Consequently, our life-cycle management retains these detections to construct more complete trajectories efficiently [24]. This approach improves the recall metric by preserving potential person tracks that might otherwise be discarded as "ghosts". As illustrated in Figure 1(a) 1(b), associating all 3D bounding boxes provides a more effective approach for identifying occluded objects. This method enhances the detection of objects that may be partially hidden or obstructed in the scene [19]. Fourthly, To demonstrate the generalization capability of our algorithm, we conduct experiments using both public datasets KITTI [9] and our own indoor office dataset. This indoor dataset is particularly valuable, as most public LiDAR datasets for person tracking focus on outdoor environments [1]. By testing across diverse scenarios, we aim to evaluate the robustness of our tracking algorithm in different conditions.

2 Related Works

2.1 2D Person Tracking

Many recent person trackers are designed based on end-to-end deep learning networks. A common approach involves integrating recurrent layers with detector modules. For instance, ROLO combines the convolutional layers of YOLO with LSTM recurrent units [16], while TrackR-CNN extends multi-object tracking to include instance segmentation [20]. Tractor++ offers an efficient tracking solution by utilizing bounding box regression to predict object positions in subsequent frames without requiring training on tracking data [2]. Other approaches focus on enhancing tracking performance through various techniques. Deep SORT integrates appearance information with the Simple Online and Realtime Tracking (SORT) technique, employing a recursive Kalman filter and frame-by-frame data association [25] [28]. This method incorporates an offline pre-training stage to learn a deep association metric using large-scale person re-identification datasets [12]. Similarly, some trackers adopt a single-stage approach, learning target detection and appearance embedding in a shared manner, often complemented by Kalman filters for location prediction [22]. While 2D trackers can leverage rich RGB information for appearance models, a resource not typically available in LiDAR-based 3D tracking, they often struggle with accurately representing 3D scene dynamics [3]. Conversely, 3D tracking methods can better handle scale variations and provide more accurate spatial information [9]. Ultimately, the design of person tracking methods should align with the specific characteristics of each modality to achieve optimal results, whether utilizing 2D image data, 3D LiDAR point clouds, or a fusion of multiple sensor inputs [24].

2.2 Lidar-based Tracking

Recent developments in Lidar-Based Tracking (LBT) have explored diverse strategies to enhance tracking performance in dynamic environments [32]. Some approaches focus solely on 3D bounding boxes detected from point clouds, offering rapid processing and comprehensive scene dynamics capture, albeit potentially underutilizing point cloud appearance features [32]. Alternative methods have incorporated custom point cloud features or employed deep networks to estimate 3D bounding boxes from single-view images [32][17]. More recent approaches have shown promising results by combining features extracted from both RGB images and point clouds [34]. Many LBT techniques rely on rule-based components. For instance, AB3DMOT, a widely-used baseline, employs Intersection over Union (IoU) for association and a Kalman filter for motion modeling [32]. Subsequent enhancements have focused on improving the association step, such as substituting IoU with Mahalanobis or L2 distance metrics [34]. Researchers have also recognized the significance of lifecycle management in tracking systems [34]. Additionally, recent studies have investigated the application of graph neural networks (GNN) to address LBT in an end-to-end manner, with a focus on data association and active tracklet classification [27][11]. As the field progresses, there is an increasing need for comprehensive studies on 3D Multi-Object Tracking (MOT) methods to identify areas for improvement and guide future research directions [34]. This systematic approach will help in addressing the challenges unique to LBT and advancing the state-of-the-art in this domain [32][34].

2.3 Kalman Filter

Kalman filters (KF) are widely used in tracking-by-detection methods for object tracking in autonomous driving due to their effectiveness in filtering noise from state estimations while handling noisy measurements [33]. However, standard KFs have limitations in real-world scenarios. They assume linear system models and Gaussian noise distribution, which may not hold in complex environments [33][29]. They struggle with non-linear motion and poor observation conditions due to fixed system covariance, leading to error accumulation in state predictions, especially during periodic occlusions of objects [29]. Additionally, spatial distortions in 3D sensor detections can result in imprecise motion parameter estimations [33]. To address these issues, advanced filtering techniques such as the Unscented Kalman Filter (UKF) and Cubature Kalman Filter (CKF) have been proposed [29][30]. These filters are designed to better handle non-linearities and approximate state distributions more accurately than standard KFs [30]. Furthermore, adaptive filters can adjust system covariance dynamically to better respond to changing environments, improving tracking performance in complex, real-world autonomous driving scenarios [30][31][13].

Refer to caption
Figure 2: The SpbTracker pipeline associates results from 3D detector predictions with predictions from person motion models. The process involves three main cases: 1. Matched Pairs: Detections and trajectories that are initially matched undergo a two-stage association process. If they remain matched after this process, they are processed through D-UKF (Discrete Unscented Kalman Filter) and LPF (Low-Pass Filter) modules. Pairs with scores exceeding the F1-score threshold are then assigned to either active tracklets or candidate tracklets. 2. Unmatched Trajectories: These are processed through CDD (Confidence Decay and Deletion) modules. Their scores are then adjusted based on their distance from the ego vehicle. Trajectories with scores falling below the death threshold are either deleted or moved to the candidate tracklet pool. 3. Unmatched Detections: Among the unmatched detections, those with scores above the F1-score threshold are added to the birth pool or the candidate tracklet pool.

3 SpbTracker

3.1 Detection Process

Our approach diverges from conventional algorithms that typically retain only high-confidence 3D detection boxes. Instead, we implement a more inclusive strategy that avoids strict confidence thresholds [33]. This method aims to preserve potential tracklets and enhance recall performance. In our pipeline, we employ the DSVT (Dynamic Sparse Voxel Transformer) detection model[21] and apply Non-Maximum Suppression (NMS) to refine the detection results. Unlike image-based approaches, our LiDAR-based method focuses on 3D space, simplifying the post-processing steps. By eschewing a fixed confidence threshold, our approach seeks to strike a balance between retaining potentially valid detections and mitigating false positives. This strategy is particularly effective in challenging scenarios where low-confidence detections may still represent genuine objects. As illustrated in Fig. 3, our experiments reveal an inverse relationship between the confidence threshold and tracking performance metrics (AMOTP and AMOTA). Lower thresholds correspond to improved metric scores, underscoring the importance of managing "ghost" objects (false positives) in enhancing recall.

Table 1: Comparison of person AP for detectors trained from scratch and from pretrained models, with and without shuffling. The experiments use the KITTI validation dataset and our custom office dataset.
Method KITTI [AP]\uparrow Office[AP]\uparrow
From scratch 49.32 64.12
From pretrained model (w/o batch shuffle)
48.32 64.21
From pretrained model (batch shuffle)
54.92 75.89
Refer to caption
Figure 3: The graphs illustrate the relationship between correlation score threshold and two metrics: AMOTP (left) and AMOTA (right) of pedestrian. As the threshold increases, both metrics decrease, indicating a decline in tracking performance. This ablation study was conducted using the DSVT detector model on the KITTI validation dataset. The results highlight the importance of carefully selecting the correlation score threshold to optimize tracking accuracy and precision.

Given the limited availability of person data in the KITTI dataset[8], we adopted a strategy to enhance our model’s performance through transfer learning and multi-dataset training. This approach is particularly effective for image-based detection tasks, where pre-trained models often yield superior results. To create a robust baseline model, we utilized a combination of another public datasets including person class. Rather than training sequentially on each dataset, we implemented a batch-wise shuffling technique. This method ensures that during each training iteration, the model is exposed to a diverse range of samples from all datasets simultaneously. Our training procedure can be summarized as follows:

  • We combined samples from KITTI, our custom dataset, and another public dataset.

  • During batch formation, we randomly shuffled samples from all datasets.

  • This shuffled batch was then used for training, allowing the model to learn from diverse data sources concurrently.

After creating this pre-trained model using the multi-dataset approach, we fine-tuned it exclusively on the KITTI dataset. This two-stage process allows our model to benefit from the rich, diverse data of multiple sources while still specializing in the specific characteristics of the KITTI dataset. This methodology aims to address the scarcity of person data in KITTI by leveraging additional datasets, potentially improving the model’s generalization capabilities and performance on the target dataset. Table 1 demonstrates that using a pretrained model with shuffling improves the Average Precision (AP).

3.2 Motion Model

Most previous methods apply simplistic motion models that fail to adequately account for the complexity of a person’s movement. persons can move freely along the X and Y axes and rotate around the Z-axis (yaw direction) according to the right-hand rule. Furthermore, analysis of person detection results reveals the difficulty in distinguishing between front and back views of individuals. Consequently, prediction models relying solely on heading angle may be inaccurate. To address these limitations, we propose formulating the state of an object trajectory as an 11-dimensional vector:

[x,y,z,θ,xv,yv,xa,ya,w,l,h][x,y,z,\theta,x_{v},y_{v},x_{a},y_{a},w,l,h] (1)

(x,y,z)(x,y,z) represents the 3D location of the object’s geometric center.(θ)(\theta) represents the object’s orientation around the Z-axis.(vx,vy)(v_{x},v_{y}) represents the object’s velocity components in the X and Y directions.(ax,ay)(a_{x},a_{y}) represents the object’s acceleration components in the X and Y directions.(w,l,h)(w,l,h) represents the object’s 3D dimensions (width, length, height)

3.3 Dynamic Unscented Kalman Filter

The standard Kalman Filter is widely used in Track-By-Detection (TBD) tracking modules for associating motion model systems with measurement systems (detection systems). Most previous methods assume linear object motion and Gaussian distribution. However, person motion is inherently non-linear, and the typical LiDAR sensor frame rate of 10Hz, coupled with TBD detection module latency of 0.3-0.7 seconds, often violates the Gaussian distribution assumption in the tracking module. To address these issues, we propose a Dynamic Unscented Kalman Filter (D-UKF). We leverage the D-UKF to estimate the trajectory state. The prediction process can be described by:

(χt1,Wt)=(Tt1,Pt1,κ)(\chi_{t-1},W_{t})=(T_{t-1},P_{t-1},\kappa) (2)
(Tt,Pt)=UT(f(χt1),Wt,Q)(T_{t},P_{t})=UT(f(\chi_{t-1}),W_{t},Q) (3)
(Dt,Pz,t)=UT(h(χt1),Wt,R)(D_{t},P_{z,t})=UT(h(\chi_{t-1}),W_{t},R) (4)

where TT denotes the prediction model, DD means the detection model, PP is the covariance matrix of the previous frame, QQ is the prediction model system’s noise, and RR is the detection model system’s noise. f()f(\cdot) represents the motion model, and h()h(\cdot) is the detection model. The motion model is defined as:

xt=xt1+vx,t1Δt+12ax,t1Δt2x_{t}=x_{t-1}+v_{x,t-1}\Delta t+\frac{1}{2}a_{x,t-1}\Delta t^{2} (5)
yt=yt1+vy,t1Δt+12ay,t1Δt2y_{t}=y_{t-1}+v_{y,t-1}\Delta t+\frac{1}{2}a_{y,t-1}\Delta t^{2} (6)
Table 2: Performance Comparison of Filters Using 10Hz Data
Dataset Method sAMOTA\uparrow AMOTA\uparrow IDs\downarrow
KITTI KF 99.12 91.02 9
KITTI UKF 99.28 92.19 8
KITTI D-UKF 99.27 92.12 8
Office KF 89.01 85.33 55
Office UKF 89.21 85.52 52
Office D-UKF 89.12 85.48 52
Table 3: Performance Comparison of Filters Using 5Hz Data
Dataset Method sAMOTA\uparrow AMOTA\uparrow IDs\downarrow
KITTI KF 94.24 88.88 25
KITTI UKF 96.88 89.12 15
KITTI D-UKF 97.15 91.77 13
Office KF 77.98 71.23 104
Office UKF 80.12 76.12 99
Office D-UKF 82.98 78.56 85

Traditional Kalman Filters use fixed measurement system covariances RR. In an ideal scenario, this error would be zero. However, in real-world applications, this is not achievable. Our D-UKF approach, when generating sigma points, spreads these points according to the system covariance. Unlike traditional KF that use fixed system noise, our SpbTracker dynamically adjusts system noise based on confidence of detection. We propose an D-UKF that modifies the measurement system covariance as follows:

Sk=2Li=0Wi[γk|k1(i)zk|k1^][γk|k1(i)zk|k1^]T+RinitS_{k}=\sum_{2L}^{i=0}W_{i}[\gamma^{(i)}_{k|k-1}-\hat{z_{k|k-1}}][\gamma^{(i)}_{k|k-1}-\hat{z_{k|k-1}}]^{T}+R_{init} (7)
νk=zkzk|k1^\nu_{k}=z_{k}-\hat{z_{k|k-1}} (8)
Rk=1confidence[(1α)Rk1+α(νkνkTSk)]R_{k}=\frac{1}{confidence}[(1-\alpha)R_{k-1}+\alpha(\nu_{k}\nu_{k}^{T}-S_{k})] (9)

This adaptive mechanism allows the filter to adjust to varying levels of measurement uncertainty, improving tracking performance across diverse scenarios without manual parameter adjustments. By incorporating this approach, our method aims to enhance robustness and generalizability in PDT, particularly in challenging environments with complex motion patterns and varying sensor characteristics. Tables 2 and 3 demonstrate the effectiveness of our Dynamic UKF in noisy environments. We conducted an ablation study using the KITTI validation dataset. For the 5Hz dataset, we analyzed every other data point in the sequence to maintain consistency.

3.4 Association

The Generalized 3D Generalized IoU (GIoU) is a widely adopted association metric in object tracking. This metric enhances the traditional IoU (Intersection over Union) by accounting for the spatial relationship between non-overlapping bounding boxes, thereby providing a more comprehensive measure of object overlap. However, GIoU solely reflects the proportion of overlap, neglecting crucial factors such as volume size and object viewpoint. To address these limitations, we incorporate the concept of Complete-IoU (CIoU) into our association framework. CIoU extends the capabilities of GIoU by considering additional geometric properties, thus offering a more nuanced approach to object association. We suggest new metric by modifying CIoU metric as Modified Complete-IoU(MCIoU). MCIoU is specifically designed to reflect the unique characteristics of person tracking, particularly taking into account the ratio of person height:

v=4.0π(arctanhsAreasarctanhtAreat)v=\frac{4.0}{\pi}(arctan\frac{h_{s}}{Area_{s}}-arctan\frac{h_{t}}{Area_{t}}) (10)
α=v(v1GIoU+1)\alpha=v(\frac{v}{1-GIoU}+1) (11)
MCIoU=GIoU+αMCIoU=GIoU+\alpha (12)

Nevertheless, IoU-based association methods inherently face challenges in Re-Identification (Re-ID) tasks, particularly when dealing with objects that have been absent from the scene for extended periods. This limitation arises from the reliance on threshold distances for matching, which can fail to associate objects that have undergone significant spatial displacement over time. To mitigate this issue, we augment our association strategy by incorporating feature similarity measures. This additional dimension allows for more robust matching, especially in scenarios where spatial proximity alone is insufficient. we utilize bird’s eye view(BEV) features sampled via region of interest(ROI) as feature similarity. Considering that the matched object ROI features of the two branches, we design the feature similarity(FS) as:

FS=exp(fs(ft)Tfs2ft2)FS=exp\left(\frac{f_{s}\left(f_{t}\right)^{T}}{\left\|f_{s}\right\|_{2}\left\|f_{t}\right\|_{2}}\right) (13)
Refer to caption
Figure 4: Comparison of association metrics

Our final association metric is derived from a weighted sum of the MCIoU score and FS measure. This composite approach leverages the strengths of both geometric and appearance-based cues, resulting in a more robust and versatile association framework capable of handling diverse tracking scenarios.

ωMCIoU+(1ω)FS,(0<ω<1)\omega MCIoU+(1-\omega)FS,(0<\omega<1) (14)

Fig. 4 presents a comparison of association metrics on the KITTI validation set. The optimal method is positioned closest to the bottom-left corner, indicating the lowest number of ID-Switches and the highest AMOTA score. Our proposed metric demonstrates superior performance among the compared methods, as evidenced by the plot.

3.5 Life Cycle Management

While previous approaches often rely exclusively on high-confidence 3D bounding boxes, our method incorporates all predicted 3D boxes to enhance recall metrics. To address the challenges associated with this comprehensive approach, we adopt a holistic strategy as follows: Firstly, we utilize the F1 score as the classification criterion. We associate the track pool with all detection results using our previously described association method. This approach, however, can potentially lead to decreased precision due to the inclusion of low-confidence detections and increased computational complexity in the association step, which employs the Hungarian algorithm. In contrast to conventional algorithms that use distance-based thresholds for association, we employ a Feature Similarity (FS) metric. Among pairs exceeding a distance threshold, we compare their FS values. Pairs with FS values above a predetermined threshold are retained in the pair memory pool. Matched pairs undergo processing through a Dynamic Unscented Kalman Filter (UKF) and a Low-Pass Filter (LPF) for confidence score smoothing. The LPF mitigates rapid fluctuations in confidence scores, applying to both matched trajectory scores and detection result scores. If the LPF result exceeds the F1-score threshold, the corresponding tracklet is classified as active. The LPF is defined as:

LPF=ωTscore+(1ω)Dscore,(0<ω<1)LPF=\omega T_{score}+(1-\omega)D_{score},(0<\omega<1) (15)
Refer to caption
Figure 5: Long-term tracking persistence

where TscoreT_{score} represents the matched trajectory confidence score, and DscoreD_{score} denotes the matched detection result confidence score. For unmatched detection results, if the score exceeds the F1-score threshold, the detection is either initialized as a new tracklet or classified as a candidate trajectory. Unmatched trajectories undergo score decay based on their distance from the ego-vehicle(or lidar), reflecting the increased likelihood of object disappearance at greater distances. Trajectories with scores falling below the death threshold are removed from the manager memory. The Confidence Decay Distance (CDD) is calculated as:

CDD=(xtxego)2+(ytyego)2+(ztzego)2MAXRANGECDD=\frac{\sqrt{(x_{t}-x_{ego})^{2}+(y_{t}-y_{ego})^{2}+(z_{t}-z_{ego})^{2}}}{MAX\>RANGE} (16)

Fig. 5 demonstrates the long-term tracking persistence capability of our life-cycle memory approach. In frame 2, SpbTracker successfully tracks a person labeled as ID-2. From frames 3 to 20, the object becomes occluded and temporarily disappears from view. Despite this extended period of occlusion, SpbTracker maintains the object’s identity in memory. When the object reappears in frame 21, SpbTracker correctly recognizes it as ID-2, avoiding an identity switch (IDsw). This description effectively conveys the key aspects of your tracking system’s performance, highlighting its ability to maintain object identity through extended periods of occlusion.

4 Experiments

4.1 Experimental Setup

We implemented our proposed methodology using Python and C++, and conducted experiments on an Intel Xeon Gold 5218R CPU (2.10GHz) with 125GB RAM. All reported results reflect online performance.

4.1.1 Datasets

We evaluated our method on two datasets: the KITTI tracking dataset, and a our own dataset. The KITTI tracking dataset was recorded in Karlsruhe, Germany, using a 64-beam LiDAR sensor at a 10Hz frame rate. Following the convention in , we used sequences 1, 6, 8, 10, 12, 13, 14, 15, 16, 18, and 19 as the validation set. Our custom Office dataset was captured using a Hesai QT128 LiDAR sensor. This sensor features a vertical Field of View (FOV) of 105.2° [-52.6°, +52.6°], a horizontal FOV of 360°, a range of 0.05 to 50 meters, and operates at a frame rate of 10Hz. We collected person tracking data in an indoor environment using this LiDAR sensor.

Table 4: A comparison of existing algorithms for the KITTI 2D Pedestrian Tracking task on the test set is presented, using the 2D Pedestrian Tracking Metric. In the results table, the best performance for each metric is highlighted in red, while the second-best performance is marked in blue.

Model Paper Input Method sMOTA\uparrow MOTA\uparrow MOTP\uparrow HOTA\uparrow IDSW\downarrow MSAMOT Sensors ’22 2D + 3D TBD 24.16 47.86 64.35 44.73 209 FNC2 TIV ’24 2D + 3D TBD 30.86 56.05 65.68 46.55 335 EagerMOT ICRA ’21 2D + 3D TBD 28.01 49.82 64.42 39.38 496 StrongFusion Sensors ’22 2D + 3D TBD 16.54 39.04 63.89 43.42 316 JRMOT IROS ’20 2D + 3D TBD 29.78 45.31 72.22 34.24 631 AB3MOT IROS ’20 3D TBD 20.80 38.13 64.54 37.81 879 PolarMOT ECCV ’22 3D GNN 24.61 46.98 64.59 43.59 270 Ours - 3D TBD 32.72 53.55 65.28 43.25 200

4.1.2 Evaluation Metrics and Criteria

For the KITTI dataset, we conducted 3D MOT evaluation on the validation set, as the test set only supports 2D MOT evaluation and its ground truth is not publicly available. We followed the KITTI convention, presenting results for Pedestrian. Regarding matching criteria, we adopted the convention from the KITTI 3D object detection benchmark, using 3D IoU to determine successful matches. Specifically, we employed 3D IoU thresholds (IoU thres) of 0.25 for pedestrian.

On KITTI, we also report sAMOTA [56] and HOTA [31] metrics, which allows us to analyze detection and association errors separately. We also report the number of identity switches (IDs) at the best performing recall.

4.2 Benchmark results

Table 5: KITTI 3D Pedestrian validation set benchmark
Model Input sAMOTA\uparrow MOTA\uparrow recall\uparrow IDs\downarrow
EagerMOT 2D+3D 92.95 93.14 93.61 36
PolarMOT 3D 94.08 93.48 93.66 9
AB3MOT 3D 73.18 66.98 72.82 1
AB3MOT* 3D 88.56 82.14 88.13 87
ours 3D 99.27 94.02 97.80 8

4.2.1 KITTI 3D Pedestrian validation set benchmark

In Table 5, we present a summary of the pedestrian tracking results. Our proposed system demonstrates superior performance compared to other modern tracking algorithms. Across all metrics, with the exception of ID switches (IDs), our method achieves the best results. It’s worth noting that AB3DMOT shows the lowest number of ID switches. However, this is primarily due to its significantly lower detection capability. To provide a fair comparison, we modified AB3DMOT to use the same detector as our SpbTracker, denoted as AB3DMOT*. With this modification, AB3DMOT* exhibits the highest number of ID switches.

Table 6: Custom Office Dataset, 3D person validation set benchmark
Model Input sAMOTA\uparrow MOTA\uparrow recall\uparrow IDs\downarrow
PolarMOT 3D 89.12 85.48 89.16 52
AB3MOT* 3D 88.42 85.77 90.11 88
ours 3D 92.12 88.70 91.98 48

4.2.2 Office 3D person tracking

In Table 6, AB3DMOT* denotes the AB3DMOT algorithm using our detector instead of its original one. The office dataset presents a more challenging environment compared to KITTI due to the presence of crowded objects. Additionally, the LiDAR sensor used in our office dataset has a low range than KITTI’s sensor and the frame rate is lower. These factors contribute to the overall lower performance results compared to those obtained on the KITTI dataset. It’s important to note that all tracking algorithm parameters, including those for our method and the other algorithms evaluated, remain consistent with the settings used in the KITTI experiments. This comparison highlights the robustness of our tracking system across different environmental conditions and sensor configurations, while also demonstrating the challenges posed by more complex scenarios such as crowded indoor environments.

4.2.3 KITTI 2D Pedestrian Tracking

In Table 4, we present the 2D Pedestrian tracking results obtained on the KITTI test set. Although our tracking is performed in 3D space, we can report 2D tracking results by projecting the 3D bounding boxes onto the image plane using camera intrinsic and extrinsic parameters. We then report the minimal axis-aligned 2D bounding boxes that fully enclose these projections as the tracks’ 2D positions. Despite focusing on 3D tracking and using 2D detections only as a secondary cue, our method achieves state-of-the-art results in terms of sMOTA, MOTA, MOTP, and IDSW metrics for the 3D modality(except for models considering only 2D image space). Notably, our sMOTA and IDSW metrics are state-of-the-art even when compared to 2D+3D multi-modal models. This performance demonstrates the effectiveness of our 3D tracking approach, which can compete with and even surpass methods that directly operate in 2D or use both 2D and 3D information. Our results highlight the potential of 3D-focused tracking methods in achieving high performance across both 3D and 2D evaluation metrics.

5 Conclusion

In this paper, we analyzed the "TBD" 3D tracking algorithm and proposed several enhancements for improved person tracking. Our contributions include a person-biased detector, MCIoU, feature similarity measures, a specific person motion model, a robust filter, and long-term life-cycle memory, all of which lead to significant performance improvements. However, the use of large-scale memory in the life-cycle model results in increased matching time. To address this, we optimized the code using C++, but further research is needed to develop a more compact algorithm. Future work will explore learning-based methods for end-to-end tracking with multi-modal, moving away from the rule-based TBD approach.

References

  • [1] Behley, J., Garbade, M., Milioto, A., Quenzel, J., Behnke, S., Stachniss, C., Gall, J.: Semantickitti: A dataset for semantic scene understanding of lidar sequences. Proceedings of the IEEE International Conference on Computer Vision 2019, 9297–9307 (2019). https://doi.org/10.1109/ICCV.2019.00937
  • [2] Bertasius, G., Torresani, L., Shi, J.: Object detection in video with spatiotemporal sampling networks. Proceedings of the European Conference on Computer Vision (ECCV) 2018, 331–346 (2018). https://doi.org/10.1007/978-3-030-01219-8_20
  • [3] Bewley, A., Ge, Z., Ott, L., Ramos, F., Upcroft, B.: Simple online and realtime tracking. In: 2016 IEEE International Conference on Image Processing (ICIP). pp. 3464–3468 (2016). https://doi.org/10.1109/ICIP.2016.7533003
  • [4] Cai, S., Zheng, W., Chen, H., Zhu, X., Liu, L., Zhang, L.: 3d multi-object tracking: A baseline and new evaluation metrics. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 11414–11424 (2021). https://doi.org/10.1109/CVPR46437.2021.01125
  • [5] Caltagirone, L., Bellone, M., Svensson, L., Wahde, M.: A survey on 3d lidar-based person detection and tracking. Sensors 20(20),  5770 (2020). https://doi.org/10.3390/s20205770
  • [6] Cho, H., Rybski, P.E., Bar-Hillel, A., Pomerleau, D.: Real-time pedestrian detection with deformable part models. Proceedings of the IEEE Intelligent Vehicles Symposium 2014, 1035–1042 (2014). https://doi.org/10.1109/IVS.2014.6856433
  • [7] Fritsch, J., Kuehnl, T., Geiger, A.: A new performance measure and evaluation benchmark for road detection algorithms. Proceedings of the International Conference on Intelligent Transportation Systems 2019, 1693–1699 (2019). https://doi.org/10.1109/ITSC.2013.6728474
  • [8] Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: The kitti dataset. The International Journal of Robotics Research 32(11), 1231–1237 (2013)
  • [9] Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? the kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 3354–3361 (2013). https://doi.org/10.1109/CVPR.2012.6248074
  • [10] Kim, A., Kim, J., Yoo, H.: Robust object tracking with adaptive detection and appearance matching. IEEE Transactions on Intelligent Transportation Systems 22(1), 354–367 (2021). https://doi.org/10.1109/TITS.2020.2967051
  • [11] Kim, A., Brasó, G., Ošep, A., Leal-Taixé, L.: Polarmot: How far can geometric relations take us in 3d multi-object tracking? In: European Conference on Computer Vision. pp. 41–58. Springer (2022)
  • [12] Leal-Taixé, L., Milan, A., Reid, I., Roth, S., Schindler, K.: Tracking the trackers: An analysis of the state of the art in multiple object tracking. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 1831–1840 (2016). https://doi.org/10.1109/CVPR.2015.7298790
  • [13] Lim, E.: Pose estimation of a drone using dynamic extended kalman filter based on a fuzzy system. In: 2021 9th International Conference on Control, Mechatronics and Automation (ICCMA). pp. 141–145. IEEE (2021)
  • [14] Ling, Z., He, D., Zhang, J., Wang, X.: Privacy-preserving human sensing for smart cities: A survey. IEEE Communications Surveys & Tutorials 22(2), 1196–1223 (2020). https://doi.org/10.1109/COMST.2019.2959041
  • [15] Liu, H., Wu, B., Sun, M., Hu, X., Hu, B., Liu, S., Lu, Q.: 3d lidar-based static and moving object detection in driving environments: A review. IEEE Intelligent Transportation Systems Magazine 10(4), 103–114 (2018). https://doi.org/10.1109/MITS.2018.2873567
  • [16] Ning, G., Zhang, Z., Huang, C., Ren, Z., Wang, H., Cai, X., He, Z.: Spatially supervised recurrent convolutional neural networks for visual object tracking. In: 2017 IEEE International Symposium on Circuits and Systems (ISCAS). pp. 1–4 (2017). https://doi.org/10.1109/ISCAS.2017.8050479
  • [17] Pang, Z., Li, Z., Wang, N.: Simpletrack: Understanding and rethinking 3d multi-object tracking. In: European Conference on Computer Vision. pp. 680–696. Springer (2022)
  • [18] Sattler, T., Maddern, W., Toft, C., Torii, A., Hammarstrand, L., Stenborg, E., Safari, D., Okutomi, M., Pollefeys, M., Sivic, J., Pajdla, T., Kahl, F., Leonard, J.: Benchmarking 6dof outdoor visual localization in changing conditions. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 8601–8610 (2018). https://doi.org/10.1109/CVPR.2018.00897
  • [19] Sun, P., Kretzschmar, H., Droz, P., Frossard, D., Casser, V., Leitner, M., Mahjourian, R., McAllister, R., Cohen, A., Zhang, B., Ondruska, P., Omari, S., Maksai, A., Texeira, M., Pollefeys, M., Funkhouser, T., Urtasun, R., Chouard, A., Sceats, C.: Scalability in perception for autonomous driving: Waymo open dataset. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020, 2446–2454 (2020). https://doi.org/10.1109/CVPR42600.2020.00252
  • [20] Voigtlaender, P., Krause, M., Osep, A., Luiten, J., Sekar, B.S., Geiger, A., Leibe, B.: Mots: Multi-object tracking and segmentation. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 7942–7951 (2019). https://doi.org/10.1109/CVPR.2019.00813
  • [21] Wang, H., Shi, C., Shi, S., Lei, M., Wang, S., He, D., Schiele, B., Wang, L.: Dsvt: Dynamic sparse voxel transformer with rotated sets. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 13520–13529 (2023)
  • [22] Wang, Y., Zhang, L., Wu, J., Zha, Z.J.: Learning multi-object tracking with multi-scale shared networks. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 10551–10560 (2020). https://doi.org/10.1109/CVPR42600.2020.01056
  • [23] Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph cnn for learning on point clouds. In: Proceedings of the ACM SIGGRAPH Symposium on Interactive 3D Graphics and Games (I3D). pp. 1–12 (2019). https://doi.org/10.1145/3326362.3326374
  • [24] Weng, X., Wang, Y., Held, D., Kitani, K.: 3d multi-object tracking: A baseline and new evaluation metrics. In: 2019 International Conference on Intelligent Robots and Systems (IROS). pp. 10–15 (2019). https://doi.org/10.1109/IROS.2019.8968044
  • [25] Wojke, N., Bewley, A., Paulus, D.: Simple online and realtime tracking with a deep association metric. In: 2017 IEEE International Conference on Image Processing (ICIP). pp. 3645–3649 (2017). https://doi.org/10.1109/ICIP.2017.8296962
  • [26] Wu, B., Nevatia, R.: Tracking of multiple, partially occluded humans based on static body part detection. Computer Vision and Image Understanding 113(8), 1131–1143 (2019). https://doi.org/10.1016/j.cviu.2009.01.005
  • [27] Yan, Z., Duckett, T., Bellotto, N.: Online learning for human classification in 3d lidar-based tracking. In: 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). pp. 864–871. IEEE (2017)
  • [28] Zhang, Y., Sun, P., Jiang, Y., Yu, D., Weng, F., Yuan, Z., Luo, P., Liu, W., Wang, X.: Bytetrack: Multi-object tracking by associating every detection box. In: European conference on computer vision. pp. 1–21. Springer (2022)
  • [29] Zhang, Y., Gao, F., Shen, S.: Kalman filter with moving reference for jump-free, multi-sensor odometry with application in autonomous driving. IEEE Transactions on Intelligent Transportation Systems (2020)
  • [30] Zhang, Y., Gao, F., Shen, S.: 3d multi-object tracking with adaptive cubature kalman filter for autonomous driving. IEEE Transactions on Intelligent Transportation Systems (2023)
  • [31] Zhang, Y., Gao, F., Shen, S.: Karnet: Kalman filter augmented recurrent neural network for learning world models in autonomous driving tasks. arXiv preprint arXiv:2305.14644 (2023)
  • [32] Zhang, Y., Gao, F., Shen, S.: Lidar-based dense pedestrian detection and tracking. IEEE Transactions on Intelligent Transportation Systems (2023)
  • [33] Zhang, Y., Gao, F., Shen, S.: Probabilistic 3d multi-object cooperative tracking for autonomous driving via differentiable multi-sensor kalman filter. arXiv preprint arXiv:2309.14655 (2023)
  • [34] Zhang, Y., Gao, F., Shen, S.: Strongfusionmot: A multi-object tracking method based on lidar-camera fusion. IEEE Transactions on Intelligent Transportation Systems (2023)
  • [35] Zhao, D., Li, Y.: A survey on 3d lidar localization for autonomous vehicles. IEEE Access 7, 107974–107986 (2019). https://doi.org/10.1109/ACCESS.2019.2931883
  • [36] Zhou, X., Wang, W., Krähenbühl, P.: Online multi-object tracking with association-aware convolutional neural networks. In: Proceedings of the European Conference on Computer Vision (ECCV). pp. 566–582 (2020). https://doi.org/10.1007/978-3-030-58580-8_34