This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

DiffRef3D: A Diffusion-based Proposal Refinement Framework for 3D Object Detection

Se-Ho Kim\equalcontrib, Inyong Koo\equalcontrib, Inyoung Lee, Byeongjun Park, Changick Kim
Abstract

Denoising diffusion models show remarkable performances in generative tasks, and their potential applications in perception tasks are gaining interest. In this paper, we introduce a novel framework named DiffRef3D which adopts the diffusion process on 3D object detection with point clouds for the first time. Specifically, we formulate the proposal refinement stage of two-stage 3D object detectors as a conditional diffusion process. During training, DiffRef3D gradually adds noise to the residuals between proposals and target objects, then applies the noisy residuals to proposals to generate hypotheses. The refinement module utilizes these hypotheses to denoise the noisy residuals and generate accurate box predictions. In the inference phase, DiffRef3D generates initial hypotheses by sampling noise from a Gaussian distribution as residuals and refines the hypotheses through iterative steps. DiffRef3D is a versatile proposal refinement framework that consistently improves the performance of existing 3D object detection models. We demonstrate the significance of DiffRef3D through extensive experiments on the KITTI benchmark. Code will be available.

1 Introduction

Denoising diffusion models have demonstrated remarkable performance in the field of image generation (Ho, Jain, and Abbeel 2020; Song, Meng, and Ermon 2020). Recently, the attempts to extend the use of the diffusion model to visual perception tasks (Wang et al. 2023; Lai et al. 2023; Gu et al. 2022; Shan et al. 2023; Chen et al. 2022) have exhibited notable success. However, these successes are limited to perception tasks in image data, while the application to perception tasks involving other modalities (e.g., LiDAR point clouds) remains unsolved. In this paper, we introduce DIffRef3D, a novel framework that utilizes the diffusion process for the task of LiDAR-based 3D object detection.

Refer to caption
Figure 1: Diffusion model for 2D and 3D object detection: (a) A diffusion model where qq and pθp_{\theta} are the forward and reverse diffusion process respectively. (b) Diffusion model for 2D object detection (DiffusionDet). (c) We formulate 3D object detection as a conditional diffusion process based on the region proposal. Noisy boxes (or hypotheses) are colored in yellow, and predictions are colored in green.

DiffRef3D is mostly inspired by DiffusionDet (Chen et al. 2022) which aims to solve 2D object detection by adopting the diffusion process to Sparse R-CNN (Sun et al. 2021). The key concept of DiffusionDet is to scatter random noisy boxes along the image scene and gradually refine the noisy boxes to the object bounding boxes by the reverse diffusion process. However, a straightforward extension of the DiffusionDet pipeline to the 3D domain encounters challenges attributed to the inherent dissimilarity between image and point cloud data. Firstly, since an image is consecutive grid-structured data, meaningful features can be extracted from a random location. In contrast, the random placement of boxes in 3D scenes often falls short of providing any information due to the sparsity of point cloud data caused by occlusion and the inherent limitation of LiDAR sensors. Secondly, while 2D objects have various sizes in an image frame due to perspective, objects in point cloud data are relatively small compared to the 3D scene. Given these characteristics of 3D scenes, scattering noisy boxes for object detection without any prior is akin to finding a needle in a haystack. Inevitably, the diffusion model on 3D point cloud space requires extended prerequisites such as coarse information about object sizes and displacements. Therefore, we design DiffRef3D to be a proposal refinement framework that enhances two-stage detectors, leveraging guidance of the region proposals generated from the first stage.

In DiffRef3D, we formulate the proposal refinement stage of two-stage detection models as a conditional diffusion process. For each proposal, our framework applies a noisy box residual and generates a hypothesis. The model is trained to denoise the residuals by the reverse diffusion process and predict object bounding boxes, using the knowledge of both the proposal and hypothesis regions. Figure 1 highlights conceptual differences between DiffusionDet and DiffRef3D, which are proposed for 2D and 3D object detection tasks, respectively. Specifically, DiffRef3D employs region proposals as the condition for both its forward and reverse diffusion process, as opposed to DiffusionDet, which is built upon an unconditional diffusion process. To this end, we introduce a hypothesis attention module (HAM), which treats features extracted from hypotheses to be conditioned on region proposals. During training, the noisy residuals are generated through the forward diffusion process from the true residuals between the proposals and target object bounding boxes. For inference, the residuals are sampled from a Gaussian distribution, and the model iteratively refines the prediction by denoising the hypotheses. We adopt the iterative sampling method from DDIM (Song, Meng, and Ermon 2020) to predict intermediate hypotheses during inference. DiffRef3D can be implemented with existing two-stage 3D object detectors and consistently improve performances compared to the baseline. We demonstrate the performance and scalability of DiffRef3D through extensive experiments held on the KITTI benchmark (Geiger, Lenz, and Urtasun 2012).

In summary, our main contributions are as follows:

  • DiffRef3D serves as a generalizable proposal refinement framework that can be applied to existing two-stage 3D object detectors, resulting in consistent performance improvements.

  • To the best of our knowledge, DiffRef3D represents the pioneering effort in employing the diffusion process for 3D object detection, interpreting the proposal refinement stage as a conditional diffusion process.

2 Related Works

2.1 LiDAR-based 3D Detection

3D object detectors can be broadly categorized into two methods: point-based methods and voxel-based methods. Point-based methods (Qi et al. 2017a, b) take the raw point clouds as inputs and employ permutation invariant operations to aggregate the point features. Voxel-based methods (Zhou and Tuzel 2018; Yan, Mao, and Li 2018) divide point clouds into a regularly structured grid and apply convolutional operations to extract features. Two-stage models employ the above methods to serve as a region proposal network (RPN), introducing the proposal refinement stage that further improves the prediction quality.

PointRCNN (Shi, Wang, and Li 2019) performs foreground segmentation and generates initial proposals using a PointNet++ (Qi et al. 2017b) backbone, and then extracts RoI features from the point clouds inside the proposals for refinement. PV-RCNN (Shi et al. 2020a) proposes a Voxel Set Abstraction module to encode the voxel-wise feature volumes into a small set of keypoints, then aggregates to the RoI-grid points to take advantage of both point-based and voxel-based methods to refine proposals. Voxel R-CNN (Deng et al. 2021) follows a typical pipeline of anchor-based two-stage models, using SECOND (Yan, Mao, and Li 2018) as RPN and refine proposals with the voxel features pooled for RoI. CenterPoint (Yin, Zhou, and Krahenbuhl 2021) predicts proposals from object centers computed from class-specific heatmap then point features are extracted from 3D centers of each face of the proposals for refinement. CT3D (Sheng et al. 2021) embeds and decodes proposals into effective proposal features based on neighboring point features for box prediction by channel-wise transformer.

There are special cases of two-stage models which utilize additional proposal refinement modules in order to adopt the cascade detection paradigm. 3D Cascade RCNN (Cai et al. 2022) iteratively refines proposals through cascade detection heads while considering point completeness score. CasA (Wu et al. 2022) progressively refines proposals by cascade refinement network while aggregating features from each stage by cascade attention module. The previous cascade detection paradigm boosts performance by leveraging the advantages of multiple predictions such as ensemble, however it lacks flexibility in sampling steps and increases memory usage for additional detection heads. In contrast, iterative sampling of diffusion-based models exploits advantages without increasing memory usage for each sampling step and sampling steps are adjustable without additional training.

2.2 Diffusion Framework on Perception Task

The diffusion framework (Dhariwal and Nichol 2021; Ho, Jain, and Abbeel 2020) has shown remarkable performance in the field of image generation. Notably, DDIM (Song, Meng, and Ermon 2020) expedites the sampling process through the utilization of a more efficient class of iterative implicit probabilistic models. Drawing attention to its denoising capabilities and performance, there have been exploding recent efforts to apply it to perception tasks (Shan et al. 2023; Zou et al. 2023) beyond generation tasks. DiffusionDet (Chen et al. 2022) first adopt the diffusion model to the 2D object detection problem by denoising random boxes to object bounding boxes via the diffusion process. DiffusionInst (Gu et al. 2022) utilizes additional mask branches and instance-aware filters upon DiffusionDet to address the instance segmentation task, and the following works (Lai et al. 2023; Wang et al. 2023) show competitive performance on semantic segmentation. However, diffusion-based perception models have been mostly explored in the image domain. Through DiffRef3D, we demonstrate the first application of the diffusion framework to a 3D perception task that involves point cloud input.

3 Preliminaries

3.1 Proposal Refinement Module

A proposal refinement module in two-stage 3D object detectors takes the initial proposals of the RPN as input and exploits the information of the RoIs to refine them. Formally, the refinement module takes a proposal XPX^{P} and outputs the box residual x^\hat{x} and the confidence score c^\hat{c}:

fP\displaystyle f^{P} =RoIPool(XP),\displaystyle=\textrm{RoIPool}(X^{P}), (1)
x^,c^\displaystyle\hat{x},\hat{c} =Det(fP).\displaystyle=\textrm{Det}(f^{P}). (2)

The RoI feature pooling operation RoIPool()\textrm{RoIPool}(\cdot) extracts proposal feature fPf^{P} from XPX^{P}. Det()\textrm{Det}(\cdot) represents the detection head, which consists of a regression branch producing x^\hat{x} and a classification branch producing c^\hat{c}. x^\hat{x} is supervised with the target residual xgtx^{gt} between the proposal and its target object box XTX^{T}, i.e., xgt=XTXPx^{gt}=X^{T}\ominus X^{P} where \ominus represents box encoding functions (Yan, Mao, and Li 2018), and the target for c^\hat{c} is calculated with the intersection over union (IoU) between XPX^{P} and XTX^{T}.

3.2 Conditional Diffusion Model

During training, a small amount of Gaussian noise is gradually added to the data sample x0x_{0} through a Markovian chain of the forward diffusion process. The forward diffusion process is formulated as:

q(xt|xt1)=𝒩(xt;1βtxt1,βt𝐈),q(x_{t}|x_{t-1})=\mathcal{N}(x_{t};\sqrt{1-\beta_{t}}x_{t-1},\beta_{t}\mathbf{I}), (3)

where βt\beta_{t} represents a noise variance schedule (Ho, Jain, and Abbeel 2020; Nichol and Dhariwal 2021). For brevity, let αt=1βt\alpha_{t}=1-\beta_{t}, αt¯=i=1tαi\bar{\alpha_{t}}=\prod_{i=1}^{t}\alpha_{i}, and a noisy sample xtx_{t} at timestep t,0tT,t,0\leq t\leq T, is derived as follows:

q(xt|x0)=𝒩(xt;αt¯,(1αt¯)𝐈),q(x_{t}|x_{0})=\mathcal{N}(x_{t};\sqrt{\bar{\alpha_{t}}},(1-\bar{\alpha_{t}})\mathbf{I}), (4)
xt=αt¯x0+(1αt¯)ϵ,where ϵ𝒩(0,𝐈).x_{t}=\sqrt{\bar{\alpha_{t}}}x_{0}+(1-\bar{\alpha_{t}})\epsilon,\textnormal{where }\epsilon\sim\mathcal{N}(0,\mathbf{I}). (5)

The conditional diffusion model learns a reverse diffusion process, aiming to progressively refine xtx_{t} back to x0x_{0} with the guidance of the conditional input yy. The reverse diffusion process is formulated as follows:

pθ(x0:T|y)=p(xT)t=1Tpθ(xt1|xt,y).p_{\theta}(x_{0:T}|y)=p(x_{T})\prod^{T}_{t=1}p_{\theta}(x_{t-1}|x_{t},y). (6)

4 DiffRef3D

In this section, we introduce DiffRef3D, a diffusion-based proposal refinement framework for 3D object detection. We formulate the refinement module of two-stage detectors as a conditional diffusion model that takes proposal XPX^{P} as the conditional input and denoises the noisy residual xtx_{t} sampled from timestep tt. DiffRef3D reconstructs the actual residual towards the object x^0(t)\hat{x}_{0}^{(t)} by the reverse diffusion process and also outputs the confidence score c^(t)\hat{c}^{(t)}:

x^0(t),c^(t)=DiffRef3D(xt,t|XP).\hat{x}_{0}^{(t)},\hat{c}^{(t)}=\textrm{DiffRef3D}(x_{t},t|X^{P}). (7)

As shown in Fig. 2, DiffRef3D can be implemented on any existing refinement module that follows the general formulation of Eq. 1, 2 by adding a simple parallel module named the hypothesis attention module (HAM). Details of the HAM architecture, training, and inference procedures are described in the following subsections.

Refer to caption
Figure 2: Diffusion-based Refinement Module. Based on an existing proposal refinement module, DiffRef3D further explores a secondary region (hypothesis) feature by processing through the hypothesis attention module.

4.1 Hypothesis Attention Module

The hypothesis XtHX^{H}_{t} is a secondary inspection region derived by applying the given box residual to the proposal. i.e., XtH=XPxtX^{H}_{t}=X^{P}\oplus x_{t}. As the baseline refinement module extracts the proposal feature fPf^{P} (Eq. 1), HAM also extracts the hypothesis feature ftHf^{H}_{t} using the same RoI feature pooling operation:

ftH=RoIPool(XtH).f^{H}_{t}=\textrm{RoIPool}(X^{H}_{t}). (8)

The hypothesis feature is further processed to incorporate the information of xtx_{t} and tt by applying a self-attention block Attn(Q,K,V)\textrm{Attn}(Q,K,V) and a temporal transformation block TT(,t)\textrm{TT}(\cdot,t).

The self-attention block is a common multi-head attention module (Vaswani et al. 2017) that takes ftHf^{H}_{t} as input for the query, key, and value. Except, we embed xtx_{t} as a positional encoding to the query. We utilize a shallow MLP gϕ()g_{\phi}(\cdot) to convert xtx_{t} as the same dimensional vector with ftHf^{H}_{t}. Formally, the attention feature of the hypothesis atHa^{H}_{t} is computed as:

atH=Attn(ftH+gϕ(xt),ftH,ftH).a^{H}_{t}=\textrm{Attn}(f^{H}_{t}+g_{\phi}(x_{t}),f^{H}_{t},f^{H}_{t}). (9)

atHa^{H}_{t} is a feature vector that represents the information of the hypothesis region with consideration of its relative displacement from the proposal.

The temporal transformation block reweights atHa^{H}_{t} with respect to tt to adjust the effect of the noise signal which gets dominant as timestep increases. Specifically, the output of HAM hth_{t} is formulated as follows:

ht=TT(atH,t)=WtatH+bth_{t}=\textrm{TT}(a^{H}_{t},t)=W_{t}\cdot a^{H}_{t}+b_{t} (10)

where Wt,btW_{t},b_{t} are the scaling and shifting factors derived from an MLP sψ()s_{\psi}(\cdot):

[Wt;bt]=sψ(t).[W_{t};b_{t}]=s_{\psi}(t). (11)

Finally, hth_{t} is aggregated with the proposal feature for prediction:

x^0(t),c^(t)=Det(fP+ht).\hat{x}^{(t)}_{0},\hat{c}^{(t)}=\textrm{Det}(f^{P}+h_{t}). (12)

4.2 Training

During the training stage, We set the reconstruction target x0=xgtx_{0}=x^{gt} and sample noisy residual xtx_{t} by the forward diffusion process. The model is trained to perform the reverse diffusion process of reconstructing the true residual, which is identical to refining a hypothesis to the target object bounding box. We elucidate the methodology for generating hypotheses from proposals and present the overall training process of DiffRef3D in pseudo-code through Algorithm 1.

Algorithm 1 DiffRef3D training procedure
1def train(pr_box, gt_box):
2 # pr_box: proposal box [B, N, 7]
3 # gt_box: ground truth box [B, *, 7]
4
5 # Compute residual and class label
6 x_0, c_gt = \
7 assign_target(pr_box, gt_box)
8
9 t = randint(0,T)
10 eps = normal(mean=0,std=1) # [B, N, 7]
11 SNR = 2
12 x_t = sqrt(1 - alpha_bar_t) * x_0 +
13 (1 - alpha_bar_t) * eps / SNR
14
15 # Predict
16 x_hat, c_hat = \
17 DiffRef3D(pr_box, x_t, t)
18
19 # Compute task specific losses
20 reg_loss = get_reg_loss(x_0, x_hat)
21 cls_loss = get_cls_loss(c_gt, c_hat)
22 return reg_loss + cls_loss

Hypothesis generation

From our attempts, spreading random boxes along the point cloud scene to detect objects fails to train 3D object detection by the diffusion process. Therefore, we propose hypothesis generation to map random signals into 3D boxes rather than mapping normalized values directly into a point cloud scene. Utilizing proposals as a reference frame to generate a hypothesis guarantees that random boxes start in close proximity to the potential object locations. Before we apply the forward diffusion process on the target residual xgtx^{gt}, normalization is preceded since the range of the residual differs according to the size of each proposal. Details of the normalization process are provided in supplementary materials. Noise scaled by signal-to-noise ratio (SNR) is applied to x0x_{0} according to Eq. 5 results in noisy residual xtx_{t}. Finally, hypothesis XtHX^{H}_{t} is obtained by applying noisy residual xtx_{t} on proposal XPX^{P}.

Training objectives

Recent diffusion models (Ho, Jain, and Abbeel 2020) are trained with L2 loss between predicted noise and actual noise added to the data sample. However, noise prediction could be replaced by direct prediction of data samples. A number of diffusion-based approaches (Chen et al. 2022; Gu et al. 2022; Ji et al. 2023) on perception tasks report that it is better to predict data samples and train on task-specific losses. Following consensus from previous works, we also trained our model with task-specific losses, e.g., binary cross-entropy loss for classification, and smooth-L1 loss for regression.

4.3 Inference

During the inference stage, DiffRef3D refines the hypothesis generated from random Gaussian noise and follows iterative sampling steps from DDIM (Song, Meng, and Ermon 2020) for progressive refinement. Starting from a random Gaussian noise xT𝒩(0,1)x_{T}\sim\mathcal{N}(0,1), DDIM progressively denoises Gaussian noise into data samples with step size Δt\Delta t, i.e., xTxTΔtx0x_{T}\rightarrow x_{T-\Delta t}\rightarrow...\rightarrow x_{0}. The progressive refinement is formulated as follows:

xtΔt=x^0(t)αtΔt+ϵ(t)1αtΔtσt2+σtϵx_{t-\Delta t}=\hat{x}^{(t)}_{0}\sqrt{\alpha_{t-\Delta t}}+\epsilon^{(t)}\sqrt{1-\alpha_{t-\Delta t}-\sigma^{2}_{t}}+\sigma_{t}\epsilon (13)
ϵ(t)\displaystyle\epsilon^{(t)} =xtαtx^0(t)1αt\displaystyle=\frac{x_{t}-\alpha_{t}\hat{x}^{(t)}_{0}}{\sqrt{1-\alpha_{t}}} (14)
σt\displaystyle\sigma_{t} =(1αtΔt1αt)(1αtαtΔt)\displaystyle=\sqrt{(\frac{1-\alpha_{t-\Delta t}}{1-\alpha_{t}})}\sqrt{(1-\frac{\alpha_{t}}{\alpha_{t-\Delta t}})} (15)

where ϵ(t)\epsilon^{(t)} represents noise term added to data sample and σt\sigma_{t} represents stochastic parameter, respectively. Algorithm 2 describes the inference procedure of DiffRef3D in pseudo-code.

Algorithm 2 DiffRef3D inference procedure
1def inference(pr_box, steps):
2 # pr_box: proposal box [B, N, 7]
3 # steps: sampling steps
4 T = 1000
5
6 t = reversed(linspace(-1,T,steps))
7 ts = list(zip(times[:-1], times[0:]))
8
9 x_t = normal(mean=0,std=1) # [B, N, 7]
10
11 for t_now, t_next in zip(ts):
12 # Predict at timestep t_now
13 x_hat, c_hat = \
14 DiffRef3D(pr_box, x_t, t_now)
15
16 pr_box = \
17 apply_residual(pr_box, x_hat)
18
19 # Estimate x_t at timestep t_next
20 x_t = \
21 ddim_step(x_t, x_hat, t_now, t_next)
22 return x_hat, c_hat

Iterative sampling

Table 1: Comparison with baseline models on the KITTI val set
Method Car 3D APR40 Ped. 3D APR40 Cyc. 3D APR40
Easy Mod. Hard Easy Mod. Hard Easy Mod. Hard
Voxel R-CNN (Deng et al. 2021) 92.24 84.57 82.30 64.97 58.07 52.66 91.83 73.13 68.40
DiffRef3D-V (1 step) 92.53 84.98 82.55 67.32 60.45 55.28 89.13 71.80 67.60
DiffRef3D-V (3 step) 92.89 84.77 82.64 69.70 63.07 57.70 93.33 74.20 69.63
PV-RCNN (Shi et al. 2020a) 92.10 84.36 82.48 64.26 56.67 51.91 88.88 71.95 66.78
DiffRef3D-PV (1 step) 91.67 84.21 81.93 64.56 58.79 54.44 92.37 72.44 68.42
DiffRef3D-PV (3 step) 91.84 84.33 82.15 65.71 59.46 55.13 93.76 73.94 69.60
CT3D (Sheng et al. 2021) 92.34 84.97 82.91 61.05 55.57 51.10 89.01 71.88 67.91
DiffRef3D-T (1 step) 92.99 85.13 82.83 65.09 59.13 54.36 88.20 71.27 67.13
DiffRef3D-T (3 step) 93.28 85.13 82.97 67.49 61.29 56.17 90.92 72.61 68.52

DiffRef3D generates hypotheses in the first sampling step using residuals sampled from a Gaussian distribution. These hypotheses perform bounding box predictions in conjunction with proposal features, and the predicted outcomes serve two purposes: firstly, as the subsequent proposal in the next sampling step, and secondly, they contribute to inferring the next-step hypothesis according to Eq. 13. The iterative sampling steps share similarities with the cascade detection paradigm (Cai et al. 2022; Wu et al. 2022); however, in the case of cascade detection, distinct weights for each iteration’s detection head are required, limiting the number of iterations. In contrast, iterative sampling employs a single-weight detection head, minimizing memory consumption and allowing customizable sampling repetitions.

5 Experiments

In this section, we delve into the intricate implementation details and outline the evaluation setup adopted to assess the performance of DiffRef3D. Moreover, we conduct ablation studies to comprehensively examine each element of DiffRef3D and affirm the validity of our design choices.

5.1 Implementation Details

Architecture

We implement DiffRef3D on three popular two-stage 3D object detectors: Voxel R-CNN (Deng et al. 2021), PV-RCNN (Shi et al. 2020a), and CT3D (Sheng et al. 2021). The models enhanced with DiffRef3D framework are denoted as DiffRef3D-V, DiffRef3D-PV, and DiffRef3D-T, respectively. The model architectures mostly follow the original baseline, and only the HAM module is newly introduced. Given the RoIPool()\textrm{RoIPool}(\cdot) output feature dimension dd, (d=96d=96 for Voxel R-CNN, 128 for PV-RCNN and CT3D), atHa^{H}_{t} and hth_{t} are also dd-dimensional vectors. Attn()\textrm{Attn}(\cdot) is a multi-head attention module with the head number of 8, and gϕ()g_{\phi}(\cdot) is a two-layer MLP with the hidden dimension of dd. sψ()s_{\psi}(\cdot) is also a two-layer MLP, with the hidden dimension of 4d4d.

Training

For consistency, training configurations for the 3D object detectors mirror those reported in the respective baseline works. In the context of the diffusion process, we set the maximum timestep TT to 1000 and we adopt a cosine noise schedule (Nichol and Dhariwal 2021) for the forward diffusion process. From our observations, SNR for signal scaling results in optimal performance at 2. During the training stage, only one hypothesis was generated for each proposal. All models were trained using 4 NVIDIA RTX 3090 GPUs, and the training duration was largely consistent with the baseline models due to the limited computational overhead introduced by the HAM.

Inference

In the inference phase, DiffRef3D can control the number of sampling steps and the number of hypothesis generation for each proposal. However, we observed that increasing the number of hypotheses yields only minimal performance improvement. Therefore, we fixed the number of hypothesis generation for each proposal as one. We also employ an ensemble method to make usage of the output of intermediate steps by averaging the predictions.

5.2 Results on KITTI Benchmark

Table 2: Comparison with state-of-the-art methods on the KITTI test set
Method Car 3D APR40 Ped. 3D APR40 Cyc. 3D APR40
Easy Mod. Hard Easy Mod. Hard Easy Mod. Hard
SECOND (Yan, Mao, and Li 2018) 83.34 72.55 65.82 48.73 40.57 37.77 71.33 52.08 45.83
PointPillars (Lang et al. 2019) 82.58 74.31 68.99 51.45 41.92 38.89 77.10 58.65 51.92
PointRCNN (Shi, Wang, and Li 2019) 86.96 76.50 71.39 47.98 39.37 36.01 74.96 58.82 52.53
STD (Yang et al. 2019) 86.61 77.63 76.06 53.08 44.24 41.97 78.89 62.53 55.87
Point-GNN (Shi and Rajkumar 2020) 88.33 79.47 72.29 51.92 43.77 40.14 78.60 63.48 57.08
3DSSD (Yang et al. 2020) 88.36 79.57 74.55 50.64 43.09 39.65 82.48 64.10 56.90
SA-SSD (He et al. 2020) 88.75 79.79 74.16 - - - - - -
PV-RCNN (Shi et al. 2020a) 90.25 81.43 76.82 52.17 43.29 40.29 78.60 63.71 57.65
Part-A2 (Shi et al. 2020b) 87.81 78.49 73.51 53.10 43.35 40.06 79.17 63.52 56.93
Voxel R-CNN (Deng et al. 2021) 90.90 81.62 77.06 - - - - - -
VoTr (Mao et al. 2021) 89.90 82.09 79.14 - - - - - -
SPG (Xu et al. 2021) 90.64 82.66 77.91 - - - 80.21 66.96 63.61
CT3D (Sheng et al. 2021) 87.83 81.77 77.16 - - - - - -
BTCDet (Xu, Zhong, and Neumann 2022) 90.64 82.86 78.09 - - - 82.81 68.68 61.81
PDV (Hu, Kuai, and Waslander 2022) 90.43 81.86 77.36 47.80 40.56 38.46 83.04 67.81 60.46
3D Cascade RCNN (Cai et al. 2022) 90.46 82.16 77.06 - - - - - -
PG-RCNN (Koo et al. 2023) 89.38 82.13 77.33 47.99 41.04 38.71 82.77 67.82 61.25
DiffRef3D-V 90.45 81.29 76.66 46.59 40.55 38.27 80.16 66.61 59.98

The KITTI benchmark (Geiger, Lenz, and Urtasun 2012) is one of the most popular benchmarks for 3D object detection evaluation. The dataset consists of 7,481 LiDAR samples for training and 7,518 LiDAR samples for testing. Training samples are divided into the train set with 3,712 samples and the val set with the remaining 3,769 samples. We present two performance evaluations for comparative analysis: For evaluation on KITTI val set, we compared DiffRef3D-V, DiffRef3D-PV, and DiffRef3D-T trained with the train set with their respective baselines. For KITTI test set, we trained a model with our framework using randomly selected 5,981 samples, validated with the remaining 1,500 samples, and evaluated the performance on the online test leaderboard. The performance were assessed across three classes: car, pedestrian, and cyclist on three different levels (easy, moderate, and hard). The results on both the val and test sets were evaluated using the mean average precision calculated at 40 recall positions.

Comparison with baseline models

Table 1 outlines the efficacy of DiffRef3D implementation across three baseline models on the KITTI val set. We present the performance of the baseline model by incorporating the officially released model weights from each study except Voxel R-CNN which is reported without pedestrian and cyclist classes results. Therefore we reproduced Voxel R-CNN detection results on three classes based on their open-source code. DiffRef3D improves detection performance for pedestrian and cyclist classes for all baseline models and difficulties, while performance remains consistent or decreases slightly in car class. We consider this is due to car class proposal boxes being sufficiently large to exploit local features which results DiffRef3D performs similar to baseline models on car class detection. DiffRef3D with sampling steps of 3 improves the baseline Voxel R-CNN, PV-RCNN, and CT3D with 5.00%, 2.79%, and 5.72% AP (R40), respectively, in the moderate pedestrian class. Moreover, DiffRef3D improves the cyclist detection performance by 1.07%, 1.99%, and 0.73%, respectively. The performance improvements primarily stem from the hypothesis generation, which contributes additional local features from individual proposals, consequently yielding greater enhancements in the pedestrian and cyclist classes. Notably, as the sampling step increases in DiffRef3D, detection performance increases in the trade-off of computation costs. Nevertheless, the performance improvement becomes less pronounced relative to the escalating computational demands as the sampling step is raised. Therefore, we present the results of the sampling steps of 1 and 3 in comparison with baseline models.

Comparison with state-of-the-art methods

To further comparisons with state-of-the-art methods, we report results of DiffRef3D-V on the KITTI test set since it shows the best overall performances among three baseline enhanced models. Therefore we trained DiffRef3D-V on a random split of training samples and reported results on the KITTI test set with a sampling step of 3. Table 2 summarizes the comparison with state-of-the-art methods. DiffRef3D-V shows comparable performance in car and cyclist with other models, while performance lags slightly on pedestrian. Nevertheless, as DiffRef3D is applicable to other two-stage 3D detectors published later, it can consistently show competitive performance with state-of-the-art models.

5.3 Ablation study

We conduct extensive ablation studies on the KITTI val set to verify our design choices of DiffRef3D.

Refer to caption
Figure 3: Visualization of iterative proposal refinement on the KITTI benchmark. DiffRef3D refines the proposals using features from the previous predictions and the hypotheses sampled at the current timestep. Each image shows hypotheses and predictions at (a) t=1000, (b) t=500, (c) t=200, and (d) the final prediction. Hypotheses are colored in yellow, predictions are colored in green, and ground truths are colored in blue.

Effectiveness of HAM and diffusion process

HAM D.P. Sampling Steps
1 2 3
59.62 61.01 61.38
59.22 60.91 61.50
60.45 62.77 63.07
Table 3: Ablation studies on the KITTI val set with AP (R40) for pedestrian in moderate difficulty with different configurations .

To investigate the effects of HAM and the diffusion process (denoted as D.P. in Table 3), we assessed performance by implementing HAM and D.P. with Voxel R-CNN as our baseline model. DiffRef3D-V refers to configuration with HAM and D.P. all enabled. During ablation study, box ensemble is applied as default in multiple sampling steps. Moreover, we also applied iterative update of box predictions on baseline either for fair comparison. Table 3 summarizes the result of the experiment. In the absence of the D.P., which implies that hypotheses are sampled from random Gaussian noise during both training and testing, AP (R40) decreased by 0.4% and 0.1% in sampling steps of 1 and 2, respectively. This suggests hypotheses generated by noise that does not follow the diffusion process distracts baseline refinement module. Integrating all of our proposed components results in AP (R40) improvements compared to baseline with 0.83%, 1.76%, and 2.69% for each sampling steps, respectively. Note that performance improvement introduced by iterative sampling steps is higher with D.P., baseline AP (R40) increased by 0.39% and DiffRef3D-V increased by 2.32% at sampling steps of 2 compared to sampling steps of 1.

Sampling steps

Sampling Steps Car Ped. Cyc. Latency
(ms)
1 92.53 69.15 89.19 72
2 92.86 69.50 93.05 113
3 92.89 69.70 93.33 152
4 92.91 70.04 94.18 185
5 93.02 70.00 92.43 232
Table 4: Ablation studies on the KITTI val set with AP (R40) for easy difficulty with varying sample steps.

We also conduct experiments to investigate the impacts of increasing the number of sampling steps on performance and inference time. As a baseline model, DiffRef3D-V trained in the KITTI train set is employed. Results of the ablation study are summarized in Table 4. Increasing the sampling steps from 1 to 5 provides an AP (R40) gain of 0.49%, 0.85%, and 2.24% in car, cyclist, and pedestrian, respectively. However, AP gain introduced by increasing the number of sampling steps become insignificant compared to increase of latency after sampling steps of 3. Therefore, we select 3 as an optimal number of sampling steps throughout the other experiments.

5.4 Qualitative Analysis

Refer to caption
Figure 4: Norm of scaling factor from temporal transformation. The impact of the hypothesis feature increases as the timestep decreases after temporal transformation.

We visualize the hypothesis-driven box prediction process facilitated by the proposal refinement process of DiffRef3D. Figure 3 illustrates iterative sampling steps of DiffRef3D in each timestep. Visualization result implies the necessity of proposals as conditional inputs due to point cloud sparsity and relatively small object size compared to point cloud scenes. Figure 3 (a) shows hypotheses generated around proposal centers tend to prioritize regions with higher object presence, while hypotheses exhibit less realistic box forms since generated from random Gaussian noise. During the iterative sampling, DiffRef3D progressively refines proposals and hypotheses toward object boxes, as demonstrated in Fig. 3 (b) and (c). This result highlights DiffRef3D successfully employing the diffusion process to iteratively denoise hypotheses and refine proposals.

To verify the temporal transformation block considers the effect of the noise signal, we visualize the norm of the scaling factor along every timestep. Results are illustrated in Fig. 4. We affirm that the norm of the scaling factor is larger at a smaller timestep, implying that the refinement module takes more attention to the hypothesis feature at the later refinement process. Therefore, the temporal transformation block controls the importance of hypothesis features before aggregating it with proposal features.

6 Conclusion

In this article, we propose a novel framework, DiffRef3D, that utilizes the diffusion process for the task of 3D object detection with point clouds. The HAM and hypothesis generation scheme we proposed demonstrated consistent performance improvements across different models by integrating the proposal refinement of a two-stage 3D object detector into the diffusion process. Experiments on the KITTI benchmark show that DiffRef3D can serve as a generalizable proposal refinement framework that can be applied to other two-stage 3D object detectors with simple implementation but effective performance improvement.

References

  • Cai et al. (2022) Cai, Q.; Pan, Y.; Yao, T.; and Mei, T. 2022. 3D Cascade RCNN: High Quality Object Detection in Point Clouds. IEEE Transactions on Image Processing.
  • Chen et al. (2022) Chen, S.; Sun, P.; Song, Y.; and Luo, P. 2022. Diffusiondet: Diffusion model for object detection. arXiv preprint arXiv:2211.09788.
  • Deng et al. (2021) Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; and Li, H. 2021. Voxel r-cnn: Towards high performance voxel-based 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, 1201–1209.
  • Dhariwal and Nichol (2021) Dhariwal, P.; and Nichol, A. 2021. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34: 8780–8794.
  • Geiger, Lenz, and Urtasun (2012) Geiger, A.; Lenz, P.; and Urtasun, R. 2012. Are we ready for autonomous driving? the kitti vision benchmark suite. In 2012 IEEE conference on computer vision and pattern recognition, 3354–3361. IEEE.
  • Gu et al. (2022) Gu, Z.; Chen, H.; Xu, Z.; Lan, J.; Meng, C.; and Wang, W. 2022. Diffusioninst: Diffusion model for instance segmentation. arXiv preprint arXiv:2212.02773.
  • He et al. (2020) He, C.; Zeng, H.; Huang, J.; Hua, X.-S.; and Zhang, L. 2020. Structure aware single-stage 3d object detection from point cloud. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 11873–11882.
  • Ho, Jain, and Abbeel (2020) Ho, J.; Jain, A.; and Abbeel, P. 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33: 6840–6851.
  • Hu, Kuai, and Waslander (2022) Hu, J. S.; Kuai, T.; and Waslander, S. L. 2022. Point density-aware voxels for lidar 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8469–8478.
  • Ji et al. (2023) Ji, Y.; Chen, Z.; Xie, E.; Hong, L.; Liu, X.; Liu, Z.; Lu, T.; Li, Z.; and Luo, P. 2023. Ddp: Diffusion model for dense visual prediction. arXiv preprint arXiv:2303.17559.
  • Koo et al. (2023) Koo, I.; Lee, I.; Kim, S.-H.; Kim, H.-S.; Jeon, W.-j.; and Kim, C. 2023. PG-RCNN: Semantic Surface Point Generation for 3D Object Detection. arXiv preprint arXiv:2307.12637.
  • Lai et al. (2023) Lai, Z.; Duan, Y.; Dai, J.; Li, Z.; Fu, Y.; Li, H.; Qiao, Y.; and Wang, W. 2023. Denoising Diffusion Semantic Segmentation with Mask Prior Modeling. arXiv preprint arXiv:2306.01721.
  • Lang et al. (2019) Lang, A. H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; and Beijbom, O. 2019. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 12697–12705.
  • Mao et al. (2021) Mao, J.; Xue, Y.; Niu, M.; Bai, H.; Feng, J.; Liang, X.; Xu, H.; and Xu, C. 2021. Voxel transformer for 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 3164–3173.
  • Nichol and Dhariwal (2021) Nichol, A. Q.; and Dhariwal, P. 2021. Improved Denoising Diffusion Probabilistic Models.
  • Qi et al. (2017a) Qi, C. R.; Su, H.; Mo, K.; and Guibas, L. J. 2017a. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, 652–660.
  • Qi et al. (2017b) Qi, C. R.; Yi, L.; Su, H.; and Guibas, L. J. 2017b. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in neural information processing systems, 30.
  • Shan et al. (2023) Shan, W.; Liu, Z.; Zhang, X.; Wang, Z.; Han, K.; Wang, S.; Ma, S.; and Gao, W. 2023. Diffusion-Based 3D Human Pose Estimation with Multi-Hypothesis Aggregation. arXiv preprint arXiv:2303.11579.
  • Sheng et al. (2021) Sheng, H.; Cai, S.; Liu, Y.; Deng, B.; Huang, J.; Hua, X.-S.; and Zhao, M.-J. 2021. Improving 3d object detection with channel-wise transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2743–2752.
  • Shi et al. (2020a) Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; and Li, H. 2020a. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10529–10538.
  • Shi, Wang, and Li (2019) Shi, S.; Wang, X.; and Li, H. 2019. Pointrcnn: 3d object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 770–779.
  • Shi et al. (2020b) Shi, S.; Wang, Z.; Shi, J.; Wang, X.; and Li, H. 2020b. From points to parts: 3d object detection from point cloud with part-aware and part-aggregation network. IEEE transactions on pattern analysis and machine intelligence, 43(8): 2647–2664.
  • Shi and Rajkumar (2020) Shi, W.; and Rajkumar, R. 2020. Point-gnn: Graph neural network for 3d object detection in a point cloud. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 1711–1719.
  • Song, Meng, and Ermon (2020) Song, J.; Meng, C.; and Ermon, S. 2020. Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502.
  • Sun et al. (2021) Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. 2021. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 14454–14463.
  • Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. Advances in neural information processing systems, 30.
  • Wang et al. (2023) Wang, H.; Cao, J.; Anwer, R. M.; Xie, J.; Khan, F. S.; and Pang, Y. 2023. DFormer: Diffusion-guided Transformer for Universal Image Segmentation. arXiv preprint arXiv:2306.03437.
  • Wu et al. (2022) Wu, H.; Deng, J.; Wen, C.; Li, X.; and Wang, C. 2022. CasA: A Cascade Attention Network for 3D Object Detection from LiDAR point clouds. IEEE Transactions on Geoscience and Remote Sensing.
  • Xu, Zhong, and Neumann (2022) Xu, Q.; Zhong, Y.; and Neumann, U. 2022. Behind the curtain: Learning occluded shapes for 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, 2893–2901.
  • Xu et al. (2021) Xu, Q.; Zhou, Y.; Wang, W.; Qi, C. R.; and Anguelov, D. 2021. Spg: Unsupervised domain adaptation for 3d object detection via semantic point generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 15446–15456.
  • Yan, Mao, and Li (2018) Yan, Y.; Mao, Y.; and Li, B. 2018. Second: Sparsely embedded convolutional detection. Sensors, 18(10): 3337.
  • Yang et al. (2020) Yang, Z.; Sun, Y.; Liu, S.; and Jia, J. 2020. 3dssd: Point-based 3d single stage object detector. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 11040–11048.
  • Yang et al. (2019) Yang, Z.; Sun, Y.; Liu, S.; Shen, X.; and Jia, J. 2019. Std: Sparse-to-dense 3d object detector for point cloud. In Proceedings of the IEEE/CVF international conference on computer vision, 1951–1960.
  • Yin, Zhou, and Krahenbuhl (2021) Yin, T.; Zhou, X.; and Krahenbuhl, P. 2021. Center-Based 3D Object Detection and Tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 11784–11793.
  • Zhou and Tuzel (2018) Zhou, Y.; and Tuzel, O. 2018. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, 4490–4499.
  • Zou et al. (2023) Zou, J.; Zhu, Z.; Ye, Y.; and Wang, X. 2023. DiffBEV: Conditional Diffusion Model for Bird’s Eye View Perception. arXiv preprint arXiv:2303.08333.