This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

MAFF-Net: Filter False Positive for 3D Vehicle Detection with Multi-modal Adaptive Feature Fusion

Zehan Zhang1,2, Ming Zhang1, Zhidong Liang1, Xian Zhao1, Ming Yang2, Wenming Tan1, and ShiLiang Pu1,∗ Shiliang Pu is the corresponding author.1 The authors are with Hikvision Research Institute, Hangzhou Hikvision Digital Technology Co. Ltd, China (e-mail: zhangzehan, zhangming15, liangzhidong, zhaoxian, tanwenming, [email protected]).2 The authors are with the Department of Automation, Shanghai Jiao Tong University, Shanghai 200240, China (e-mail: [email protected]; [email protected]).
Abstract

3D vehicle detection based on multi-modal fusion is an important task of many applications such as autonomous driving. Although significant progress has been made, we still observe two aspects that need to be further improvement: First, the specific gain that camera images can bring to 3D detection is seldom explored by previous works. Second, many fusion algorithms run slowly, which is essential for applications with high real-time requirements(autonomous driving). To this end, we propose an end-to-end trainable single-stage multi-modal feature adaptive network in this paper, which uses image information to effectively reduce false positive of 3D detection and has a fast detection speed. A multi-modal adaptive feature fusion module based on channel attention mechanism is proposed to enable the network to adaptively use the feature of each modal. Based on the above mechanism, two fusion technologies are proposed to adapt to different usage scenarios: PointAttentionFusion is suitable for filtering simple false positive and faster; DenseAttentionFusion is suitable for filtering more difficult false positive and has better overall performance. Experimental results on the KITTI dataset demonstrate significant improvement in filtering false positive over the approach using only point cloud data. Furthermore, the proposed method can provide competitive results and has the fastest speed compared to the published state-of-the-art multi-modal methods in the KITTI benchmark.

I INTRODUCTION

Background. Since 3D vehicle detection is a crucial part of perception in autonomous driving, a lot of research work has been invested in the industry and academia. 3D detection can be achieved through images, lidar point clouds or multi-modal data. In this study, we consider the fusion problem of lidar point clouds and RGB images. Point clouds provide very accurate depth information, but are accompanied by low resolution and texture information. On the other hand, images have ambiguous depth information but can provide fine-grained texture and color information. This provides an attractive research opportunity for how to design a model that can take full advantages of the two types of data.

Reality. As a pioneering attempt of fusion methods, MV3D proposed a multi-view fusion mechanism to explore multi-modal collaboration methods. Based on this method, AVOD and ContFuse further proposed more diverse fusion pipelines to improve the performance of 3D detection. However, some recently published methods such as PointPillars[1], Second[2], PointRCNN[3], PartA2[4], STD[5] and 3DSSD[6] significantly outperform these methods using only lidar point clouds. In fact, despite some recent fusion studies[7, 8, 9, 10, 11], the top methods on the KITTI[12] leaderboard are still lidar only[13, 14, 15]. Does this mean that the 3D point cloud is sufficient for 3D detection? Or is there any way for RGB images to effectively supplement 3D detection?

Refer to caption
Figure 1: Sample scenes of LIDAR point clouds prone to false detection from KITTI dataset. The 3D structures of the hierarchical trees on the left and objects clusters on the right are similar to that of vehicles, so that they are easy to be misdetected as vehicles. The proposed MAFF-Net will adaptively add RGB image features to the point cloud network, making it easier to distinguish these ambiguous objects.

Analysis. The answer is clear by analyzing the characteristics of point cloud data and RGB image data. Consider the two examples in Fig. 1, where the hierarchical trees in the left and the object clusters in the right are very similar to vehicles in the lidar modality. This makes it difficult to distinguish specific objects using only point clouds. But, the above objects are easy to distinguish in the image, which shows that adding images to the point cloud network can help effectively reduce 3D false positive(FP). However, in the currently published fusion methods, almost no attention has been paid to the specific effects of images for 3D detection, especially in eliminating FP. In this research, we focus on the false positive filtering effect of images for 3D vehicle detection: we hope that when the image is fused with the point cloud, FP can be effectively reduced without reducing true positive(TP).

Challenges. However, how to effectively integrate 2D images into the 3D detection pipeline is still an open question. A naive way is to directly concatenate the raw RGB feature to the 3D point cloud feature, because the point-to-pixel correspondence can be established through projection. However, the image often has noise information such as occlusion and truncation. In these cases, the wrong image feature will be obtained after 3D points are projected onto the image. Therefore, simply using point-wise projected image features will degrade the performance of 3D detection. Moreover, we also notice that the detection speeds of current fusion methods are generally slow. The reasons for the slow speed are: (i) the fusion methods generally adopt a two-stage way that uses ROI-pooling[16] to fuse the image with the point cloud; (ii) the fusion methods usually use 2D detection or segmentation tasks to obtain high-level image features, and then fuse the features with the point cloud. The second way is equivalent to doing a 2D task first and then doing 3D task, which slows down the speed of the 3D detection.

Contributions. In this paper, to address the above challenges, we design a multi-modal adaptive feature fusion network named MAFF-Net based on the successful PointPillars architecture. It is an end-to-end single-stage fusion 3D detection method, which only uses the raw RGB features and does not rely on the high-level features of any 2D tasks. It is also able to eliminate the interference information of two modalities through channel-wise attention module, so that the network can make full use of the advantages of image and point cloud features, to achieve the purpose that reducing FP and reserving TP at the same time.

Specifically, we have developed two fusion technologies: (i) PointAttentionFusion(PAF): This is a relatively fast and simple fusion method where the corresponding projection image features and 3D point features are concatenated to obtain preliminary mixed features, which are used as the input of point-wise channel attention module to learn the channel-wise importance of point cloud and image features. The learned attention features and the original modal features are then concatenated to obtain the final point-wise fusion feature, which is then jointly processed by the PointPillars pipelines. (ii) DenseAttentionFusion(DAF): In this technology, point cloud and image features are divided into three types of pillar-wise features: point cloud feature, image feature, and point cloud image feature that is obtained by appending RGB feature to the point cloud feature. The above three features are concatenated and sent to the pillar-wise channel attention module to generate attention feature. Finally, the attention feature is concatenated with the previous three modalities features to form the pillar-wise fusion feature, which is further used by the subsequent 2D CNN network. Compared with the DenseAttentionFusion, PointAttentionFusion is more concise, so it is faster and more suitable for eliminating some simple false detections. DenseAttentionFusion is a complex and dense fusion method. Its performance is better than PointAttentionFusion and it is more effective for some difficult false detections.

The main contributions of this study are summarized as follows:

  • This paper proposes an end-to-end trainable single-stage multi-modal fusion method for reducing 3D false positive.

  • This paper proposes two different fusion techniques based on channel attention mechanism for different usage scenarios.

  • Experiments on the KITTI dataset demonstrate that the proposed method can effectively reduce false positive while maintaining true positive. It achieves competitive results compared with the published state-of-the-art fusion methods with a fast detection speed of 32 Hz.

  • Extensive analysis and visualization are conducted to understand the design principles of the proposed method and the usefulness of the attention mechanism for multi-modal fusion.

II RELATED WORK

II-A 3D Vehicle Detection based on Point Cloud

3D vehicle detection using only point clouds can be roughly divided into 3 types: Bird’s Eye View(BEV)-based, Voxel-based, and Point-based. BEV-based methods[17, 18] first use rules[19] to convert the point cloud into the BEV form, and then uses the 2D CNN network for learning and prediction[20]. Voxel-based methods divides the point cloud space into regularly arranged voxel blocks, use rules[21, 22, 23] or neural networks[24, 25] to encode the voxels, and then adopt 3D or 2D convolution[2, 26, 4, 27, 28] for 3D detection. PointPillars[1] further simplified voxels into pillars, realizing a real-time one-stage 3D detection method. The voxel-based methods are conducive to accurate 3D proposal generation but are also limited by the receptive field of 2D/3D convolution. The point-based method mainly uses PointNet[29, 30] technology to encode 3D points, and then uses more strategies[3, 5, 6] to process 3D points to achieve the purpose of accurately predicting 3D vehicles. Most of these point-based methods are based on the PointNet series, which makes the receptive fields of point cloud feature learning more flexible. Recently, there has been a trend of fusing point-based and voxel-based methods[15, 13] to utilize the best of two worlds, thereby improving 3D detection performance.

II-B 3D Vehicle Detection based on Multi-modal Fusion

Refer to caption
Figure 2: Overview of the proposed MAFF-Net. The net is an end-to-end point cloud and image fusion method. It uses point cloud and image as input, outputs a set of pillar-wise fusion features through the MAFF module, and then scatter the fusion features back to a 2D pseudo-image for a 2D convolutional neural network according to the spatial relationship of the pillars. The features from the 2D backbone are used by the detection head to predict 3D bounding boxes for vehicles.

In order to take advantages of the camera and lidar sensors, various fusion methods have also been proposed. MV3D[31] is a pioneering work in the fusion method. It combines the lidar BEV representation, front view and image together, and proposes a two-stage network. AVOD[32] fuses the features of the BEV and the image in the middle layer of the convolution to predict 3D objects. ContFuse[33] uses continuous convolution to fuse images and lidar features on different resolutions. MMF adds ground estimation and depth estimation to the fusion framework, and learns better fusion feature representations while jointly learning multi-tasks. [34, 35, 36] first use camera images to generate proposals and then exploit some methods to process the lidar points in these regions to generate 3D objects. [10, 37] use 2D tasks (detection or segmentation) to obtain the feature representations of the image, then fuse these feature representations with the point cloud, and finally use the architecture of VoxelNet to process the fused features.

Although various multi-modal fusion methods have been proposed, they have seldom paid attention to the specific gains brought by the image for 3D detection and the detection speeds are often very slow. In this study, we focus on using images to eliminate the FP of 3D vehicle detection and hope to obtain a fast multi-modal fusion method.

III Proposed Method

In this section, we introduce the proposed single-stage point cloud and image feature fusion 3D vehicle detector MAFF-Net in detail.

III-A Raw RGB Input

Compared with lidar point clouds, images have rich color and texture information, which is very effective for overcoming the false detection of point clouds, especially the background false detection. In this paper, we only use the raw RGB features of the image as the input of MAFF-Net. MAFF-Net does not rely on any 2D annotation information and the high-level features of any 2D tasks.

III-B PointPillars

The PointPillars structure is used as the base 3D detection network for two main reasons: (i) it achieves a good balance between speed and performance: with excellent performance, it has a very fast detection speed. A lot of work in academia and industry is also based on this algorithm[27, 28, 38, 39, 40, 41]. (ii) It provides a natural and effective interface for fusing image features at different granularities in 3D space such as points and pillars. The network used in this study is described in [1]. For completeness, this section briefly reviews PointPillars. This algorithm consists of three modules: (i) a Pillar Feature Net(PFN); (ii) 2D Convolutional Neural Networks; (iii) a Detection Head.

PFN is a feature learning network designed to encode raw point clouds for individual pillars. All non-empty pillars are encoded by PFN and share the same network parameters. The encoded features are scattered back to the original pillar positions to construct a pseudo-image. The pseudo-image features are forwarded through a series of 2D convolutional blocks to extract high-level features. The features are then used by a detection head to generate the targets.

III-C Multimodal Adaptive Feature Fusion(MAFF)

In this study, two concise technologies that can fuse raw RGB data with the point cloud data are proposed to filter false positive in 3D detection.

PointAttentionFusion(PAF): This is an early and simple fusion approach in which each 3D point is aggregated through an image feature and two attention features to achieve the adaptive fusion features. Fig. 3 presents the procedure of this technology.

The method first uses the calibration matrix to project each 3D point onto the image, and the projected image features can be obtained according to the corresponding projection location index. Note that these features are the raw RGB features. In order to make the projected image features and point cloud features compatible, a simple fully connected network called MLPPD\mathrm{MLP_{PD}} is applied to map the image features to appropriate dimensions. The MLPPD\mathrm{MLP_{PD}} is composed of a set of blocks and each block consists of a linear layer, a BN layer, and a ReLU layer. Next, the point cloud features and the mapped image features are concatenated channel-wise to obtain point-wise extended features. However, since the image has many noise factors, such as occlusion, truncation, etc., the extended features will introduce interference information.

Refer to caption
Figure 3: Architecture of the proposed MAFF module with PointAttentionFusion technology. Different colors represent the features of different modalities

To address this issue, we adopt the point-wise channel attention module, which uses expanded features to adaptively estimate the importance of each type of features in a channel-wise manner. First, feed the extended features into a fully connected layer that includes a linear layer, a ReLU layer, and a linear layer. Then output the feature weights through a sigmoid, and finally multiply the weights with the corresponding features in a channel-wise manner to get the attention features. The above process is used to process the point cloud features and image features respectively. The specific forms of channel attention are as follows:

𝐅𝐚.𝐏\displaystyle\mathbf{F_{a.P}} =𝐅𝐏σ(MLPP(𝐅𝐄))Point cloud attention\displaystyle=\mathbf{F_{P}}\otimes\underbrace{\sigma(\mathrm{MLP_{P}}(\mathbf{F_{E}}))}_{\text{Point cloud attention}} (1)
𝐅𝐚.𝐈\displaystyle\mathbf{F_{a.I}} =𝐅𝐈σ(MLPI(𝐅𝐄))Image attention\displaystyle=\mathbf{F_{I}}\otimes\underbrace{\sigma(\mathrm{MLP_{I}}(\mathbf{F_{E}}))}_{\text{Image attention}}

where 𝐅𝐏\mathbf{F_{P}} and 𝐅𝐈\mathbf{F_{I}} represent point cloud and image features, respectively, 𝐅𝐚.𝐏\mathbf{F_{a.P}} and 𝐅𝐚.𝐈\mathbf{F_{a.I}} are the corresponding point-wise attention features, 𝐅𝐄\mathbf{F_{E}} is the extended point image features, σ\sigma is the sigmoid activation function and \otimes is the element-wise product operator. After obtaining the attention features, concatenate 𝐅𝐏\mathbf{F_{P}}, 𝐅𝐈\mathbf{F_{I}}, 𝐅𝐚.𝐏\mathbf{F_{a.P}} and 𝐅𝐚.𝐈\mathbf{F_{a.I}} channel-wise to get the point-wise fusion feature. Then divide the 3D point cloud space into pillars, followed by grouping the points to pillars. Finally, the simplified version of PointNet is adopted to generate the pillar-wise fusion features.

The advantage of this method is that since the fusion method is simple and the added networks are all shallow networks, the approach can achieve fast detection speed. Moreover, the approach can learn to summarize useful information from both modalities through PFN layer because of the early stage fusion strategy.

DenseAttentionFusion(DAF): In contrast to PointAttentionFusion that fuses features in a concise manner, DenseAttentionFusion employs a relatively complex fusion strategy, where features are divided into three forms and fused together. As shown in Fig. 4, the three features are the point cloud features, the image features, and the extended point image features 𝐅𝐄\mathbf{F_{E}} described in PointAttentionFusion.

Refer to caption
Figure 4: Architecture of the proposed MAFF module with DenseAttentionFusion technology. Different colors represent the features of different modalities

In PointAttentionFusion, only extended point image features 𝐅𝐄\mathbf{F_{E}} is used for subsequent work. However, after PFN learning, the feature is equivalent to the global features of the point cloud and the image. This feature will lose some characteristics of the original two modalities. Thus, a naive approach would be to put the original point cloud features and image features back into the global features. However, due to the noise information of the image features and the redundancy of the three features, blindly fusing the three features will introduce lots of interference information, resulting in the degradation of detection performance. In the following, we describe a novel dense attention fusion method that effectively combines the three types of features.

First, after projecting 3D points onto the image, the corresponding RGB features can be obtained. The RGB features have two uses, one is to expand the point cloud features in the same way as PointAttentionFusion to obtain the point cloud image features, and the other can regard this point-wise image feature as another feature form of 3D points. In this way, we can obtain point cloud features, point cloud image mixed features, and projected image features. Because the above three features are all in point-wise form, they can use the x,yx,y and zz coordinates of the 3D point to perform pillar operations, respectively. Here, we can get the pillar representations of three different features, and then three PFNs with the same structure are adopted to encode the pillars to generate three type of pillar-wise features.

Next, we use the pillar-wise channel attention module to adaptively estimate the importance of each pillar feature. The above three features are concatenated together as the input of this module. The attention features of the three features are estimated by the following attention maps:

𝐅𝐚.𝐏\displaystyle\mathbf{F_{a.P}} =𝐅𝐏σ(MLPP(𝐅𝐂))Point cloud attention\displaystyle=\mathbf{F_{P}}\otimes\underbrace{\sigma(\mathrm{MLP_{P}}(\mathbf{F_{C}}))}_{\text{Point cloud attention}} (2)
𝐅𝐚.𝐏𝐈\displaystyle\mathbf{F_{a.PI}} =𝐅𝐏𝐈σ(MLPPI(𝐅𝐂))Point Image attention\displaystyle=\mathbf{F_{PI}}\otimes\underbrace{\sigma(\mathrm{MLP_{PI}}(\mathbf{F_{C}}))}_{\text{Point Image attention}}
𝐅𝐚.𝐈\displaystyle\mathbf{F_{a.I}} =𝐅𝐈σ(MLPI(𝐅𝐂))Image attention\displaystyle=\mathbf{F_{I}}\otimes\underbrace{\sigma(\mathrm{MLP_{I}}(\mathbf{F_{C}}))}_{\text{Image attention}}
𝐅𝐀\displaystyle\mathbf{F_{A}} =𝐅𝐚.𝐏+𝐅𝐚.𝐏𝐈+𝐅𝐚.𝐈\displaystyle=\mathbf{F_{a.P}}+\mathbf{F_{a.PI}}+\mathbf{F_{a.I}}

where 𝐅𝐏\mathbf{F_{P}}, 𝐅𝐏𝐈\mathbf{F_{PI}} and 𝐅𝐈\mathbf{F_{I}} represent the pillar-wise point cloud, point image and image features, respectively, 𝐅𝐂\mathbf{F_{C}} is the concatenated feature of the above three features, 𝐅𝐚.𝐏\mathbf{F_{a.P}}, 𝐅𝐚.𝐏𝐈\mathbf{F_{a.PI}} and 𝐅𝐚.𝐈\mathbf{F_{a.I}} are the corresponding pillar-wise attention features, 𝐅𝐀\mathbf{F_{A}} is the pillar-wise attention features that need to be obtained, σ\sigma is the sigmoid activation function and \otimes represents the element-wise product operator. After obtaining the attention features, concatenate 𝐅𝐏\mathbf{F_{P}}, 𝐅𝐏𝐈\mathbf{F_{PI}}, 𝐅𝐈\mathbf{F_{I}}, and 𝐅𝐀\mathbf{F_{A}} channel-wise to get the pillar-wise fusion features.

Although DenseAttentionFusion is a relatively complex fusion strategy, it has the following advantages. First, it fully retains the original characteristics of the three features of point cloud, image, and point cloud image, while minimizing the noise impact of each feature. This makes its detection performance better than PointAttentionFusion. Second, the fusion is based on Pillars features, which can reduce the dependence on the availability of high-resolution 3D points.

III-D Training Details

Network Architecture: For the fairness of comparison, we keep most of the settings of PointPillars as described in [1] except for some newly added module structures. According to the type of fusion, the input and output dimensions of the PFN layers and the MLP layers are different. Note that, except for the input dimensions, the structure of 2D convolutional network in the proposed MAFF-Net is the same as PointPillars.

For PointAttentionFusion, the configuration of image dimension prediction MLPPD\mathrm{MLP_{PD}} is (3,96,16). The MLPP\mathrm{MLP_{P}} and MLPI\mathrm{MLP_{I}} in the attention module have (9+16,25,9) and (9+16,25,16) configurations respectively, where 9 is the dimension of the point cloud feature and 16 is the dimension of the image feature. Next, the configuration of PFN layer is (9+16+9+16,64), which leads to the input dimension of the subsequent 2D convolutional network is 64.

For DenseAttentionFusion, the configuration of MLPPD\mathrm{MLP_{PD}} is the same as the MLPPD\mathrm{MLP_{PD}} in PointAttentionFusion. PFNP\mathrm{PFN_{P}}, PFNPI\mathrm{PFN_{PI}} and PFNI\mathrm{PFN_{I}} have configurations of (16, 64), (9+16, 64) and (3, 64) respectively. The configuration of the three MLPs in the attention module are all (64*3, 64*3, 64), but their weights are not shared. The final fusion feature is combined by four types of features, so its dimension is 64*4, which leads to the input dimension of the 2D convolutional network is 64*4.

Loss: We use the same loss functions described in PointPillars. The loss function is divided into 3 types, namely the localization regression loss loc\mathcal{L}_{loc}, the classification loss cls\mathcal{L}_{cls}, and the orientation loss dir\mathcal{L}_{dir}. loc\mathcal{L}_{loc} uses SmoothL1 function to define the loss for (x,y,z,w,l,h,θ)(x,y,z,w,l,h,\theta), cls\mathcal{L}_{cls} uses focal loss, and dir\mathcal{L}_{dir} uses a softmax classification loss. The overall loss function can be defined as:

total=1Npos(βlocloc+βclscls+βdirdir)\mathcal{L}_{total}=\frac{1}{{N_{pos}}}\left({\beta_{loc}\mathcal{L}_{loc}{\rm{+}}\beta_{cls}\mathcal{L}_{cls}{\rm{+}}\beta_{dir}\mathcal{L}_{dir}}\right) (3)

where βloc=2.0,βcls=1.0,βdir=0.2\beta_{loc}=2.0,\ \beta_{cls}=1.0,\ \beta_{dir}=0.2, and NposN_{pos} is the number of positive anchors.

Data Augmentation: Our two fusion methods both project the 3D points to the image in a point-wise manner, so all the data augmentation methods of PointPillars can be used, which is similar to PointPainting[9]. Data augmentation adopts sample objects from database, augment ground truths independently, and perform global augmentations for the whole point cloud and all boxes.

IV EXPERIMENTS

IV-A Dataset and Metric

Dataset and Implementation: This study uses the KITTI benchmark dataset [12] to evaluate the proposed fusion method, which includes 7481 training samples and 7518 testing samples. There are three levels of difficulty: easy, moderate and hard, which are assessed based on the object height in image, occlusion and truncation. The rank of leaderboard is based on the moderate result. We follow the general approach proposed by [31] to split the training samples into a training set of 3712 samples and a validation set of 3769 samples. The proposed MAFF-Net is compared with previously published method on the car category. Inference of the entire network is carried out on Tesla V100 GPU. The code of PointPillars used in this study comes from (httP://github.com/nutonomy/second.pytorch).

Metric: The metric of KITTI is defined by average precision(AP) on 40 recall positions of the Precision/Recall curve [42] with IoU=0.7. This metric is a good indicator of the overall performance, but it cannot effectively reflect the details of the performance, such as the FP that we are concerned about. To show the performance of filtering FP, a simple way can be used: for each recall position, the number of TP is the same. According to the formula of precision precision=TPTP+FPprecision=\frac{TP}{TP+FP}, when TP remains unchanged, the less FP, the higher precision. Thus, the precision at a single recall position can reflect the performance of the algorithm processing FP. In this study, we use precision of five recall positions (0.725,0.75,0.775,0.80.725,0.75,0.775,0.8) to verify the proposed method performance for filtering FP. The reason for choosing these positions is that the scores of detection at these positions are generally at a medium level and it is difficult to filter vehicles only using the score threshold, which will lead to many FP.

TABLE I: 3D detection performance(%) comparison on the KITTI validation set (Car). Top-1 method is highlighted in BOLD.
Method Recall=0.725 Recall=0.75 Recall=0.775 Recall=0.8
easy moderate hard mAP easy moderate hard mAP easy moderate hard mAP easy moderate hard mAP
PointPillars 95.17 86.21 81.75 87.71 94.29 83.68 77.41 85.13 93.49 79.99 71.39 81.62 92.41 73.86 66.03 77.44
MAFF-Net(PAF) 96.48 87.56 82.83 88.96 96.16 85.25 78.91 86.78 95.34 81.58 73.31 83.41 94.44 75.69 67.99 79.37
MAFF-Net(DAF) 95.64 88.35 83.53 89.17 95.03 86.08 80.11 87.07 94.23 83.05 73.56 83.61 93.49 78.08 68.82 80.13
Improvement +1.31 +2.13 +1.78 +1.46 +1.87 +2.40 +2.69 +1.94 +1.86 +3.06 +2.17 +1.99 +2.03 +4.22 +2.79 +2.69

IV-B Evaluation on KITTI Validation Set

Evaluation on filtering FP. Table I shows the comparison of the 3D detection performance of PointPillars and MAFF-Net at different recall positions. The performance of each position has been significantly improved. As recall increases (in this case, the detection score decreases and FP is easier to increase), the improvement of mAP gradually increases. For the moderate category that is concerned on the KITTI benchmark, when the recall is 0.8, the proposed fusion method improves the performance by 4.22%. For a single recall position, only reducing FP can improve performance because of the same number of TP. Therefore, the performance improvement on each single recall position shows that MAFF-Net is effective for filtering FP.

TABLE II: The number of TP and FP for 3D detection under different score thresholds (Car). BG means background. Top-1 method is highlighted in BOLD.
Method Score Threshold = 0.4 Score Threshold = 0.1
TP FP FP(BG) TP FP FP(BG)
PointPillars 8606 4428 2346 8783 26237 22403
MAFF-Net(PAF) 8636 4080 2018 8802 24446 20478
MAFF-Net(DAF) 8627 3933 1906 8811 23330 19585
Improvement number/rate - -495 -440 - -2907 -2818

In order to reflect the performance of filtering FP more intuitively, Table II lists the number of TP and FP for the 3D vehicle detection under different score thresholds. When the score are 0.4 and 0.1, the FP are reduced by 11.18% and 11.08%, respectively, and the FP caused by the background are reduced by 18.75% and 12.58%, respectively. It shows that the proposed two fusion methods can effectively reduce FP while slightly improving TP, especially for reducing FP caused by background.

TABLE III: Comparison of 3D and BEV AP with/without attention module on the KITTI validation set (40 recall positions).
Method 3D(Car) BEV(Car)
easy moderate hard easy moderate hard
PointPillars(baseline) 87.79 78.44 74.06 92.55 88.32 86.52
MAFF-Net(PointFusion) 89.02 77.76 74.51 93.55 87.52 84.78
MAFF-Net(PAF) 88.86 79.30 74.71 93.10 89.25 86.67
MAFF-Net(DenseFusion) 88.05 77.20 73.73 93.10 87.21 84.29
MAFF-Net(DAF) 88.88 79.37 74.68 93.23 89.31 86.61

Analysis of the attention mechanisms. We visualize the point cloud features in DenseAttentionFusion before (𝐅𝐏\mathbf{F_{P}}) and after (𝐅𝐚.𝐏\mathbf{F_{a.P}}) using attention, as shown in Figure 6. After using the attention module, the background area is fully suppressed, so that the geometric shape of the vehicles can be highlighted. Moreover, for some objects that are similar to the vehicle in 3D structure, the attention module can eliminate some of their shapes, so that their 3D structures are no longer similar to the vehicle.

In order to further show the role of the attention module, Table III shows the performance comparison of the KITTI validation set with and without the attention module. When the attention module is not used, compared to PointPillars, the two fusion methods have improved performance in the easy category, but the performance in the moderate and hard categories decreases. In the moderate and hard categories, interference information is very serious. In these cases, when the point cloud is projected to the image, it will get the wrong image features, which will lead to the degradation of detection performance. After adding the attention module, the MAFF-Net is able to adapt to the weight of each modal feature, thereby improving the detection performance.

Refer to caption
Figure 5: Qualitative analysis of results from KITTI validation dataset by projecting 3D bounding boxes into image for clearer visualization. Fist line(top): MAFF-Net with DenseAttentionFusion, Second Line: MAFF-Net with PointAttentionFusion, Third line: PointPillars(baseline), Last Line(bottom): Spatial distribution of 3D point cloud. Green 3D boxes indicate prediction results. Red rectangles highlight false positive.
Refer to caption
Figure 6: Visualization of features before/after attention. First line(top): image, Second line: point cloud feature before attention in DenseAttentionFusion, Third line: point cloud feature after attention in DenseAttentionFusion. Green rectangular draws the area that is easy to be false detected. The attention module can effectively suppress the background area and enhance the shape of the vehicle.

The visualization of Detection Results. Figure 5 provides a qualitative analysis of some 3D detection results. It can be observed from Figure 5 that the 3D structures of many false detection objects are similar to that of vehicles, such as the hierarchical tree clusters on the left side of the second column image, the object clusters on the right side of the third column image, and wall and fence on the left side of the fourth column image. These conditions will make it difficult for the lidar point cloud to distinguish the objects. However, MAFF-Net can effectively reduce the above false detection objects by using image information, which demonstrates the effectiveness of the proposed algorithm.

IV-C Evaluation on KITTI Testing Set

We evaluate the proposed MAFF-Net with DenseAttentionFusion on the KITTI testing set by submitting the test results to the official server. The results are summarized in Table IV. It can be observed that MAFF-Net with DenseAttentionFusion can have competitive results compared with other state-of-the-art multimodal fusion algorithms. The proposed method has the fastest speed, achieves top rank in one category and 2nd rank in three categories within the fusion methods. It is worth pointing out that the results of the test set we submitted(PointPillars(baseline)) using the code given in PointPillars[1] are not ideal, for example, it is lower than the original PointPillars 1.1% in the 3D moderate category. However, even if only the low-performance PointPillars are used, the detection performance of MAFF-Net is still better than the original PointPillars in most categories. In 3D moderate, MAFF-Net is 1.9% higher than PointPillars (baseline) and 0.7% higher than PointPillars. Therefore, if a more powerful point cloud backbone network can be used, MAFF-Net will achieve better performance.

TABLE IV: Comparison of 3D and BEV AP with other multimodal published methods on the KITTI testing set. Top-2 methods are highlighted in BOLD.
Method 3D(Car) BEV(Car) Speed
easy moderate hard easy moderate hard (Hz)
MV3D[31] 74.97 63.63 54.00 86.62 78.93 69.80 2.80
AVOD-FPN[32] 83.07 71.76 65.73 90.99 84.82 79.62 10.00
ContFuse[33] 83.68 68.78 61.67 94.07 85.35 75.88 16.70
F-ConvNet[35] 87.36 76.39 66.69 91.51 85.84 76.11 2.1
F-pointnet[34] 82.19 69.79 60.59 91.17 84.67 74.77 5.90
PI-RCNN[11] 84.37 74.82 70.03 91.44 85.81 81.00 10.00
PointPillars[1] 82.58 74.31 68.99 90.07 86.56 82.81 62.00
PointPillars(baseline) 83.11 73.12 67.73 90.06 86.64 79.19 62.00
MAFF-Net(DAF) 85.52 75.04 67.61 90.79 87.34 77.66 24.00

V CONCLUSION

In this paper, we have explored how RGB images can assist 3D vehicle detection and proposed an end-to-end single-stage feature adaptive fusion network by extending the recently proposed PointPillars, to achieve fast speed and effectively eliminate false positive. Based on the channel attention mechanism, we propose two fusion techniques: PointAttentionFusion uses the channel attention mechanism to perform point-wise feature fusion of the two modalities; DointAttentionFusion converts the image and point cloud into three modalities, and then performs pillar-wise feature fusion of multi-modalities. Researchers can choose different fusion methods according to actual needs. Evaluation on the KITTI dataset demonstrates significant improvement in filtering false positive over the approach that uses only a single modality. Furthermore, the proposed method yields competitive results and has the fastest speed compared to the published state-of-the-art multi-modal methods in the KITTI benchmark. In the future, we plan to explore how images can bring more performance enhancements to 3D vehicle detection and the gains images can bring to multi-class detection networks.

References

  • [1] A. H. Lang, S. Vora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom, “Pointpillars: Fast encoders for object detection from point clouds,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 12 697–12 705.
  • [2] Y. Yan, Y. Mao, and B. Li, “Second: Sparsely embedded convolutional detection,” Sensors, vol. 18, no. 10, p. 3337, 2018.
  • [3] S. Shi, X. Wang, and H. Li, “Pointrcnn: 3d object proposal generation and detection from point cloud,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 770–779.
  • [4] S. Shi, Z. Wang, J. Shi, X. Wang, and H. Li, “From points to parts: 3d object detection from point cloud with part-aware and part-aggregation network,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
  • [5] Z. Yang, Y. Sun, S. Liu, X. Shen, and J. Jia, “Std: Sparse-to-dense 3d object detector for point cloud,” in Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 1951–1960.
  • [6] Z. Yang, Y. Sun, S. Liu, and J. Jia, “3dssd: Point-based 3d single stage object detector,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 040–11 048.
  • [7] J. H. Yoo, Y. Kim, J. S. Kim, and J. W. Choi, “3d-cvf: Generating joint camera and lidar features using cross-view spatial feature fusion for 3d object detection,” arXiv preprint arXiv:2004.12636, 2020.
  • [8] M. Liang, B. Yang, Y. Chen, R. Hu, and R. Urtasun, “Multi-task multi-sensor fusion for 3d object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 7345–7353.
  • [9] S. Vora, A. H. Lang, B. Helou, and O. Beijbom, “Pointpainting: Sequential fusion for 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 4604–4612.
  • [10] V. A. Sindagi, Y. Zhou, and O. Tuzel, “Mvx-net: Multimodal voxelnet for 3d object detection,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 7276–7282.
  • [11] L. Xie, C. Xiang, Z. Yu, G. Xu, Z. Yang, D. Cai, and X. He, “Pi-rcnn: An efficient multi-sensor 3d object detector with point-based attentive cont-conv fusion module.” in AAAI, 2020, pp. 12 460–12 467.
  • [12] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition.   IEEE, 2012, pp. 3354–3361.
  • [13] S. Shi, C. Guo, L. Jiang, Z. Wang, J. Shi, X. Wang, and H. Li, “Pv-rcnn: Point-voxel feature set abstraction for 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10 529–10 538.
  • [14] Z. Liang, M. Zhang, Z. Zhang, X. Zhao, and S. Pu, “Rangercnn: Towards fast and accurate 3d object detection with range image representation,” arXiv preprint arXiv:2009.00206, 2020.
  • [15] C. He, H. Zeng, J. Huang, X.-S. Hua, and L. Zhang, “Structure aware single-stage 3d object detection from point cloud,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 873–11 882.
  • [16] R. Girshick, “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448.
  • [17] M. Simony, S. Milzy, K. Amendey, and H.-M. Gross, “Complex-yolo: An euler-region-proposal for real-time 3d object detection on point clouds,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 0–0.
  • [18] J. Redmon and A. Farhadi, “Yolo9000: better, faster, stronger,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 7263–7271.
  • [19] B. Yang, W. Luo, and R. Urtasun, “Pixor: Real-time 3d object detection from point clouds,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2018, pp. 7652–7660.
  • [20] Y. Zeng, Y. Hu, S. Liu, J. Ye, Y. Han, X. Li, and N. Sun, “Rt3d: Real-time 3-d vehicle detection in lidar point cloud for autonomous driving,” IEEE Robotics and Automation Letters, vol. 3, no. 4, pp. 3434–3440, 2018.
  • [21] D. Z. Wang and I. Posner, “Voting for voting in online point cloud object detection.” in Robotics: Science and Systems, vol. 1, no. 3, 2015, pp. 10–15 607.
  • [22] M. Engelcke, D. Rao, D. Z. Wang, C. H. Tong, and I. Posner, “Vote3deep: Fast object detection in 3d point clouds using efficient convolutional neural networks,” in 2017 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2017, pp. 1355–1361.
  • [23] B. Li, “3d fully convolutional network for vehicle detection in point cloud,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2017, pp. 1513–1518.
  • [24] Y. Zhou and O. Tuzel, “Voxelnet: End-to-end learning for point cloud based 3d object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4490–4499.
  • [25] P. Yun, L. Tai, Y. Wang, C. Liu, and M. Liu, “Focal loss in 3d object detection,” IEEE Robotics and Automation Letters, vol. 4, no. 2, pp. 1263–1270, 2019.
  • [26] H. Yi, S. Shi, M. Ding, J. Sun, K. Xu, H. Zhou, Z. Wang, S. Li, and G. Wang, “Segvoxelnet: Exploring semantic context and depth-aware features for 3d vehicle detection from point cloud,” in 2020 International Conference on Robotics and Automation (ICRA).   IEEE, 2020, pp. 2274–2280.
  • [27] Z. Liu, X. Zhao, T. Huang, R. Hu, Y. Zhou, and X. Bai, “Tanet: Robust 3d object detection from point clouds with triple attention.” in AAAI, 2020, pp. 11 677–11 684.
  • [28] M. Ye, S. Xu, and T. Cao, “Hvnet: Hybrid voxel network for lidar based 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 1631–1640.
  • [29] C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 652–660.
  • [30] C. R. Qi, L. Yi, H. Su, and L. J. Guibas, “Pointnet++: Deep hierarchical feature learning on point sets in a metric space,” in Advances in neural information processing systems, 2017, pp. 5099–5108.
  • [31] X. Chen, H. Ma, J. Wan, B. Li, and T. Xia, “Multi-view 3d object detection network for autonomous driving,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 1907–1915.
  • [32] J. Ku, M. Mozifian, J. Lee, A. Harakeh, and S. Waslander, “Joint 3d proposal generation and object detection from view aggregation,” IROS, 2018.
  • [33] M. Liang, B. Yang, S. Wang, and R. Urtasun, “Deep continuous fusion for multi-sensor 3d object detection,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 641–656.
  • [34] C. R. Qi, W. Liu, C. Wu, H. Su, and L. J. Guibas, “Frustum pointnets for 3d object detection from rgb-d data,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 918–927.
  • [35] Z. Wang and K. Jia, “Frustum convnet: Sliding frustums to aggregate local point-wise features for amodal,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2019, pp. 1742–1749.
  • [36] K. Shin, Y. P. Kwon, and M. Tomizuka, “Roarnet: A robust 3d object detection based on region approximation refinement,” in 2019 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2019, pp. 2510–2515.
  • [37] J. Dou, J. Xue, and J. Fang, “Seg-voxelnet for 3d vehicle detection from rgb and lidar data,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 4362–4368.
  • [38] H. Kuang, B. Wang, J. An, M. Zhang, and Z. Zhang, “Voxel-fpn: Multi-scale voxel feature aggregation for 3d object detection from lidar point clouds,” Sensors, vol. 20, no. 3, p. 704, 2020.
  • [39] H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu, A. Krishnan, Y. Pan, G. Baldan, and O. Beijbom, “nuscenes: A multimodal dataset for autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 621–11 631.
  • [40] Y. Zhou, P. Sun, Y. Zhang, D. Anguelov, J. Gao, T. Ouyang, J. Guo, J. Ngiam, and V. Vasudevan, “End-to-end multi-view fusion for 3d object detection in lidar point clouds,” in Conference on Robot Learning, 2020, pp. 923–932.
  • [41] Y. Wang, A. Fathi, A. Kundu, D. Ross, C. Pantofaru, T. Funkhouser, and J. Solomon, “Pillar-based object detection for autonomous driving,” arXiv preprint arXiv:2007.10323, 2020.
  • [42] A. Simonelli, S. R. Bulo, L. Porzi, M. López-Antequera, and P. Kontschieder, “Disentangling monocular 3d object detection,” in Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 1991–1999.