Faraway-Frustum: Dealing with Lidar Sparsity for 3D Object Detection using Fusion
Abstract
Learned pointcloud representations do not generalize well with an increase in distance to the sensor. For example, at a range greater than 60 meters, the sparsity of lidar pointclouds reaches to a point where even humans cannot discern object shapes from each other. However, this distance should not be considered very far for fast-moving vehicles: A vehicle can traverse 60 meters under two seconds while moving at 70 mph. For safe and robust driving automation, acute 3D object detection at these ranges is indispensable. Against this backdrop, we introduce faraway-frustum: a novel fusion strategy for detecting faraway objects. The main strategy is to depend solely on the 2D vision for recognizing object class, as object shape does not change drastically with an increase in depth, and use pointcloud data for object localization in the 3D space for faraway objects. For closer objects, we use learned pointcloud representations instead, following state-of-the-art. This strategy alleviates the main shortcoming of object detection with learned pointcloud representations. Experiments on the KITTI dataset demonstrate that our method outperforms state-of-the-art by a considerable margin for faraway object detection in bird’s-eye-view and 3D. Our code is open-source and publicly available: https://github.com/dongfang-steven-yang/faraway-frustum.
I INTRODUCTION
3D/Bird’s eye view (BEV) object detection is a critical task for many robotics applications. Existing lidar-based methods show good performance for close to medium range objects. However, a closer look at the state-of-the-art exposes an inherent problem: learned pointcloud representations do not generalize well with an increase in sparsity. This is not a surprising phenomenon. At a range greater than sixty meters, lidar pointcloud sparsity reaches a point where even humans cannot discern object shapes from each other. For example, in KITTI 3D/BEV object detection benchmark [1], the state-of-the-art 3D object detection performance is remarkable. But when these high performing models face objects that are located at 60 meters and beyond, mean average precision drops to almost zero. We believe this is an important issue for automated driving. For instance, detecting faraway objects can offer more time for the automated vehicle to make better decisions.

3D/BEV object detection for faraway objects is challenging, and state-of-the-art (SOTA) lidar-based detectors [2, 3, 4, 5, 6] do not perform well for this task. We believe this is caused by sparsity and near-random scattering of the few points obtained from faraway objects. Learned representations from close to medium range objects do not generalize to faraway cases, and since SOTA approaches are primarily deep neural networks, they cannot learn the representations of faraway cases.
RGB-pointcloud fusion is a common strategy [7, 8, 9, 10, 11, 12] to increase 3D object detection performance. For example, some works [7], [12] focus on using 2D detection results to generate frustum-based search spaces in pointclouds. As shown in Fig. 1, a faraway object in the RGB image domain usually contains around 400 pixels, which can be easier to recognize with a mature 2D detector. As such, a fusion-based approach can be a good candidate for faraway object detection. However, even though the aforementioned studies use RGB imagery to boost detection performance, they still depend on solely learned pointcloud representations to localize objects in 3D.
In this work, we propose an alternative 3D/BEV detector, Faraway-Frustum, to address the problem of faraway object detection. We follow the idea of frustum generation but use clustering instead of a neural network to estimate an initial object location in the cropped pointcloud for faraway objects. We still train a neural network to regress bounding box shape and refine depth. The overview of the proposed method is shown in Fig. 2. We first use 2D instance segmentation masks (or 2D bounding boxes) for each object in the RGB image space to generate frustums in the pointcloud space and find the corresponding lidar points for each object. Then, a pointcloud clustering technique is applied to estimate the 3D centroid of the object. By comparing the centroid with a faraway threshold, an object is determined to be treated as a faraway object or not. If so, a 3D bounding box is regressed by our Faraway Frustum Network (FF-Net) to the object based on the estimated centroid and frustum pointcloud . If it is not a faraway object, instead of clustering the raw pointcloud, learned representations are directly used for 3D box fitting, following SOTA.
To evaluate the proposed method, we conducted bench-marking experiments using the KITTI dataset [1]. In KITTI, the average number of lidar points for each faraway object (e.g. pedestrians over 60 meters and cars over 75 meters) is ten or less, which supports our motivation that an alternative approach is necessary for faraway objects instead of directly using pointcloud-driven neural network approaches. The experimental results demonstrated that our method outperforms the SOTA methods on faraway object detection, which indicates that our method effectively fuses the RGB data with very sparse pointcloud. As shown in Fig. 6, our proposed method successfully detects faraway objects, where SOTA methods (fusion or pointcloud only) fail. Our main contributions can be summarized as follows:
Introduction of a novel fusion strategy: depending solely on 2D vision for object-class recognition and using frustum-cropped pointcloud data with clustering for 3D object localization.
Showing that using clustering with cropped, very sparse raw pointcloud data is a better strategy than using learned representations for faraway 3D object detection. As shown in Fig. 1, within very sparse pointclouds, the shape of objects changes drastically and randomly. As such, using representations learned mostly from closer objects is not useful.
Demonstrating state-of-the-art 3D object detectors’ failure with objects at a distance over sixty meters in the KITTI dataset. The proposed faraway-frustum approach outperforms SOTA with a significant margin.
II RELATED WORK
In this section, we briefly review state-of-the-art 3D/BEV object detection methods. We divide them into two main categories: pointcloud only methods and RGB-pointcloud fusion methods. In the second category, we mainly discuss feature-based fusion and frustum-based fusion. We also discuss their performance of detecting faraway objects.
Pointcloud only methods. One way of processing pointcloud is based on voxels [3, 4, 2, 13]. Such methods first convert the pointcloud into voxel grids and then learn the representation of each voxel. 3D/BEV detection is achieved with the learned voxel representation. Alternatively, raw pointcloud data can be used directly utilizing PointNet-based architectures [14] and make 3D/BEV object detections [15, 16, 6]. These methods are robust for most objects. However, pointcloud only methods all have the difficulty of detecting faraway objects, because lidar points of faraway objects are too sparse to be voxelized and learned, leading to no detection result in most cases.
Feature-based Fusion. Feature-based fusion methods try to make the pointcloud data and the RGB data complement each other. One way is to fuse the information from pointcloud into RGB image. For example, [8] fuses the features from Region of Interest (RoI) in both 2D image and 2D depth map, and then conducts 3D box regression. MV3D [9] projects the lidar pointcloud to two-view image representations (Bird view and Front view). The features and information extracted from these two image representations and the RGB image are then fed into a region-based fusion network for 3D object detection. AVOD [10] firstly generates a BEV map from a voxel grid representation of the lidar pointcloud. The features extracted from both the BEV map and the RGB image are fused for 3D object detection through a first-stage region proposal network and a second-stage detector network. The main problem of these methods is the loss of 3D geometric information of lidar pointcloud because of using the pointcloud’s 2D representations, leading to some errors in locating small objects such as pedestrians. Feature-based fusion can also be achieved by fusing the information from image space into pointcloud. For example, [11] extracts the geometric features in 3D and color features in 2D from RGB-D images and then fuses them for 3D object detection. MVX-Net [17] fuses the RGB image and pointcloud point-wise or voxel-wise. The features extracted from the RGB image by a pre-trained 2D CNN are fused with the pointcloud in a voxel-based network to do 3D object detection. PointPainting [18] assigns the semantic feature to each lidar point by fusing the 2D detection result from the RGB image, thus achieving better results in pointcloud-based neural network detector. Since these approaches heavily rely on the pointcloud feature, they still can not generate good results for faraway objects with sparse lidar points.
Frustum-based Fusion. Frustum-based fusion methods use the detection results from 2D image to generate frustums for pointcloud, hence reducing the search space in 3D. An early and classic method is Frustum PointNets [7]. This method first generates a frustum for each object detected in 2D, then applies a PointNet-based approach to do instance segmentation and 3D box estimation in each frustum. Some work [19, 20] have improved the process of frustum generation by filtering out some background noise, and there is work focusing on changing the content of frustums. For example, Frustum ConvNet [12] generates a sequence of sub-frustums via sliding in the original 3D frustum. Frustum Voxnet [21] voxelizes parts of the frustum instead of using the whole frustum space, which offers more accurate representations around the area of interest. Some other researchers aimed to provide more fusion information. For example, one work [22] combines the pointcloud features in the frustum with the image features in the 2D bounding box as early-fusion and then apply a PointNet-based detector. Another work [23] fuses their own BEV detection results with 3D/BEV results from Frustum PointNets as late-fusion. These perform well for most objects. But unfortunately, for faraway objects with sparse lidar points, pure neural network based approaches cannot generalize well.

One recent work [24] achieved good 3D/BEV pedestrian detection results around 30 meters. In our work, we extend the range significantly, and detect pedestrians at 60 meters and beyond, where most SOTA approaches completely fail.
III PROPOSED METHOD

An overview of our proposed method is shown in Fig. 2 and Algorithm 1. Our method takes both the RGB image and lidar pointcloud as input and outputs 3D/BEV bounding boxes with class id . There are three main stages in our method: frustum generation, centroid estimation, and depth-refinement with box regression. Each stage will be illustrated in detail in the following subsections.
III-A Frustum Generation
2D instance segmentation. 2D instance segmentation serves as the base of frustum generation. It takes an image as input and outputs the 2D object detection results containing 2D bounding boxes and semantic masks.
In this work, we use 2D instance segmentation framework Mask R-CNN [25] to obtain 2D object information from image :
(1) |
where represents the Mask R-CNN framework. is the instance segmentation result, which is a 4-tuple consisting of class label , 2D bounding box , 2D semantic mask , and confidence score for object .
Frustum generation. We use the 2D results to generate frustums and to further identify the lidar points that correspond to each object . With the known transformation between the camera and the lidar, we use the semantic mask for 2D-to-3D projection, as shown in Fig. 4(b). Then the corresponding pointcloud can be identified from the raw lidar pointcloud based on the frustum:
(2) |
The mask-frustum based projection is our main approach. As shown in Fig. 4(c), we believe that using the semantic mask can exclude some noise points that do not belong to the target object, e.g., the points from occluded objects or the background. As an alternative, we also tested box-frustum based projection , which is used as a comparison with our main approach.
III-B Centroid Estimation
Centroid estimation. With obtained from the frustum, we then estimate the 3D centroid for object :
(3) |
The 3D object centroid plays two key roles in our method. One is to use the depth to determine whether object should be treated as a faraway object. The other is to further generate the 3D/BEV detection results for faraway objects.
Based on our observation in KITTI dataset, no matter whether the pointcloud in the frustum is dense or sparse, there are always some points on the object’s surface. Thus, we adopt a fast clustering technique using histograms to estimate the 3D object centroid.
First, for all points in the pointcloud , the histogram of all the coordinate values in each axis is generated (here we have 3 axes x, y, and z). For the histogram of each axis, we define the edges of every bin in the histogram as , , and the count of values belonging to each bin as , , where is the number of bins. Then, we identify the bin with the largest count value. This indicates that most of the points are concentrated within this bin. The corresponding index will be obtained by . Finally, the centroid value of an axis, for example, the centroid of x-axis, , can be obtained by The centroid values and for the other two axes is estimated in the same way.
Faraway threshold. Using the estimated centroid , we conduct faraway judgement for each object . We set different faraway thresholds for different object classes based on the statistics of the number of ground truth lidar points in each object. As shown in Fig. 5, we first draw a line for the objects that have 10 lidar points, then we approximately select the such that most objects of distance larger than have less than 10 points. To determine whether object should be treated as a faraway object, we compare the estimated distance with of the corresponding object class . If , then it is a faraway object, otherwise not.
III-C Box Regression
To obtain the 3D/BEV bounding box for object , based on the estimated object centroid , we need to estimate the box shape: length , width , height , and the orientation . If object is a faraway object, directly using learned representations from state-of-the-art models is not a good choice because they do not generalize well from dense pointcloud to very sparse pointcloud. Furthermore, the estimated object centroid (especially the depth) may still be quiet far from the box center. As such, we propose to use a light model named Faraway Frustum Network (FF-Net) only for faraway objects to refine the depth and regress the shape of the 3D bounding box with the input of the object class and a 2D projection of the frustum pointcloud, as shown in Fig. 3.
Pointcloud Projection. 2D projection of the frustum pointcloud is generated from cooridinate transformation for pintcloud, as shown in Fig. 3 (a). After conducting frustum generation for raw pointcloud , the frustum pointcloud is obtained. First, the camera coordinate system is rotated to the center view of the frustum to build frustum coordinate system. Second, after histogram-based clustering and centroid estimation, the frustum coordinate system is transformed to the centroid coordinate system with the estimated centroid at origin. At last, lidar points inside of the frustum are all projected into the centroid coorinate system in bird’s eye view. The projected frustum pointcloud in centroid coordinate system is taken as the 2D projection of the frustum pointcloud.

Box Regression. FF-Net takes the object class and a 2D projection of the frustum pointcloud whose origin is the estimated centroid as input, and combines a MobileNet-based [26] backbone network with a multi-output regression head as shown in Fig. 3 (b).
The estimated centroid is considered as the origin of the input projection, and the goal of the FF-net is to shift this origin to the real center of the 3D bounding box (i.e. to transform centroid coordinate to 3D object coordinate as shwon in Fig. 3 (a)). Furthermore, another goal is to regress the box shape, achieved by minimizing the loss of the regressed length, width and height of the box. We use mean absolute error (MAE) to compute the loss ,,,,,, of box centroid ) and shape ( respectively. By summing these losses, FF-Net is trained and optimized with multi-task losses .
Finally, for a faraway object, we take the shifted depth and the regressed 3D bounding box shape from the output of FF-Net and combine them with () from the estimated centroid. We assign a 3D bounding box to the faraway object as:
(4) |
It should be noted that the class id is directly obtained with Mask R-CNN. If object is not a faraway object, we switch to using learned representations, following SOTA (e.g. Frustum-PointNets [7], PV-RCNN [2]). In this case, 3D bounding box and class id for a non-faraway object are obtained by .

IV EXPERIMENTS
We utilized the KITTI dataset [1] to conduct our experiments. We specifically extracted faraway objects in KITTI and investigated them separately. Details of dataset preparation, evaluation metrics, and implementation are described below.
Dataset preparation. First we analyzed the statistics of the original KITTI dataset by evaluating the distribution of the objects at different distances and having different numbers of lidar points, as shown in Fig. 5. It is obvious that as the distance increases, the number of lidar points in an object decreases. That is to say, the pointcloud is very sparse for faraway objects. We selected the faraway threshold for cars as 75 meters and pedestrians as 60 meters. We also split the KITTI dataset into the train set (3724 frames) and the validation set (3757 frames).

Evaluation metric. The major evaluation metric is the mean average precision (mAP) with a given IoU threshold, as suggested by the KITTI dataset. We use both the official benchmark and a specific benchmark for faraway objects. In the benchmark for faraway objects, we only evaluate faraway objects and we use a specific IoU threshold (0.1) for the mAP. We use low IoU threshold because detecting the object with a small overlap with the ground truth is still helpful at long distances, compared to no detection at all. The mAP results are computed with 11 recall positions which is the same as in previous works.
We also evaluate the faraway objects using average IoU (aIoU), which is defined as , where is the total number of faraway objects and is the sum of the IoU values calculated based on the ground truth and the predicted bounding box.
Implementation. Our method uses the instance-level semantic segmentation method (Mask R-CNN [25] pre-trained on COCO dataset [27]) to generate 2D object information from image space. Our first approach (ours1) uses 2D semantic masks to generate frustums in 3D. The second approach (ours2) uses 2D bounding boxes to generate frustums. As a baseline (ours3), we also use ground truth 2D bounding boxes provided by KITTI to generate frustums. All of our approaches are combined with PV-RCNN [2] for non-faraway objects. The Faraway Frustum Network (FF-Net) is trained using the Adam optimizer with early stopping. FF-Net is trained with the whole training set, but during inference only faraway objects are fed to the FF-Net.
V RESULTS
Quantitative results. Table I shows the average IoU results for BEV detection of faraway objects in the KITTI validation dataset. All of our methods outperform SOTA methods with higher average IoU (at least 0.051 and at most 0.157). And surprisingly, none of the methods except for ours can achieve an average of 0.1 IoU for both pedestrians and cars. This result not only demonstrates the effectiveness of our method, but also underlines an important shortcoming of SOTA methods. Furthermore, we believe finding the exact shape of faraway objects is not a priority. As long as we obtain the 3D/BEV detection result with even a small IoU (e.g. 0.1), it can still be very useful for certain applications such as automated driving. In other words, a 0.1 IoU detection is better than a false negative. As such, we set IoU threshold to 0.1 for faraway objects in the mAP comparison.
Method |
|
|
||
---|---|---|---|---|
|
|
|||
Frustum PointNets [7] | 0.000 | 0.000 | ||
SECOND [3] | 0.036 | 0.009 | ||
PointPillars [4] | 0.072 | 0.000 | ||
PV-RCNN [2] | 0.051 | 0.018 | ||
Ours1 (mask + PV-RCNN) | 0.123 | 0.157 | ||
Ours2 (box + PV-RCNN) | 0.124 | 0.150 |
-
*
Name explanation: Ped. (Pedestrian).
-
**
The bold result means the best in all methods, and the blue result represents the second place.
|
|
|
|
|||||
---|---|---|---|---|---|---|---|---|
Method |
|
|||||||
|
|
|||||||
Frustum PointNets [7] | 00.00 | 00.00 | 00.00 | 00.00 | ||||
SECOND [3] | 13.63 | 13.63 | 09.09 | 09.09 | ||||
PointPillars [4] | 22.40 | 22.40 | 00.00 | 00.00 | ||||
PV-RCNN [2] | 19.69 | 19.69 | 18.18 | 18.18 | ||||
Ours1 (mask + PV-RCNN) | 44.54 | 44.54 | 34.70 | 45.27 | ||||
Ours2 (box + PV-RCNN) | 31.95 | 45.45 | 46.90 | 46.90 |
-
*
Name explanation: Ped. (Pedestrian).
-
**
The bold result means the best in all methods, and the blue result represents the second place. We set the experimental IoU threshold as 0.1 for faraway pedestrians because in the current stage, it is extremely difficult to precisely locate faraway objects, while detecting faraway objects even with low IoU is still practical.
|
|
|||||||
Method |
|
|
||||||
Easy | Mod | Hard | Easy | Mod | Hard | |||
PV-RCNN [2] | 69.53/73.32 | 66.02/67.42 | 62.91/65.70 | 96.73/97.53 | 93.18/94.77 | 85.76/94.78 | ||
Ours1 (mask + PV-RCNN) | 71.65/73.02 | 67.29/68.73 | 62.08/66.87 | 96.73/97.54 | 93.17/94.75 | 85.76/94.77 | ||
Ours2 (box + PV-RCNN) | 71.74/73.11 | 67.29/68.73 | 62.08/66.88 | 96.73/97.54 | 93.17/94.75 | 85.76/94.77 |
-
For the non-faraway official Easy/Mod/Hard benchmark, our method performs as well as the baseline SOTA method (PV-RCNN). This result shows that the proposed method can be used to improve the faraway detection performance without sacrificing the non-faraway detection performance.
The mAP results of faraway 3D/BEV detection over KITTI validation dataset for pedestrians and cars are shown in Table II. For faraway pedestrians (over 60 meters), our methods (ours1 and ours2) outperform SOTA methods on 3D/BEV detection with large mAP margins (BEV: at least 22.14% and at most 45.45%, 3D: at least 9.55% and at most 44.54%). For faraway cars (over 75 meters), our methods (ours1 and ours2) outperform SOTA methods again with a higher mAP (BEV: at least 27.09% and at most 46.90%, 3D: at least 16.52% and at most 46.90%).
The mAP results of 3D/BEV detection over KITTI validation dataset for pedestrians and cars are shown in Table III. For the non-faraway official Easy/Mod/Hard benchmark, our method performs as well as the baseline SOTA method (PV-RCNN).
All the above results demonstrate that our method achieves better performance on faraway object detection without impairing the overall performance of SOTA methods.
Qualitative results. Fig. 6 shows an example of visual results of different methods for faraway object detection. We compared our method with PV-RCNN [2] and Frustum PointNets [7]. In frame (a), Frustum PointNets mistakenly detects the pole as a pedestrian, and PV-RCNN even has no result. However, for detecting the faraway pedestrian near the pole, only our detector succeeds. In frame (b), state-of-the-art methods all fail in detecting the faraway car. In contrast, our method successfully detects that car in 3D.
VI CONCLUSION
In this paper, we proposed an alternative 3D/BEV detector, named Faraway-Frustum, to deal with lidar sparsity of faraway objects. Our method takes the advantages of relatively dense image data to find faraway objects, and circumvents the disadvantages of pointcloud-driven neural networks working on very sparse points. Moreover, our alternative detector can be flexibly combined with a state-of-the-art method to form an overall 3D/BEV object detection system via setting faraway thresholds.
The experiments demonstrated the feasibility of our approach, but they also exposed a significant shortcoming of state-of-the-art object detection methods: Relying on learned representations of very sparse lidar points to detect faraway objects is not a good strategy.
ACKNOWLEDGMENT
Material reported here was supported by the United States Department of Transportation under Award Number 69A3551747111 for the Mobility21 University Transportation Center.
References
- [1] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2012, pp. 3354–3361.
- [2] S. Shi, C. Guo, L. Jiang, Z. Wang, J. Shi, X. Wang, and H. Li, “Pv-rcnn: Point-voxel feature set abstraction for 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10 529–10 538.
- [3] Y. Yan, Y. Mao, and B. Li, “Second: Sparsely embedded convolutional detection,” Sensors, vol. 18, no. 10, p. 3337, 2018.
- [4] A. H. Lang, S. Vora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom, “Pointpillars: Fast encoders for object detection from point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 12 697–12 705.
- [5] Z. Liu, X. Zhao, T. Huang, R. Hu, Y. Zhou, and X. Bai, “Tanet: Robust 3d object detection from point clouds with triple attention.” in AAAI, 2020, pp. 11 677–11 684.
- [6] Z. Yang, Y. Sun, S. Liu, and J. Jia, “3dssd: Point-based 3d single stage object detector,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 040–11 048.
- [7] C. R. Qi, W. Liu, C. Wu, H. Su, and L. J. Guibas, “Frustum pointnets for 3d object detection from rgb-d data,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, pp. 918–927.
- [8] Z. Deng and L. Jan Latecki, “Amodal detection of 3d objects: Inferring 3d bounding boxes from 2d ones in rgb-depth images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2017, pp. 5762–5770.
- [9] X. Chen, H. Ma, J. Wan, B. Li, and T. Xia, “Multi-view 3d object detection network for autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2017, pp. 1907–1915.
- [10] J. Ku, M. Mozifian, J. Lee, A. Harakeh, and S. L. Waslander, “Joint 3d proposal generation and object detection from view aggregation,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2018, pp. 1–8.
- [11] S. Song and J. Xiao, “Deep sliding shapes for amodal 3d object detection in rgb-d images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2016, pp. 808–816.
- [12] Z. Wang and K. Jia, “Frustum convnet: Sliding frustums to aggregate local point-wise features for amodal,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2019, pp. 1742–1749.
- [13] M. Ye, S. Xu, and T. Cao, “Hvnet: Hybrid voxel network for lidar based 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 1631–1640.
- [14] C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 652–660.
- [15] S. Shi, X. Wang, and H. Li, “Pointrcnn: 3d object proposal generation and detection from point cloud,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 770–779.
- [16] Z. Yang, Y. Sun, S. Liu, X. Shen, and J. Jia, “Std: Sparse-to-dense 3d object detector for point cloud,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 1951–1960.
- [17] V. A. Sindagi, Y. Zhou, and O. Tuzel, “Mvx-net: Multimodal voxelnet for 3d object detection,” in 2019 International Conference on Robotics and Automation (ICRA). IEEE, 2019, pp. 7276–7282.
- [18] S. Vora, A. H. Lang, B. Helou, and O. Beijbom, “Pointpainting: Sequential fusion for 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 4604–4612.
- [19] X. Du, M. H. Ang, S. Karaman, and D. Rus, “A general pipeline for 3d detection of vehicles,” in 2018 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2018, pp. 3194–3200.
- [20] K. Shin, Y. P. Kwon, and M. Tomizuka, “Roarnet: A robust 3d object detection based on region approximation refinement,” in 2019 IEEE Intelligent Vehicles Symposium (IV). IEEE, 2019, pp. 2510–2515.
- [21] X. Shen and I. Stamos, “Frustum voxnet for 3d object detection from rgb-d or depth images,” in The IEEE Winter Conference on Applications of Computer Vision, 2020, pp. 1698–1706.
- [22] X. Zhao, Z. Liu, R. Hu, and K. Huang, “3d object detection using scale invariant and feature reweighting networks,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, 2019, pp. 9267–9274.
- [23] P. Cao, H. Chen, Y. Zhang, and G. Wang, “Multi-view frustum pointnet for object detection in autonomous driving,” in 2019 IEEE International Conference on Image Processing (ICIP). IEEE, 2019, pp. 3896–3899.
- [24] M. Fürst, O. Wasenmüller, and D. Stricker, “Lrpd: Long range 3d pedestrian detection leveraging specific strengths of lidar and rgb,” in 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2020, pp. 1–7.
- [25] K. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
- [26] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.
- [27] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European conference on computer vision. Springer, 2014, pp. 740–755.