YOLOStereo3D: A Step Back to 2D for Efficient Stereo 3D Detection
Abstract
Object detection in 3D with stereo cameras is an important problem in computer vision, and is particularly crucial in low-cost autonomous mobile robots without LiDARs. Nowadays, most of the best-performing frameworks for stereo 3D object detection are based on dense depth reconstruction from disparity estimation, making them extremely computationally expensive. To enable real-world deployments of vision detection with binocular images, we take a step back to gain insights from 2D image-based detection frameworks and enhance them with stereo features. We incorporate knowledge and the inference structure from real-time one-stage 2D/3D object detector and introduce a light-weight stereo matching module. Our proposed framework, YOLOStereo3D, is trained on one single GPU and runs at more than ten fps. It demonstrates performance comparable to state-of-the-art stereo 3D detection frameworks without usage of LiDAR data. The code will be published in https://github.com/Owen-Liuyuxuan/visualDet3D.
I Introduction
3D object detection is a fundamental problem in computer vision, and a crucial engineering problem for autonomous vehicles and mobile robots [1] [2]. With two horizontally-aligned RGB cameras with known displacement, it is possible to estimate depth by triangulation according to the pin-hole camera model. Though this is a non-direct measurement for depth and is, in most cases, less robust than LiDAR-based approaches, the binocular setup is generally much cheaper and it is promising for low-cost applications such as mobile robots and autonomous logistic vehicles.

Many of the state-of-the-art frameworks in stereo 3D object detection stem from the idea of pseudo-LiDAR and are motivated by general stereo matching algorithms. However, in 3D object detection, the model should focus on foreground objects. It is expected to be as accurate as possible since a disparity error of one or two pixels would cause a large error in terms of real-world distance. Many researchers have delved deep into these problems to improve the performance of pseudo-LiDAR-based algorithms; some directly fine-tune the estimation of point clouds to improve performance [3],[4], while others utilize instance segmentation to focus the stereo matching network on foreground pixels [5], [6], [7]. However, a high-performance disparity estimation network, e.g., PSMNet [8], usually takes more than 300 ms per frame on modern hardware on the KITTI dataset [9] and requires a huge GPU memory to train. These issues hinder the deployment of stereo systems on low-cost robotic applications.
Many of the works mentioned above have shown in practice that transforming images into 3D features is usually sub-optimal and computationally expensive. To improve the efficiency of stereo 3D detection algorithms while maintaining as much of their performance as possible, we propose selecting a different architecture. Instead of casting the problem as a 3D detection problem with less accurate point clouds, we take a step back and treat it as a monocular 3D detection task with enhanced stereo features, which is the fundamental motivation of this work.
The framework we propose, YOLOStereo3D, is a light-weight one-stage stereo 3D detection network (Section III-A). To efficiently produce powerful stereo features, we re-introduce the pixel-wise correlation module to construct the cost-volume, instead of the popular concatenation-based module (Section III-B1). Such a module produces a thin 2D feature map where each channel corresponds to a disparity hypothesis in stereo matching. We then apply this module hierarchically to efficiently produce stereo features as 2D feature maps (Section III-B2), and we densely fuse these features (Section III-B3) to form the base-feature of detection heads. The network is trained end-to-end without the use of LiDAR data (Section III-C).
The main contributions of this paper are three-fold.
-
•
For the inference architecture, we incorporate and optimize the inference pipeline from one-stage monocular 3D detection into stereo 3D detection.
-
•
For the design of the network, we introduce a point-wise correlation module in stereo detection tasks and propose a hierarchical, densely-connected structure to utilize stereo features from multiple scales.
-
•
For the experimental results, the proposed YOLOStereo3D produces competitive results on the KITTI 3D benchmark without using point clouds and with an inference time of less than 0.1 seconds per frame.
II Related Works
II-A Stereo Matching
Stereo matching algorithms focus on estimating the disparity between binocular images. The current state-of-the-art frameworks for stereo matching apply siamese networks for feature extraction from two images, and construct 3D cost volumes to search the disparity value on each pixel exhaustively. Early research applied the dot-product between binocular feature maps, with the resulting correlation directly forming an estimation of the disparity distribution [10] [11]. PSMNet [8] and GCNet [12] constructed concatenation-based cost volumes and applied multiple 3D convolutions to produce disparity outputs. The recent FADNet managed to perform a fast stereo estimation with a point-wise correlation module [13]. Zhang et al. proposed stereo focal loss to improve the loss function formulation in disparity estimation [14]. Our work, similar to many other stereo 3D object detection algorithms, is developed upon these studies and utilizes the stereo matching features to boost detection performance.
II-B Visual 3D Object Detection
II-B1 Stereo 3D Object Detection
Stereo 3D object detection is usually considered as a tractable but computationally hard problem. Recent advances in stereo 3D object detection algorithms are based on the idea of pseudo-LiDAR [15]. DispRCNN [5], ZoomNet [6], and OC Stereo [7] applied instance segmentation on binocular images to construct a local point cloud for each detected instance to improve the accuracy of disparity estimation on foreground objects. Pseudo-LiDAR++ [3] recognized that uniform 3D convolution might not be suitable to process the disparity cost volume, and transformation to the depth cost volume may be needed.
We point out that all the aforementioned algorithms require more than 0.3 seconds runtime per frame. Moreover, Pseudo-LiDAR++ [3], ZoomNet [6], OC Stereo [7] and DSGN [16] required point cloud data during training or need point cloud data to help the training process. DispRCNN [5] and the baseline Pseudo-LiDAR [15] required off-the-shelf disparity modules, which are usually trained with depth images or point cloud data.
YOLOStereo3D is a light-weight model that performs most of the convolution operation in the perspective view, and the training and inference are significantly lighter and faster than all methods mentioned above. Moreover, the training process of YOLOStereo3D does not depend on point-cloud data.
II-B2 Monocular 3D Object Detection
Monocular 3D object detection is an ill-posed problem, but it provides many insights into how depth information can be estimated from a single image. Tom et al. [17] demonstrated that a typical monocular depth estimation network mainly estimates depth from the vertical position of an object. The authors [17] provided the theoretical background for pseudo-LiDAR in monocular detection [18][19]. YOLOStereo3D is built upon the inference structure of M3D-RPN [20] and GAC [21] and further enhances the final features with stereo matching results.
III Methods

In this section, we elaborate on the network structure and methods applied in this paper. First, we introduce the output definition and data-preprocessing tricks imported into and optimized for YOLOStereo3D [21]. Second, we re-introduce the light-weight cost volume that speeds up stereo matching and present the hierarchical densely-connected structure that fully exploits such thin features. Finally, we deliver the loss function as well as the training and inference scheme of YOLOStereo3D.
III-A Anchors Definition and Preprocessings
Since we adopt the inference structure of a monocular 3D object detection framework, we need to import the basic definition of anchors and we propose multiple optimized processing methods. In this subsection, we present some of the preprocessing on the input and output of the network.
III-A1 3D Anchors and Statistical Priors
Each anchor is described by 12 regressed parameters including for the 2D bounding boxes; for the 3D centers of objects on the left image; corresponding to the width, height and length of the 3D bounding boxes respectively; and to estimate the observation angle/orientation of objects.
We observe that and correspond to the same rectangular bounding box results in 3D object detection. As a result, we instead predict in the regression branch. We also add a classification channel to predict if to eliminate ambiguity, which intuitively means whether or not the object is facing the camera.
We incorporate 3D statistic priors into 2D anchors to improve the regression results. To collect prior statistics of the anchors, we iterate through the training set, and for each anchor box of different shapes, collect all the objects assigned to this anchor based on the IoU metric. Then, we compute the mean and the variance of , , for each box.
Furthermore, we explicitly exploit scene-specific knowledge for autonomous vehicles and utilize the statistical information from the anchor boxes. During training, we project dense anchor boxes into 3D with the mean depth value and filter out anchor boxes that are far away from the ground plane based on their, as displayed in Figure 2.
For multi-class training, since the statistics for different types of obstacles, e.g., cars and pedestrians, are significantly different, we compute 3D priors for each category, separately. During training, we filter out anchor boxes dynamically based on the categories assigned. During inference, we also filter out anchor boxes dynamically based on the anchors’ local categorical predictions.
III-A2 Data Augmentation for Stereo 3D Detection
Data augmentation is useful to improve the generalization ability in deep learning applications. However, the nature of stereo 3D detection limits the number of possible augmentation choices. We follow [20] to apply photometric distortion concurrently on binocular images. We also follow [22] to apply random flipping online during training. Random flipping includes flipping both RGB images, flipping the position and orientation of objects, and then switching left/right images.
III-B Multi-Scale Stereo Feature Extraction
The extraction of stereo features is one of the most time-consuming parts for many pre-existing stereo 3D object detection algorithms. In this subsection, we re-introduce the cost volume formulation based on dot-product/cosine-similarity and present the hierarchical structure to utilize these features effectively.
III-B1 Light-weight Cost Volume
Current state-of-the-art stereo matching algorithms usually construct 3D cost volume with concatenation, where the module iteratively shifts the right feature map horizontally over the left feature map, and at each step, concatenate the two features at each overlapping pixel. For binocular feature maps with the shape , the shape of the output tensor is . In this paper, we follow [10] and [13] to apply a normalized dot-product to construct a thin cost volume. Such a module compute correlation between two overlapping pixels of the feature maps instead. The shape of the output tensor becomes .
The stereo matching process can be much faster. Consider two input feature maps of , which is a common shape of a KITTI image down scaled by 4. The forward pass of concatenation-based cost volume construction takes about 200 ms while the correlation-based cost volume takes about 7 ms on an Nvidia-1080Ti.
However, the number of output channels is smaller, which could cause the network to be numerically skewed towards monocular features during the fusion stage and downsampling the stereo matching results could induce further information loss. We ease these two problems with densely connected ghost modules [23] and a hierarchical fusion structure.
III-B2 Densely Connected Ghost Module
As mentioned in Section III-B1, we need to expand the width of the features to guide the network to skewed towards features produced by stereo matching.
Han et al. propose the ghost module, which is an efficient module to produce redundant features [23]. It applies depthwise convolution to produce extra features, which requires significantly fewer parameters and FLOPs. We go one step further and densely concatenate the original input features with the output of the original ghost module, thereby tripling the number of channels. As indicated in Figure 1, the mauve blocks in (b) are the results from ghost module and others denote densly connected residuals.
Such a module preserves more information before downsampling and also rebalances the number of channels between stereo features and monocular semantic features during the fusion phase.
III-B3 Hierachical Multi-scale Fusion Structure
To minimize the information loss during the stereo matching phase while keeping the computational time tractable, we engineer a hierarchical fusion scheme. At the downsampling level of and , we construct a light-weight cost volume of a max-disparity of 96 and 192, respectively. As shown in Figure 1, they are fed into a densely connected ghost module, downsampled, and concatenated with features at a smaller scale. At a downsampling level of , we first downsample the number of channels with convolution. We then construct a small concatenation-based cost volume (also flattened to be a 2D feature map) to preserve more semantic information from the right images.
This arrangement can also be justified with high-level reasoning. Features with higher resolution are usually local features with higher frequency portions, which are suitable for dense and accurate disparity estimation. In contrast, features with low resolution contain semantic information at a larger scale.
III-C Training Scheme and Loss Function
The overall network structure is presented in Figure 1. Multi-scale features from binocular images are extracted and fused into stereo features to construct hierarchical cost volumes. The stereo feature map is concatenated with the last feature map of the left image and fed to the regression/classification branch. The stereo feature map is also fed into a decoder to predict a disparity map trained with an auxiliary loss. The auxiliary loss can regularize the training process.
III-C1 Auxiliary Disparity Supervision in Training
As pointed out by Chen et al. [16], disparity supervision is important to improve detection performance. We also observe a similar phenomenon in our framework. Without disparity supervision, the network may not be guided to produce local features useful in stereo matching to fully utilize the geometric potential of binocular images, and the network could be trapped in a local minimum similar to that of a monocular detection network.
We upsample the output of the final stereo features to , and supervise the prediction with a sparse ”ground truth” disparity derived from the traditional block matching algorithm in OpenCV [24] during training. During evaluation and testing, this disparity estimation branch is disabled to improve efficiency.
Though the disparity from the block matching algorithm is coarse and sparse, we empirically show that it significantly improves the network’s performance.
Methods | Easy/Moderate/Hard | Time |
---|---|---|
RT3DStereo[25] | 29.90 %/23.28 %/ 18.96 % | 0.08s |
StereoRCNN[22] | 47.58 %/30.23 %/ 23.72 % | 0.30s |
Pseudo-LiDAR*[19] | 54.53 %/34.05 %/ 28.25 % | 0.40s |
OC Stereo*[7] | 55.15 %/37.60 %/ 30.25 % | 0.35s |
ZoomNet*[6] | 55.98 %/38.64 %/ 30.97 % | 0.35s |
Disp R-CNN(velo)*[5] | 59.58 %/39.34 %/ 31.99 % | 0.42s |
Pseudo-LiDAR++*[3] | 61.11 %/42.43 %/ 36.99 % | 0.40s |
DSGN*[16] | 73.50 %/52.18 %/ 45.14 % | 0.67s |
Ours YOLOStereo3D | 65.68 %/41.25 %/ 30.42 % | 0.08s |
III-C2 Loss Function
We apply focal loss [26][27] on classification, and smoothed-L1 loss [28] on bounding box regression. We follow the scheme of [14] to apply stereo focal loss on the auxiliary disparity estimation. First, we compute the expected distribution of disparity with a hard-coded variance :
Where represents the disparity and indicates the predicted confidence at disparity . Then, following [14], stereo focal loss is defined as:
where is the max-disparity, is the focus weight, presents the set of pixels involved, and , represents the expected and predicted distribution map of disaparity .
The final loss function is simply the sum of the three losses.
IV Experiments
We evaluate our method on the KITTI Object Detection Benchmark [9]. The dataset consists of 7,481 training frames and 7,518 test frames. Chen et al. [29] further split the training set into 3,712 training frames and 3,769 validation frames. In this section, we provide further training details and show the performance of YOLOStereo3D on the test set to compare it with existing models.
IV-A Implementation and Training Details
Modern deep learning frameworks are sensitive to hyperparameters choices, and critical design choices could profoundly influence the final performance. We introduce some crucial design choices before showing the performance, and the code will be made open source upon publication.
We first determine the structure and the hyperparameters of the network on Chen’s split [29]. Then, we retrain the final network on the entire training set with the same hyperparameters before uploading the results for testing onto the KITTI server. An ablation study is also conducted on the validation set of Chen’s split.
The backbone of the network is ResNet-34 [30]. The top 100 pixels of each image are cropped to speed up inference and training. The cropped input images are scaled to . The network is trained with a batch size of 4 on a single Nvidia 1080Ti GPU (it takes about 7 GB of GPU memory, significantly less than other SOTA stereo detection algorithms) for 50 epochs on the KITTI training dataset. During inference, the network is fed one image at a time, and the total average processing time, including file IO, is about 0.08 s per frame. In contrast, most other stereo-based networks in the KITTI benchmark are several times slower.
IV-B Results on Test Set
The results are presented in Table I alongside those of other state-of-the-art stereo 3D detection algorithms.
The proposed YOLOStereo3D is fast and outperforms many pseudo-LiDAR methods or local point cloud methods and is the best performing algorithm without LiDAR usage. It also outperforms DSGN [16] on pedestrian detection without an additional training schedule.
Methods | 3D Easy/Moderate/Hard | Time |
---|---|---|
M3D-RPN[20] | 14.76 % / 9.71 % / 7.42 % | 0.16s |
RTM3D[31] | 14.41 % / 10.34 % / 8.77 % | 0.05s |
AM3D[32] | 16.50 % / 10.74 % / 9.52 % | 0.40s |
D4LCN[33] | 16.65 % / 11.72 % / 9.51 % | 0.20s |
Ours | 19.24 %/ 12.37 %/ 8.67 % | 0.05s |
IV-C Test results for Monocular 3D Setting
To verify the effectiveness of the proposed anchor pre-processing techniques, we further test them in the task of monocular 3D object detection. Recall that we claim YOLOStereo3D being a monocular detector enhanced with stereo features. By taking away the image from the right camera, the multi-scale fusion module, and the disparity estimation branch, we obtain a standalone monocular detector. We enhance the backbone to be ResNet-101 [30]. Following the proposed YOLOStereo3D, we compute the statistic for each anchor box and filter out deviated anchor boxes during training.
We also follow M3D-RPN [20] to post-process the prediction results to maximize the 2D-3D coherence. Notice that in YOLOStereo3D, we empirically find this post-processing step deteriorate the final performance, but it is beneficial in the monocular setting.
V Model Analysis and Discussion
In this section, we further analyze the performance of YOLOStereo3D and discuss the effectiveness of several important design choices. The baseline model here is only trained on the ”Car” type. We first conduct an ablation study to validate the contribution of anchor preprocessing, hierarchical fusion, and the densely connected ghost module on the validation set. Then, we present and discuss some qualitative results.
Methods | 3D Easy/Moderate/Hard |
---|---|
YOLOStereo3D | 72.06 %/ 46.58 %/ 35.53 % |
w/o Anchor Prior | 65.09 %/ 41.38 %/ 30.90 % |
w/o Anchor Filtering | 71.37 %/ 45.03 %/ 35.83 % |
w/o Channel Expand | 64.16 %/ 39.96 %/ 30.02 % |
w Naive Channel-expand | 70.70 %/ 45.74 %/ 34.87 % |
w/o Scale 8 | 70.80 %/ 45.71 %/ 35.86 % |
w/o Scale 16 | 68.64 %/ 44.54 %/ 33.95 % |
w/o Disparity supervision | 62.58 %/ 39.09 %/ 30.34 % |
w PC supervision | 72.05 %/ 46.59 %/ 35.62 % |
YOLOMono3D | 21.66 %/ 14.20 %/ 11.07 % |
Mono w/o Prior | 19.90 %/ 13.36 %/ 9.68 % |
Mono w/o Filtering | 20.50 %/ 13.45 %/ 10.50 % |

V-A Ablation Study
V-A1 Anchor Preprocessing
We first test the effectiveness of including statistical information in each anchor. In the first experiment, instead of predicting a depth value normalized by the depth prior, the network outputs a transformed depth output , where , following [34]. In the second experiment, we do not filter out anchors during training, and the training loss is evaluated with all anchors. We conduct these two experiments in both the stereo setting and the monocular setting. The results are presented in Table IV respectively.
From the two table, we can observe that anchor priors significantly boost the performance of the network, and filtering out irrelevant anchors during training is also helpful. The performance gain can be observed in both monocular 3D detection and stereo 3D detection. We suggest that we can ease the difficulty of depth inferencing by properly defining and preprocessing anchors specifically for 3D scene understanding in autonomous driving.
The improvement we apply on anchors can also be applied and verified in monocular 3D detection. We further provide ablation experiments to validate the effectiveness of these processing methods under monocular 3D detection setting.
V-A2 Densely Connected Ghost Module
Densely-connected ghost modules are useful in expanding the number of channels in stereo processing. We conduct two experiments to verify its effectiveness. In the first experiment, we use a BasicBlock in resnet [30] to replace the ghost module without expanding the number of channels, resulting in fewer channels during the fusion between RGB features and stereo features. In the second experiment, we directly upsample the number of channels with convolution before feeding the tensor into a BasicBlock.
We can observe from Table IV that the densely-connected ghost module is useful in improving the network’s capability. From the first experiment, we demonstrate that expanding the number of channels is crucial for the network’s performance. In the second experiment, we further show that the densely-connected ghost module is better at preserving information than the naive convolution.
V-A3 Hierachical Fusion
We respectively disable the stereo matching output on scale 8/16 to produce two networks to justify the usage of multi-scale fusion. We can observe from Table IV that the results of the two ablated models are inferior to that of the baseline model. We also point out that the forward pass of stereo matching modules on scale 8/16 is several times faster than that on scale 4.
As a result, we argue that hierarchically fusing stereo features from scale 8 and 16 is worth the effort. The baseline structure of hierarchical fusion in YOLOStereo3D achieves a fair balance between speed and performance.
V-A4 Disparity Supervision
We also have an ablation study on the importance of disparity supervision. Similar to the conclusion in DSGN [16], disparity supervision significantly boosts the performance of the network.
In the experiment, we show that such supervision is essential, but the results are not sensitive to the accuracy of the ”target” disparity map. The insight is that the network may only need slight regularizations in stereo matching submodules. The auxiliary loss is required to drive the network from falling back to a naive local optimal of monocular 3D object detection.
V-B Qualitative Results
We show qualitative validation results in Figure 3. The model displayed is YOLOStereo3D sharing the same hyperparameters as the model submitted to the KITTI server, but it is only trained on the training sub-split.
From the RGB images, we can observe that most of the successful predictions of YOLOStereo3D are visually consistent with the context. As shown in the bird’s-eye-view images, though the disparity estimation may not correctly align with the ground truth 3D bounding boxes, the bounding box predictions from YOLOStereo3D are still reasonably accurate.
The examples suggest that priors in anchor heads and the fusion between stereo matching features and RGB features could help the network to produce more visually consistent predictions and make the network more robust against potentially misleading disparity matching results.
VI Conclusion
In this paper, we presented YOLOStereo3D, an efficient stereo 3D object detection framework. This work’s major contribution is to take a step back to consider stereo 3D object detection as an enhanced monocular detection problem, rather than as an inaccurate LiDAR-based detection problem. To achieve this, we first incorporated knowledge from real-time monocular 3D object detection frameworks and used priors in anchors for depth inference. Then, we introduced the point-wise correlation module into the detection problems. Finally, we used a hierarchical fusion framework that balances information preservation and computational burden. We tested YOLOStereo3D on the KITTI Object Detection Benchmark. The model produces competitive results among stereo frameworks while running at more than ten frames per second without any usage of LiDAR data.
It should be noted that we are converting the features from the right image to the left image. In other words, the computational roles of the two images are not equal. Information loss in the right image is significant. As a result, when an object is occluded in the left image but is more visible in the right image, the model could be significantly sub-optimal.
Nevertheless, since the model can achieve a competitive result with only one GPU and a short training time, YOLOStereo3D lowers the bar of stereo detection research. With a significantly faster inference speed and competitive performance, YOLOStereo3D can also boost the deployment of stereo setups on self-driving cars and mobile robots in the future.
References
- [1] B. Yang, M. Liang, and R. Urtasun, “Hdnet: Exploiting hd maps for 3d object detection,” in Proceedings of The 2nd Conference on Robot Learning, ser. Proceedings of Machine Learning Research, A. Billard, A. Dragan, J. Peters, and J. Morimoto, Eds., vol. 87. PMLR, 29–31 Oct 2018, pp. 146–155. [Online]. Available: http://proceedings.mlr.press/v87/yang18b.html
- [2] Y. Zhou, P. Sun, Y. Zhang, D. Anguelov, J. Gao, T. Ouyang, J. Guo, J. Ngiam, and V. Vasudevan, “End-to-end multi-view fusion for 3d object detection in lidar point clouds,” in Proceedings of the Conference on Robot Learning, ser. Proceedings of Machine Learning Research, L. P. Kaelbling, D. Kragic, and K. Sugiura, Eds., vol. 100. PMLR, 30 Oct–01 Nov 2020, pp. 923–932. [Online]. Available: http://proceedings.mlr.press/v100/zhou20a.html
- [3] Y. You, Y. Wang, W.-L. Chao, D. Garg, G. Pleiss, B. Hariharan, M. Campbell, and K. Q. Weinberger, “Pseudo-lidar++: Accurate depth for 3d object detection in autonomous driving,” in International Conference on Learning Representations (ICLR), 2019.
- [4] R. Qian, D. Garg, Y. Wang, Y. You, S. Belongie, B. Hariharan, M. Campbell, K. Q. Weinberger, and W.-L. Chao, “End-to-end pseudo-lidar for image-based 3d object detection,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [5] J. Sun, L. Chen, Y. Xie, S. Zhang, Q. Jiang, and a. H. B. Xiaowei Zhou, “Disp r-cnn: Stereo 3d object detection via shape prior guided instance disparity estimation,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [6] Z. Xu, W. Zhang, X. Ye, X. Tan, W. Yang, S. Wen, E. Ding, A. Meng, and L. Huang, “Zoomnet: Part-aware adaptive zooming neural network for 3d object detection,” arXiv preprint arXiv:2003.00529, 2020.
- [7] A. D. Pon, J. Ku, C. Li, and S. L. Waslander, “Object-centric stereo matching for 3d object detection,” arXiv preprint arXiv:1909.07566, 2019.
- [8] J. Chang and Y. Chen, “Pyramid stereo matching network,” CoRR, vol. abs/1803.08669, 2018. [Online]. Available: http://arxiv.org/abs/1803.08669
- [9] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2012.
- [10] W. Luo, A. G. Schwing, and R. Urtasun, “Efficient deep learning for stereo matching,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 5695–5703.
- [11] J. Zbontar and Y. LeCun, “Stereo matching by training a convolutional neural network to compare image patches,” CoRR, vol. abs/1510.05970, 2015. [Online]. Available: http://arxiv.org/abs/1510.05970
- [12] A. Kendall, H. Martirosyan, S. Dasgupta, P. Henry, R. Kennedy, A. Bachrach, and A. Bry, “End-to-end learning of geometry and context for deep stereo regression,” CoRR, vol. abs/1703.04309, 2017. [Online]. Available: http://arxiv.org/abs/1703.04309
- [13] Q. Wang, S. Shi, S. Zheng, K. Zhao, and X. Chu, “Fadnet: A fast and accurate network for disparity estimation,” arXiv preprint arXiv:2003.10758, 2020.
- [14] Y. Zhang, Y. Chen, X. Bai, S. Yu, K. Yu, Z. Li, and K. Yang, “Adaptive unimodal cost volume filtering for deep stereo matching,” arXiv preprint arXiv:1909.03751, 2019.
- [15] Y. Wang, W.-L. Chao, D. Garg, B. Hariharan, M. Campbell, and K. Weinberger, “Pseudo-lidar from visual depth estimation: Bridging the gap in 3d object detection for autonomous driving,” arXiv preprint arXiv:1812.07179, 2018.
- [16] Y. Chen, S. Liu, X. Shen, and J. Jia, “Dsgn: Deep stereo geometry network for 3d object detection,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [17] T. van Dijk and G. C. H. E. de Croon, “How do neural networks see depth in single images?” CoRR, vol. abs/1905.07005, 2019. [Online]. Available: http://arxiv.org/abs/1905.07005
- [18] J. M. U. Vianney, S. Aich, and B. Liu, “Refinedmpl: Refined monocular pseudolidar for 3d object detection in autonomous driving,” 11 2019.
- [19] X. Weng and K. Kitani, “Monocular 3d object detection with pseudo-lidar point cloud,” CoRR, vol. abs/1903.09847, 2019. [Online]. Available: http://arxiv.org/abs/1903.09847
- [20] G. Brazil and X. Liu, “M3D-RPN: monocular 3d region proposal network for object detection,” CoRR, vol. abs/1907.06038, 2019. [Online]. Available: http://arxiv.org/abs/1907.06038
- [21] Y. Liu, Y. Yixuan, and M. Liu, “Ground-aware monocular 3d object detection for autonomous driving,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 919–926, 2021.
- [22] P. Li, X. Chen, and S. Shen, “Stereo R-CNN based 3d object detection for autonomous driving,” CoRR, vol. abs/1902.09738, 2019. [Online]. Available: http://arxiv.org/abs/1902.09738
- [23] K. Han, Y. Wang, Q. Tian, J. Guo, C. Xu, and C. Xu, “Ghostnet: More features from cheap operations,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [24] B. G., “The OpenCV Library,” Dr. Dobb’s Journal of Software Tools, 2000.
- [25] K. Hendrik, S. Niels, and S. Christoph, “Realtime 3d object detection for automated driving using stereo vision and semantic information,” 10 2019, pp. 1405–1410.
- [26] P. Yun, L. Tai, Y. Wang, and M. Liu, “Focal loss in 3d object detection,” CoRR, vol. abs/1809.06065, 2018. [Online]. Available: http://arxiv.org/abs/1809.06065
- [27] T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollar, “Focal loss for dense object detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. PP, pp. 1–1, 07 2018.
- [28] R. B. Girshick, “Fast R-CNN,” CoRR, vol. abs/1504.08083, 2015. [Online]. Available: http://arxiv.org/abs/1504.08083
- [29] X. Chen, K. Kundu, Y. Zhu, A. G. Berneshawi, H. Ma, S. Fidler, and R. Urtasun, “3d object proposals for accurate object class detection,” in Advances in Neural Information Processing Systems 28, C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, Eds. Curran Associates, Inc., 2015, pp. 424–432. [Online]. Available: http://papers.nips.cc/paper/5644-3d-object-proposals-for-accurate-object-class-detection.pdf
- [30] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” CoRR, vol. abs/1512.03385, 2015. [Online]. Available: http://arxiv.org/abs/1512.03385
- [31] P.-X. Li, H. Zhao, P. Liu, and F. Cao, “Rtm3d: Real-time monocular 3d detection from object keypoints for autonomous driving,” ArXiv, vol. abs/2001.03343, 2020.
- [32] X. M. and/ Zhihui Wang, H. Li, W. Ouyang, and P. Zhang, “Accurate monocular 3d object detection via color-embedded 3d reconstruction for autonomous driving,” CoRR, vol. abs/1903.11444, 2019. [Online]. Available: http://arxiv.org/abs/1903.11444
- [33] M. Ding, Y. Huo, H. Yi, Z. Wang, J. Shi, Z. Lu, and P. Luo, “Learning depth-guided convolutions for monocular 3d object detection,” 12 2019.
- [34] Y. Chen, L. Tai, K. Sun, and M. Li, “Monopair: Monocular 3d object detection using pairwise spatial relationships,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2020.