This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

S3M-Net: Joint Learning of Semantic Segmentation and Stereo Matching for Autonomous Driving

Zhiyuan Wu\scalerel —{}^{\href https://orcid.org/0009-0001-7253-7603\,}, Yi Feng\scalerel —{}^{\href https://orcid.org/0009-0005-4885-0850\,}, Chuang-Wei Liu\scalerel —{}^{\href https://orcid.org/0000-0003-0260-6236\,}, Fisher Yu\scalerel —{}^{\href https://orcid.org/0000-0001-8829-7344\,},
Qijun Chen\scalerel —{}^{\href https://orcid.org/0000-0001-5644-1188\,}, , and Rui Fan\scalerel —{}^{\href https://orcid.org/0000-0003-2593-6596\,}
This research was supported by the Science and Technology Commission of Shanghai Municipal under Grant 22511104500, the National Natural Science Foundation of China under Grant 62233013, the Fundamental Research Funds for the Central Universities, and Xiaomi Young Talents Program. (Corresponding author: Rui Fan)Zhiyuan Wu, Yi Feng, Chuang-Wei Liu, Qijun Chen, and Rui Fan are with the College of Electronics & Information Engineering, Shanghai Research Institute for Intelligent Autonomous Systems, the State Key Laboratory of Intelligent Autonomous Systems, and Frontiers Science Center for Intelligent Autonomous Systems, Tongji University, Shanghai 201804, China (e-mails: [email protected], [email protected], {cwliu, qjchen, rfan}@tongji.edu.cn).Fisher Yu is with the Department of Information Technology and Electrical Engineering, ETH Zürich, Sternwartstrasse 7, 8092 Zürich, Switzerland. email: [email protected].
Abstract

Semantic segmentation and stereo matching are two essential components of 3D environmental perception systems for autonomous driving. Nevertheless, conventional approaches often address these two problems independently, employing separate models for each task. This approach poses practical limitations in real-world scenarios, particularly when computational resources are scarce or real-time performance is imperative. Hence, in this article, we introduce S3M-Net, a novel joint learning framework developed to perform semantic segmentation and stereo matching simultaneously. Specifically, S3M-Net shares the features extracted from RGB images between both tasks, resulting in an improved overall scene understanding capability. This feature sharing process is realized using a feature fusion adaption (FFA) module, which effectively transforms the shared features into semantic space and subsequently fuses them with the encoded disparity features. The entire joint learning framework is trained by minimizing a novel semantic consistency-guided (SCG) loss, which places emphasis on the structural consistency in both tasks. Extensive experimental results conducted on the vKITTI2 and KITTI datasets demonstrate the effectiveness of our proposed joint learning framework and its superior performance compared to other state-of-the-art single-task networks. Our project webpage is accessible at mias.group/S3M-Net.

Index Terms:
semantic segmentation, stereo matching, environmental perception, autonomous driving, joint learning.

I Introduction

3D environmental perception stands as a critical and foundational aspect of autonomous driving [1, 2]. Semantic segmentation and stereo matching are two key functionalities in 3D environmental perception systems [3, 54, 5]. The former provides a comprehensive pixel-level understanding of the environment, while the latter simulates human binocular vision to acquire accurate and dense depth information [6]. The combined utilization of both functionalities has become the mainstream approach in recent years [7, 8, 9, 10, 11, 12].

In recent years, the research focus in semantic segmentation has shifted from single-modal networks [13, 14, 15, 16, 17, 18] with a single encoder to feature-fusion networks with dual encoders [19, 20, 21, 22]. The latter type of networks extract heterogeneous features from RGB-X data, where “X” can represent various forms of spatial geometric information, e.g., depth images generated from LiDAR point clouds and surface normal maps obtained through depth-to-normal translation [23]. These heterogeneous features are subsequently fused to achieve a more comprehensive understanding of the environment [21]. However, a critical drawback of feature-fusion networks is their dependency on the availability of the “X” data, which can pose limitations in scenarios where LiDARs are not present. Additionally, when the accuracy of the “X” data is not satisfactory, such as due to variations in camera-LiDAR calibration, the fusion of these heterogeneous features can potentially lead to a degradation in the overall performance of semantic segmentation [24]. While a stereo camera can serve as a practical and cost-effective alternative to LiDARs for depth information acquisition, the incorporation of a separate stereo matching network introduces additional computations, and therefore, poses difficulties in achieving real-time processing speeds for the entire system [9]. Moreover, stereo matching and semantic segmentation share the same input, and the representations from RGB images can be more informative when they are jointly learned by both tasks.

The joint learning of multiple interconnected 3D environmental perception tasks introduces a form of regularization that has demonstrated superiority over uniform complexity penalization in reducing over-fitting [25]. Furthermore, rather than employing separate models for semantic segmentation and stereo matching, joint learning can potentially reduce computational complexity [7, 8, 9, 10, 11, 12], as shared learning representations can be used for both tasks. This can be advantageous in real-time or resource-constrained applications. Moreover, joint learning enables end-to-end optimization of the entire system, allowing the model to adapt to the specific challenges of both tasks simultaneously. Consequently, this can lead to improved performance when compared to models trained separately for each task [12]. In addition, stereo matching can occasionally produce ambiguously estimated disparities, particularly in texture-less or occluded regions [26]. Semantic segmentation can provide informative contextual information that helps disambiguate such cases, ultimately leading to more reliable disparity estimations [9]. Regrettably, the joint learning of semantic segmentation and stereo matching, especially within feature-fusion networks or when faced with a scarcity of training samples, has received relatively limited attention in this research area and calls for further investigation.

Therefore, in this article, we present Semantic Segmentation and Stereo Matching Network (S3M-Net), a joint framework to simultaneously predict both semantic and disparity information. S3M-Net begins with the extraction of features from stereo images. These features are then processed by a multi-level gate recurrent unit (GRU) operator to generate a disparity map. Simultaneously, these features are shared with the semantic segmentation task via a feature fusion adaptation (FFA) module. Building upon our prior work SNE-RoadSeg [3], we extract additional features from the estimated disparity map. Finally, a densely-connected skip connection decoder is employed to decode the fused features and generate the semantic predictions. S3M-Net is trained in a fully supervised manner by minimizing a semantic consistency-guided (SCG) joint learning loss. Extensive experiments conducted on the vKITTI2 [27] and KITTI 2015 [28] datasets unequivocally demonstrate the effectiveness and superior performance of our proposed S3M-Net.

In summary, the main contributions of this article include:

  • S3M-Net, a joint learning framework designed to address semantic segmentation and stereo matching simultaneously, where both tasks collaboratively leverage the features extracted from RGB images, enhancing the overall understanding of the driving scenario;

  • A feature fusion adaption module to transform the shared feature maps into semantic space and subsequently fuse them with encoded disparity features;

  • A semantic consistency-guided loss function to supervise the training process of the joint learning framework, emphasizing on the structural consistency in both tasks.

The remainder of this article is organized as follows: Sect. II provides a review of related work. Sect. III introduces our proposed S3M-Net. Sect. IV presents the experimental results and compares our framework with other state-of-the-art (SoTA) approaches. In Sect. V, we discuss the advantages and limitations of our method. Finally, we conclude this article in Sect. VI.

II Literature Review

II-A Semantic Segmentation

Semantic segmentation has been a long-standing problem in the field of computer vision over the past decade [6, 29]. The SoTA networks in this research area can generally be classified into two categories: (1) single-modal networks with a single encoder and (2) feature-fusion networks with multiple encoders [3, 30, 31]. In the early attempts to tackle semantic segmentation, researchers primarily focused on encoder-decoder architectures for pixel-level classification. Notable examples include SegNet [13], U-Net [14], PSPNet [15], the DeepLab series [16, 17], and Transformer-based networks [32, 33, 34]. The encoder extracts hierarchical deep features from the input image, while the decoder produces the segmentation map by upsampling and combining the features from different encoder layers. However, these networks are limited in their ability to effectively combine deep features extracted from different modalities (or sources) of visual information. As a result, they often struggle to produce accurate segmentation results in challenging scenarios characterized by poor lighting and illumination conditions [3]. Therefore, researchers have turned their focus towards feature-fusion networks that can effectively integrate deep features learned from multiple modalities (or sources) of visual information. This problem is commonly referred to as “RGB-X semantic segmentation”, where “X” represents the additional modality (or source) of visual information, in addition to the RGB images. The most representative feature-fusion networks based on convolutional neural networks (CNNs) include FuseNet [19], MFNet [35], RTFNet [20], and our previous works SNE-RoadSeg series [3, 21]. Furthermore, Transformer-based RGB-X semantic segmentation networks, such as OFF-Net [22] and RoadFormer [24], have been recently introduced. In this article, we design our S3M-Net based on the SNE-RoadSeg architecture and explore more effective solutions for the feature fusion operation.

II-B Stereo Matching

Conventional explicit programming-based stereo matching algorithms (local, global, and semi-global) generally consist of four main procedures: (1) cost computation, (2) cost aggregation, (3) disparity optimization, and (4) disparity refinement [26]. The performance of these algorithms has been significantly outperformed by end-to-end deep stereo networks, thanks to the recent advancements in deep learning techniques. PSMNet [36], GwcNet [37], AANet [38], LEA-Stereo [39], RAFT-Stereo [40], and CRE-Stereo [41] are six representative end-to-end deep stereo networks proposed in recent years. PSMNet [36] employs a spatial pyramid to capture multi-scale information and employs multiple 3D convolutional layers to exploit both local and global contexts for cost computation. GwcNet [37], on the other hand, builds upon the foundation of PSMNet by constructing the cost volume via group-wise correlation, thereby enhancing the 3D stacked hourglass network. In light of the computational demands of 3D convolutions, researchers have actively sought ways to minimize the trade-off between efficiency and accuracy in stereo matching. For example, LEA-Stereo [39] introduces the neural architecture search (NAS) [42] technique to stereo matching. This pioneering approach results in the first end-to-end hierarchical NAS framework for deep stereo matching. RAFT-Stereo [40], a rectified stereo matching method that draws inspiration from the optical flow estimation network RAFT [43], leverages the RAFT architecture to perform accurate and real-time stereo matching inference. The network utilizes recurrent structures to refine correlation features and enhance the disparity estimation accuracy. CRE-Stereo [41], another recent prior art based on recurrent refinement (to update disparities in a coarse-to-fine manner) and adaptive group correlation (to mitigate the impact of erroneous rectification), yields more compelling disparity estimation results. In this article, we develop our S3M-Net based on the RAFT-Stereo architecture.

Refer to caption
Figure 1: The architecture of our proposed S3M-Net for end-to-end joint learning of semantic segmentation and stereo matching.

II-C Multi-Task Joint Learning for Semantic Segmentation and Stereo Matching

Existing frameworks that jointly address semantic segmentation and stereo matching generally focus on improving disparity accuracy by leveraging semantic information [7, 8, 9, 11, 12, 10], while the discussion regarding the utilization of disparity information to enhance semantic segmentation at the feature level for joint learning remains limited, except for the exploration of “RGB-X semantic segmentation” discussed in Sect. II-A. Nevertheless, these prior arts either require a large amount of well-annotated training data or involve intricate training strategies for the joint learning of both tasks. For instance, SegStereo [7] and DispSegNet [8] require an initial unsupervised training phase on the large-scale Cityscapes [44] dataset, followed by a subsequent supervised fine-tuning on the smaller KITTI 2012 and 2015 [45, 28] datasets. Similarly, the studies presented in [9, 11, 12] involve the pre-training of their spatial branches (performing stereo matching) on the large-scale SceneFlow [46] dataset, followed by the fine-tuning of both semantic and spatial branches on the KITTI 2012 and 2015 datasets [45, 28]. DSNet [10] adopts a different joint learning strategy in which the training alternates between the semantic segmentation and stereo matching networks, with each network being frozen during the training of the other. However, achieving simultaneous convergence of the two networks can be challenging, as the shared features are not learned in an end-to-end manner. Additionally, we were unable to locate publicly available source code (in PyTorch or TensorFlow) for these prior arts, and re-implementations carry the risk of introducing errors. In contrast to the aforementioned approaches, our proposed S3M-Net is trained in an end-to-end fashion and capable of jointly learning semantic segmentation and stereo matching even when the training data are limited.

III Methodology

As illustrated in Fig. 1, our proposed S3M-Net consists of five main components:

  1. (1)

    Joint encoder to extract shared features from RGB images;

  2. (2)

    Multi-level GRU update operator to refine disparity maps;

  3. (3)

    Feature fusion adaptation module to transform shared features into the semantic space and fuse them with features extracted from the disparity maps;

  4. (4)

    Densely-connected skip connection decoder to decode fused features and produce final semantic predictions;

  5. (5)

    Semantic consistency-guided loss to supervise the entire joint learning process.

III-A Joint Encoder

Given a pair of well-rectified stereo images 𝑰L\boldsymbol{I}^{L}, 𝑰RH×W×3\boldsymbol{I}^{R}\in\mathbb{R}^{H\times W\times 3}, where HH and WW denote their height and width, respectively, we employ a joint encoder consisting of a series of residual blocks and downsampling layers to extract features L={𝑭1L,,𝑭nL}\mathcal{F}^{L}=\left\{\boldsymbol{F}^{L}_{1},\dots,\boldsymbol{F}^{L}_{n}\right\} and R={𝑭1R,,𝑭nR}\mathcal{F}^{R}=\left\{\boldsymbol{F}^{R}_{1},\dots,\boldsymbol{F}^{R}_{n}\right\} from 𝑰L\boldsymbol{I}^{L} and 𝑰R\boldsymbol{I}^{R}, respectively. L\mathcal{F}^{L} is subsequently shared with the semantic segmentation task.

III-B Multi-Level GRU Update Operator

Using the features L\mathcal{F}^{L} and R\mathcal{F}^{R} extracted by the joint encoder, we first construct an initial 3D correlation volume 𝑪1H×W×W\boldsymbol{C}_{1}\in\mathbb{R}^{H\times W\times W} as follows:

𝑪1(i,j,k)=𝑭nL(i,j,:)𝑭nR(i,k,:),\boldsymbol{C}_{1}(i,j,k)={\boldsymbol{F}^{L}_{n}}(i,j,:)\cdot{\boldsymbol{F}^{R}_{n}}(i,k,:), (1)

where ii represents the ii-th row, and jj and kk represent to the jj-th and kk-th columns in the left and right shared feature maps, respectively. We then construct a pyramid of 3D correlation volumes 𝒞={𝑪1,,𝑪m}\mathcal{C}=\left\{\boldsymbol{C}_{1},\dots,\boldsymbol{C}_{m}\right\} by downsampling 𝑪1\boldsymbol{C}_{1} with average pooling operations. The mm-th 3D correlation volume 𝑪mH×W×W2m1\boldsymbol{C}_{m}\in\mathbb{R}^{H\times W\times\frac{W}{2^{m-1}}} is constructed from the (m1m-1)-th 3D correlation volume 𝑪m1\boldsymbol{C}_{m-1} using 1D average pooling with a kernel size of 2 and a stride of 2. Inspired by RAFT-Stereo [40], we adopt a multi-level GRU update operator to refine a sequence of disparity maps 𝒟={𝑫1,,𝑫n}\mathcal{D}=\left\{\boldsymbol{D}_{1},\dots,\boldsymbol{D}_{n}\right\}, where 𝑫iH×W\boldsymbol{D}_{i}\in\mathbb{R}^{H\times W} (i=1,,ni=1,\dots,n). This refinement process is performed in a coarse-to-fine manner, starting from an initial disparity map 𝑫0\boldsymbol{D}_{0} in which all disparities are initialized to 0.

III-C Feature Fusion Adaptation Module

Refer to caption
Figure 2: An illustration of our proposed FFA module.

In stereo matching, a lower number of channels, e.g., the 256 channels utilized in RAFT-Stereo [40], is often sufficient for capturing relevant features for 1D correspondence search, especially when considering computational efficiency. On the other hand, semantic segmentation requires pixel-level classification and a more in-depth scene understanding. It benefits from complex feature representations that can capture fine-grained details and object boundaries, making a larger number of channels, e.g., the 2048 channels employed in SNE-RoadSeg [3], advantageous for this task. Therefore, we introduce the FFA module to align the channels and resolutions between the disparity and semantic feature maps during joint learning.

As illustrated in Fig 2, given the left shared feature maps L={𝑭1L,,𝑭nL}\mathcal{F}^{L}=\left\{\boldsymbol{F}^{L}_{1},\dots,\boldsymbol{F}^{L}_{n}\right\} and the disparity map pyramid 𝒟={𝑫1,,𝑫n}\mathcal{D}=\left\{\boldsymbol{D}_{1},\dots,\boldsymbol{D}_{n}\right\}, we obtain the adapted fused feature sequence F={𝑭1F,,𝑭nF}\mathcal{F}^{F}=\left\{\boldsymbol{F}^{F}_{1},\dots,\boldsymbol{F}^{F}_{n}\right\} using our proposed FFA module, which can be formulated as follows:

𝑭iF=𝒜i(L)iD(𝑫n),\boldsymbol{F}^{F}_{i}=\mathcal{A}_{i}(\mathcal{F}^{L})\oplus\mathcal{E}^{D}_{i}(\boldsymbol{D}_{n}), (2)

where D\mathcal{E}^{D} denotes the disparity map encoding operation, \oplus denotes the feature fusion operation, and 𝒜i\mathcal{A}_{i} is defined as our feature adaptation operation, as formulated as follows:

𝒜i(L)={(𝑭2i1L),in+12(𝑭i1Fi1D(𝑫n)),i>n+12,\mathcal{A}_{i}(\mathcal{F}^{L})=\left\{\begin{aligned} &\mathcal{R}(\boldsymbol{F}^{L}_{2i-1}),&&i\leq\frac{n+1}{2}\\ &\mathcal{E}(\boldsymbol{F}^{F}_{i-1}\oplus\mathcal{E}^{D}_{i-1}(\boldsymbol{D}_{n})),&&i>\frac{n+1}{2}\end{aligned},\right. (3)

where \mathcal{R} represents the remapping operation from the shared feature space to the semantic feature space, and \mathcal{E} represents the encoding operation for the semantic feature maps.

Specifically, for the remapping operation \mathcal{R}, we employ 3×33\times 3 convolutional layers with a stride of 2 and padding of 1, each followed by a batch normalization layer and a rectified linear unit (ReLU) activation layer, adapting the feature map channels to 64, 256, and 512, respectively. Regarding the disparity encoding operation D\mathcal{E}^{D}, we employ ResNet-152 [47] as the backbone network to extract features from the last disparity map 𝑫n\boldsymbol{D}_{n}. In ResNet-152, the first block consists of a convolutional layer, a batch normalization layer, and a ReLU activation layer. Then, a max pooling layer and four residual layers are sequentially applied to progressively increase the number of feature map channels.

Similarly, we utilize the residual block for the encoding operation \mathcal{E} on the semantic feature maps, resulting in feature maps with 1024 and 2048 channels. The fused features F\mathcal{F}^{F} contain both texture and spatial geometric information, thereby enhancing semantic scene understanding. We conduct an ablation study for different feature fusion modules in Sect. IV-F.

III-D Densely-Connected Skip Connection Decoder

We employ the decoder introduced in our previous work SNE-RoadSeg [3] to decode the fused features and generate the semantic prediction. In this encoder, three convolutional layers in the feature extractor and the upsampling layer share the same parameters: a 3×33\times 3 kernel size, a stride of 1, and a padding of 1. In the final layer, features are upsampled to create the prediction map with NN channels, where NN denotes the number of semantic classes.

III-E Semantic Consistency-Guided Joint Learning Loss

The loss function employed in our joint learning framework should guide the supervision of both the semantic segmentation and stereo matching tasks. Gradient smoothness between the disparity and semantic segmentation maps typically aligns closely, particularly at inter-class boundaries, where traditional training strategies tend to result in more errors due to factors such as occlusion and reflection. In light of this, we propose an SCG loss function to supervise the entire joint learning process, which leverages semantic consistency to optimize the training of S3M-Net.

Given the ground-truth semantic segmentation map 𝑴GH×W\boldsymbol{M}^{G}\in\mathbb{R}^{H\times W}, Each pixel 𝒑\boldsymbol{p} of 𝑴G\boldsymbol{M}^{G} can be written as follows:

𝑴G(𝒑){1,,C},\boldsymbol{M}^{G}(\boldsymbol{p})\in\{1,\dots,C\}, (4)

where CC refers to the number of the semantic classes. We construct an extended 3D volume 𝑽3DH×W×C{\boldsymbol{V}^{3D}}\in\mathbb{R}^{H\times W\times C} using the following expression:

𝑽c3D(𝒑)=δ(𝑴G(𝒑),c),\boldsymbol{V}^{3D}_{c}(\boldsymbol{p})=\delta(\boldsymbol{M}^{G}(\boldsymbol{p}),c), (5)

where cc represents the cc-th channel in the volume, and δ\delta denotes the Kronecker Delta function [48]. As a result, each channel of the volume can be regarded as a binary segmentation map of the cc-th class. To emphasize semantic consistency, We use an average pooling operation for each channel to obtain the inter-class volume 𝑽IH×W×C\boldsymbol{V}^{I}\in\mathbb{R}^{H\times W\times C}:

𝑽I=𝒫(𝑽3D),\boldsymbol{V}^{I}=\mathcal{P}(\boldsymbol{V}^{3D}), (6)

where 𝒫\mathcal{P} denotes the average pooling operation. Furthermore, we apply a normalization operation:

𝑽N(𝒑)=e(2𝑽I(𝒑)1)2.\boldsymbol{V}^{N}(\boldsymbol{p})=e^{-(2\boldsymbol{V}^{I}(\boldsymbol{p})-1)^{2}}. (7)

to obtain a normalized volume 𝑽NH×W×C\boldsymbol{V}^{N}\in\mathbb{R}^{H\times W\times C}. We then map 𝑽N\boldsymbol{V}^{N} to a semantic consistency-guided weight map 𝑾H×W\boldsymbol{W}\in\mathbb{R}^{H\times W} through:

𝑾(𝒑)=maxc{𝑽cN(𝒑)}.\boldsymbol{W}(\boldsymbol{p})=\operatorname*{max}\limits_{c}\left\{\boldsymbol{V}_{c}^{N}(\boldsymbol{p})\right\}. (8)

The total loss function

scg=ss+sm\mathcal{L}_{scg}=\mathcal{L}_{ss}+\mathcal{L}_{sm} (9)

consists of an SCG semantic segmentation loss ss\mathcal{L}_{ss} and an SCG stereo matching loss sm\mathcal{L}_{sm}. ss\mathcal{L}_{ss} is formulated as follows:

ss=1Ni=1Nc=1C[(1α)+α𝑾(𝒑)]yc(𝒑)log(y^c(𝒑)),\mathcal{L}_{ss}=-\frac{1}{N}\sum_{i=1}^{N}\sum_{c=1}^{C}[(1-\alpha)+\alpha\boldsymbol{W}(\boldsymbol{p})]y_{c}(\boldsymbol{p})\log(\hat{y}_{c}(\boldsymbol{p})), (10)

where NN denotes the pixel number, CC represents the class number, y^c(𝒑)\hat{y}_{c}(\boldsymbol{p}) denotes the predicted probability of 𝒑\boldsymbol{p} belonging to class cc, yc(𝒑)y_{c}(\boldsymbol{p}) represents the ground-truth label for 𝒑\boldsymbol{p} in class cc, and α\alpha denotes the loss weight. Based on the ablation study detailed in Sect. IV-F, we set the value of α\alpha to 0.1. Moreover, sm\mathcal{L}_{sm} is formulated as follows:

sm=i=1N[(1α)+α𝑾(𝒑)]γNi𝑫G𝑫i1,\mathcal{L}_{sm}=\sum_{i=1}^{N}[(1-\alpha)+\alpha\boldsymbol{W}(\boldsymbol{p})]\gamma^{N-i}\left\|\boldsymbol{D}^{G}-\boldsymbol{D}_{i}\right\|_{1}, (11)

where 𝑫G\boldsymbol{D}^{G} represents the ground-truth disparity map and 𝑫i\boldsymbol{D}_{i} denotes the ii-th disparity map in 𝒟\mathcal{D}. α\alpha is set to 0.1 and γ\gamma is set to 0.9.

TABLE I: Comparisons of SoTA semantic segmentation networks on the vKITTI2 [27] dataset. The symbol \uparrow indicates that a higher value corresponds to better performance. The best results are shown in bold font.
Category Networks Acc (%) \uparrow mAcc (%) \uparrow mIoU (%) \uparrow fwIoU (%) \uparrow Pre (%) \uparrow Rec (%) \uparrow FSc (%) \uparrow
Single-Modal SegNet [13] 59.29 32.54 23.93 48.17 66.10 66.73 61.11
U-Net [14] 62.71 37.65 29.83 55.10 75.80 67.67 65.26
PSPNet [15] 76.26 53.53 44.81 69.30 81.55 79.68 75.38
DeepLabv3+ [17] 92.19 63.15 56.90 87.15 89.00 92.71 90.16
HRNet [49] 74.79 40.82 32.47 63.23 73.69 76.50 73.39
BiSeNet V2 [50] 81.77 51.07 44.45 74.71 83.23 82.19 80.67
Segmenter [32] 90.39 60.33 52.99 83.47 88.05 87.89 87.70
SegFormer [33] 94.75 70.56 64.98 90.49 93.57 93.62 93.46
Mask2Former [34] 89.29 64.58 57.14 83.84 90.75 87.23 87.19
DDRNet [51] 70.80 40.32 32.10 61.44 76.35 71.67 70.57
Feature-Fusion FuseNet [19] 49.42 31.21 22.56 41.07 79.39 50.67 47.50
MFNet [35] 76.22 51.50 43.41 68.82 82.46 78.65 73.80
RTFNet [20] 85.22 49.47 42.59 77.69 83.74 89.17 84.41
SNE-RoadSeg [3] 83.64 60.85 52.56 75.14 83.44 81.66 77.77
OFF-Net [22] 90.84 61.51 55.27 84.69 89.24 86.71 86.15
RoadFormer [24] 97.54 86.58 80.83 95.34 96.99 96.86 96.91
S3M-Net 98.27 88.28 84.25 96.92 98.29 98.32 98.28
S3M-Net w/ SCG loss 98.32 88.24 84.18 96.98 98.37 98.28 98.31
TABLE II: Comparisons of SoTA semantic segmentation networks on the KITTI 2015 [28] dataset. The symbol \uparrow indicates that a higher value corresponds to better performance. The best results are shown in bold font.
Category Networks Acc (%) \uparrow mAcc (%) \uparrow mIoU (%) \uparrow fwIoU (%) \uparrow Pre (%) \uparrow Rec (%) \uparrow FSc (%) \uparrow
Single-Modal SegNet [13] 59.63 31.98 22.61 43.98 55.25 67.49 57.29
U-Net [14] 69.02 41.15 30.64 55.65 69.11 77.65 71.04
PSPNet [15] 80.03 44.97 38.15 68.62 79.29 82.66 79.59
DeepLabv3+ [17] 82.33 50.15 42.79 72.43 83.85 87.18 84.59
HRNet [49] 63.42 31.68 22.78 49.40 55.10 67.71 57.21
BiSeNet V2 [50] 73.68 41.66 32.71 60.55 68.35 81.79 72.37
Segmenter [32] 84.53 50.77 43.63 74.72 82.99 87.15 84.41
SegFormer [33] 88.28 59.23 51.39 80.53 87.15 90.85 88.46
Mask2Former [34] 84.35 54.33 45.87 75.56 84.74 89.12 85.92
DDRNet [51] 62.12 31.61 22.63 48.15 57.09 68.98 59.07
Feature-Fusion FuseNet [19] 41.79 19.05 11.38 27.53 44.14 44.35 37.68
MFNet [35] 81.02 48.13 40.70 70.42 82.85 85.73 82.36
RTFNet [20] 71.61 39.26 30.35 57.98 69.52 85.16 74.28
SNE-RoadSeg [3] 79.46 51.91 41.56 69.22 81.45 87.05 82.91
OFF-Net [22] 75.84 40.13 33.13 64.02 77.48 72.19 70.62
RoadFormer [24] 90.05 62.34 55.13 83.40 91.65 91.39 91.11
S3M-Net 90.01 62.48 54.33 83.44 88.96 93.52 90.65
S3M-Net w/ SCG loss 90.66 65.90 57.80 84.53 90.85 93.55 91.80
Refer to caption
Figure 3: Qualitative experimental results of semantic segmentation on the vKITTI2 [27] dataset: (a) RGB images; (b)-(k) semantic segmentation results achieved by SegNet [13], U-Net [14], PSPNet [15], DeepLabv3+ [17], HRNet [49], BiSeNet V2 [50], Segmenter [32], SegFormer [33], Mask2Former [34], and DDRNet [51], respectively; (l)-(q) semantic segmentation results achieved by FuseNet [19], MFNet [35], RTFNet [20], SNE-RoadSeg [3], OFF-Net [22], and RoadFormer [24], respectively; (r)-(s) semantic segmentation results achieved by our proposed S3M-Net w/o and w/ the use of the SCG loss, respectively; (t) ground truth annotations.
Refer to caption
Figure 4: Qualitative experimental results of semantic segmentation on the KITTI 2015 [28] dataset: (a) RGB images; (b)-(k) semantic segmentation results achieved by SegNet [13], U-Net [14], PSPNet [15], DeepLabv3+ [17], HRNet [49], BiSeNet V2 [50], Segmenter [32], SegFormer [33], Mask2Former [34], and DDRNet [51], respectively; (l)-(q) semantic segmentation results achieved by FuseNet [19], MFNet [35], RTFNet [20], SNE-RoadSeg [3], OFF-Net [22], and RoadFormer [24], respectively; (r)-(s) semantic segmentation results achieved by our proposed S3M-Net w/o and w/ the use of the SCG loss, respectively; (t) ground truth annotations.
Refer to caption
Figure 5: Qualitative experimental results of stereo matching on the vKITTI2 [27] dataset: (a) left RGB images; (b)-(j) disparity maps estimated using PSMNet [36], GwcNet [37], AANet [38], LEA-Stereo [39], RAFT-Stereo [40], CRE-Stereo [41], ACVNet [52], PCWNet[53], and IGEV-Stereo [54], respectively; (k)-(l) disparity maps estimated using our proposed S3M-Net w/o and w/ the use of the SCG loss, respectively.
Refer to caption
Figure 6: Qualitative experimental results of stereo matching on the KITTI 2015 [28] dataset: (a) left RGB images; (b)-(j) disparity maps estimated using PSMNet [36], GwcNet [37], AANet [38], LEA-Stereo [39], RAFT-Stereo [40], CRE-Stereo [41], ACVNet [52], PCWNet[53], and IGEV-Stereo [54], respectively; (k)-(l) disparity maps estimated using our proposed S3M-Net w/o and w/ the use of the SCG loss, respectively.
Refer to caption
Figure 7: Ablation study on the selection of α\alpha in the SCG joint learning loss function on the KITTI 2015 [28] dataset.

IV Experiments

In this article, we conduct extensive experiments to evaluate the performance of our developed S3M-Net both quantitatively and qualitatively. The following subsections provide details on the used datasets, experimental set-up, evaluation metrics, and the comprehensive evaluation of our proposed method.

IV-A Datasets

Since the training of our network requires both semantic and disparity annotations, we employ two public datasets to evaluate its performance: the vKITTI2 [27] dataset (synthetic yet large-scale) and the KITTI 2015 [28] dataset (real-world yet modest-scale). Their details are as follows:

  • The vKITTI2 dataset contains virtual replicas of five sequences from the KITTI dataset. It provides 15 classes for the semantic segmentation tasks. Dense ground-truth disparity maps are acquired through depth rendering using a virtual engine. In our experiments, we randomly select 700 pairs of stereo images, along with their semantic and disparity annotations to evaluate the effectiveness and robustness of our proposed S3M-Net, where 500 pairs are used for model training and the remaining 200 pairs are used for model validation.

  • The KITTI 2015 dataset contains 400 pairs of stereo images captured in real-world driving scenarios, with 200 pairs containing ground truth and the other 200 pairs lacking ground truth. It provides 19 classes for the semantic segmentation tasks (in alignment with the Cityscapes [44] dataset). Sparse disparity ground truth is obtained using a Velodyne HDL-64E LiDAR. In our experiments, we allocate 70% of the dataset for training, while the remaining portion is used as the test set.

IV-B Experimental Setup

Our experiments are conducted on an NVIDIA RTX 3090 GPU. The batch size is set to 1. The maximum disparity search range is set to 192 pixels. All images are cropped to 1000 ×\times 320 pixels before feeding into the network. We utilize the AdamW [55] optimizer for model training, setting the epsilon and weight decay parameters to 10810^{-8} and 10510^{-5}, respectively. The initial learning rate is set to 2×1042\times 10^{-4}. Training lasts for 100K iterations on the vKITTI2 dataset and 20K iterations on the KITTI 2015 dataset. We employ traditional data augmentation techniques to enhance the robustness of the models.

IV-C Evaluation Metrics

Since our proposed S3M-Net simultaneously performs semantic segmentation and stereo matching, we evaluate the performance of both tasks in our experiments.

We utilize seven evaluation metrics to quantify the performance of semantic segmentation: (1) accuracy (Acc), (2) mean accuracy (mAcc), (3) mean intersection over union (mIoU), (4) frequency-weighted intersection over union (fwIoU) [56], (5) precision (Pre), (6) recall (Rec), and (7) F1-score (FSc).

Additionally, we use two evaluation metrics: (1) the average end-point error (EPE) and (2) the percentage of error pixels (PEP), setting the tolerance for the latter to 1.0 and 3.0 pixels, respectively, to quantify the performance of stereo matching.

IV-D Semantic Segmentation Performance

The qualitative experimental results on the vKITTI2 and KITTI datasets are presented in Figs. 3 and 4, respectively, while the quantitative experimental results on the vKITTI2 and KITTI datasets are given in Tables I and II, respectively. These results suggest that S3M-Net outperforms other SoTA single-modal and feature-fusion networks (including CNN-based and Transformer-based methods) across all evaluation metrics on both datasets. Specifically, it is noteworthy that when the entire joint learning framework is trained by minimizing our proposed SCG loss, S3M-Net achieves the best performance on the KITTI dataset across all evaluation metrics except for Pre. Compared with SoTA methods, it shows improvements of 5.71%5.71\% in mAcc, 4.84%4.84\% in mIoU, 1.35%1.35\% in fwIoU, and 0.76%0.76\% in FSc, respectively. Similarly, it outperforms other networks on the vKITTI2 dataset in most evaluation metrics, with improvements of 1.72%1.72\% in fwIoU and 1.44%1.44\% in FSc. However, for mAcc, mIoU, and Rec, its performance is comparable to that of S3M-Net trained without using the SCG loss. Additionally, it is obvious that feature-fusion networks consistently outperform single-modal networks, particularly under challenging weather and lighting conditions. This observation aligns with our expectations, as feature-fusion networks leverage both RGB images and disparity maps, allowing them to effectively learn informative spatial geometric representations. However, SoTA feature-fusion networks may exhibit higher error rates in distant regions. For instance, FuseNet and SNE-RoadSeg demonstrate poor performance in the sky. We attribute this phenomenon to the deep structure of the encoders, where distinguishing distant objects using disparity features becomes challenging, and the feature fusion process amplifies the influence of the disparity feature. In contrast, within our proposed joint learning framework, we can extract more informative features benefiting from both tasks, irrespective of dataset size. This improvement is likely due to the fact that joint learning of multiple interconnected tasks introduces a form of regularization, which has shown its superiority over uniform complexity penalization in reducing over-fitting.

TABLE III: Comparisons of SoTA stereo matching network on the vKITTI2 [27] dataset. The symbol \downarrow indicates that a lower value corresponds to better performance. The best results are shown in bold font.
Networks EPE (pixels) \downarrow PEP (%) \downarrow
>1>1 pixel >3>3 pixels
PSMNet [36] 0.68 10.31 3.77
GwcNet [37] 0.65 9.72 3.69
AANet [38] 1.36 15.61 6.98
LEA-Stereo [39] 0.83 13.33 4.84
RAFT-Stereo [40] 0.40 5.88 2.67
CRE-Stereo [41] 0.63 10.35 3.90
ACVNet [52] 0.61 9.41 3.45
PCW-Net [53] 0.63 9.45 3.49
IGEV-Stereo [54] 0.47 7.15 3.09
S3M-Net 0.39 5.59 2.55
S3M-Net w/ SCG loss 0.38 5.56 2.55
 
TABLE IV: Comparisons of SoTA stereo matching network on the KITTI 2015 [28] dataset. The symbol \downarrow indicates that a lower value corresponds to better performance. The best results are shown in bold font.
Networks EPE (pixels) \downarrow PEP (%) \downarrow
>1>1 pixel >3>3 pixels
PSMNet [36] 0.74 16.12 2.61
GwcNet [37] 0.68 14.21 2.01
AANet [38] 1.10 22.67 5.37
LEA-Stereo [39] 0.83 18.67 3.22
RAFT-Stereo [40] 0.60 10.78 1.96
CRE-Stereo [41] 0.92 19.68 3.35
ACVNet [52] 0.68 13.93 2.10
PCW-Net [53] 0.70 14.81 2.43
IGEV-Stereo [54] 0.62 12.15 1.99
S3M-Net 0.56 10.33 1.72
S3M-Net w/ SCG loss 0.55 10.02 1.62
 

IV-E Stereo Matching Performance

The qualitative experimental results on the vKITTI2 and KITTI datasets are given in Figs. 5 and 6, respectively, while the quantitative experimental results on the vKITTI2 and KITTI datasets are presented in Tables III and IV, respectively. These results suggest that S3M-Net outperforms other SoTA stereo matching networks across all evaluation metrics on both datasets. Specifically, S3M-Net trained with and without using the SCG loss achieves the top and second-best overall performances, respectively. S3M-Net, when trained without using the SCG loss, demonstrates improvements of 2.50%2.50\%-71.32%71.32\% in EPE, 4.17%4.17\%-64.19%64.19\% in PEP 1.0, and 4.49%4.49\%-67.97%67.97\% in PEP 3.0. On the other hand, S3M-Net, when trained with the SCG loss, shows improvements of 5.00%5.00\%-72.06%72.06\% in EPE, 5.44%5.44\%-64.38%64.38\% in PEP 1.0, and 4.49%4.49\%-69.83%69.83\% in PEP 3.0. We attribute these improvements to the feature sharing and fusion strategies applied in S3M-Net. First, sharing features with the semantic segmentation task allows S3M-Net to learn stereo matching effectively even with limited training data. Second, as discussed above, stereo matching can sometimes produce ambiguous disparity estimations, especially in occluded or texture-less areas. The pursuit of semantic consistency helps resolve such ambiguities, leading to more reliable disparity estimation results. In Fig. 6, it is evident that regions lacking disparity ground truth frequently have substantial errors. Previous stereo matching algorithms have endeavored to tackle this issue through knowledge distillation with pre-trained models [38]. Nevertheless, our S3M-Net successfully overcomes this challenge by leveraging semantic information.

TABLE V: Ablation study on feature fusion strategy in our FFA module on the KITTI [28] 2015 dataset.
Fusion Strategy Acc (%) \uparrow mAcc (%) \uparrow mIoU (%) \uparrow fwIoU (%) \uparrow Pre (%) \uparrow Rec (%) \uparrow FSc (%) \uparrow
Addition 90.01 62.48 54.33 83.44 88.96 93.52 90.65
Concatenation 86.88 57.43 48.40 78.50 85.96 93.71 88.92
CFM [57] 86.87 57.41 48.77 79.13 85.41 92.05 87.63
DDPM [58] 86.52 58.65 49.51 78.14 85.56 93.34 88.44
SA Gate [59] 87.62 61.77 52.10 80.55 86.74 92.31 88.52
SWS [60, 61, 62] 87.94 58.11 49.64 80.25 86.63 93.34 89.20

IV-F Ablation studies

In this subsection, we first conduct an ablation study on the selection of loss weight α\alpha in (11). Fig. 7 shows the quantitative experimental results with respect to different α\alpha in the range of 0.00.0 to 0.40.4 for both semantic segmentation and stereo matching. It can be obvious that when α=0.1\alpha=0.1, S3M-Net achieves the best overall performance for both tasks. Further weight tuning is possible, but it should be approached cautiously, especially when dealing with limited data to avoid over-fitting.

Furthermore, we conduct an additional ablation study on the feature fusion strategy in our proposed FFA module. As shown in Table V, when using the addition operation to fuse heterogeneous features, the FFA module consistently achieves the best performance across all evaluation metrics, compared to other feature fusion strategies, including concatenation, cross feature module (CFM) [57], dynamic dilated pyramid module (DDPM) [58], separation-and-aggregation gate (SA Gate) [59], and softmax weighted sum (SWS) used in AysmFusion [60], CEN [61], and TokenFusion [62].

V Discussion

The experimental results shown in Sect. IV provide strong support for the claims made in Sect. I. First, the joint learning of semantic segmentation and stereo matching, two interconnected environmental perception tasks, using our proposed S3M-Net introduces a form of regularization that has shown its effectiveness in reducing overfitting, particularly in scenarios where training data are limited. Secondly, this end-to-end joint learning framework yields improved performance when compared to the models trained separately for each task. Finally, the pursuit of semantic consistency in joint learning helps reduce ambiguous disparity estimations in texture-less or occluded regions. We believe that our proposed S3M-Net can be readily deployed in autonomous vehicles after addressing the following limitations:

  • S3M-Net requires both semantic and disparity annotations, and collecting data with such ground truth remains a labor-intensive process. Therefore, the exploration of potential solutions such as semi-supervised or low/few-shot semantic segmentation and un/self-supervised stereo matching is a promising avenue for future research.

  • S3M-Net achieves a processing speed of 0.66 fps when processing input RGB images with a resolution of 1248 ×\times 384 pixels. We believe that further computational efficiency optimizations are necessary before deploying S3M-Net in autonomous vehicles.

VI Conclusion

This article introduced S3M-Net, an effective solution for joint learning of semantic segmentation and stereo matching. We have made three significant contributions in this work: (1) the development of an entire joint learning framework that shares features between both tasks and fuses heterogeneous features to improve semantic segmentation, (2) a feature fusion adaption module designed to enable effective feature sharing between the two tasks, and (3) a semantic consistency-guided joint learning loss that emphasizes structural consistency in both tasks. We conducted extensive experiments on the vKITTI2 (synthetic and large) and KITTI (real-world and small) datasets to validate the effectiveness of our framework, the FFA module, and the training loss. Our results demonstrate the superior performance of our approach compared to all other existing methods.

References

  • [1] B. Ranft and C. Stiller, “The role of machine vision for intelligent vehicles,” IEEE Transactions on Intelligent vehicles, vol. 1, no. 1, pp. 8–19, 2016.
  • [2] D. L. Fisher, M. Lohrenz, D. Moore, E. D. Nadler, and J. K. Pollard, “Humans and intelligent vehicles: The hope, the help, and the harm,” IEEE Transactions on Intelligent Vehicles, vol. 1, no. 1, pp. 56–67, 2016.
  • [3] R. Fan, H. Wang, P. Cai, and M. Liu, “SNE-RoadSeg: Incorporating surface normal information into semantic segmentation for accurate freespace detection,” in European Conference on Computer Vision (ECCV).   Springer, 2020, pp. 340–356.
  • [4] G. Xu, X. Wang, X. Ding, and X. Yang, “Iterative geometry encoding volume for stereo matching,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 21 919–21 928.
  • [5] Z. Rao, B. Xiong, M. He, Y. Dai, R. He, Z. Shen, and X. Li, “Masked representation learning for domain generalized stereo matching,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 5435–5444.
  • [6] R. Fan, S. Guo, and M. J. Bocus, Autonomous Driving Perception.   Springer, 2023.
  • [7] G. Yang, H. Zhao, J. Shi, Z. Deng, and J. Jia, “SegStereo: Exploiting semantic information for disparity estimation,” in European Conference on Computer Vision (ECCV), 2018, pp. 636–651.
  • [8] J. Zhang, K. A. Skinner, R. Vasudevan, and M. Johnson-Roberson, “DispSegNet: Leveraging semantics for end-to-end learning of disparity estimation from stereo imagery,” IEEE Robotics and Automation Letters, vol. 4, no. 2, pp. 1162–1169, 2019.
  • [9] Z. Wu, X. Wu, X. Zhang, S. Wang, and L. Ju, “Semantic stereo matching with pyramid cost volumes,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019, pp. 7484–7493.
  • [10] W. Zhan, X. Ou, Y. Yang, and L. Chen, “DSNet: Joint learning for scene segmentation and disparity estimation,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 2946–2952.
  • [11] P. L. Dovesi, M. Poggi, L. Andraghetti, M. Martí, H. Kjellström, A. Pieropan, and S. Mattoccia, “Real-time semantic stereo matching,” in 2020 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2020, pp. 10 780–10 787.
  • [12] S. Chen, Z. Xiang, C. Qiao, Y. Chen, and T. Bai, “SGNet: Semantics guided deep stereo matching,” in Proceedings of the Asian Conference on Computer Vision (ACCV), 2020, pp. 106–122.
  • [13] V. Badrinarayanan, A. Kendall, and R. Cipolla, “SegNet: A deep convolutional encoder-decoder architecture for image segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 12, pp. 2481–2495, 2017.
  • [14] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI).   Springer, 2015, pp. 234–241.
  • [15] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 2881–2890.
  • [16] L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” arXiv preprint arXiv:1706.05587, 2017.
  • [17] L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 801–818.
  • [18] Z. Huang, X. Wang, L. Huang, C. Huang, Y. Wei, and W. Liu, “CCNet: Criss-cross attention for semantic segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019, pp. 603–612.
  • [19] C. Hazirbas, L. Ma, C. Domokos, and D. Cremers, “FuseNet: Incorporating depth into semantic segmentation via fusion-based cnn architecture,” in Proceedings of the Asian Conference on Computer Vision (ACCV).   Springer, 2017, pp. 213–228.
  • [20] Y. Sun, W. Zuo, and M. Liu, “RTFNet: RGB-Thermal fusion network for semantic segmentation of urban scenes,” IEEE Robotics and Automation Letters, vol. 4, no. 3, pp. 2576–2583, 2019.
  • [21] H. Wang, R. Fan, P. Cai, and M. Liu, “SNE-RoadSeg+: Rethinking depth-normal translation and deep supervision for freespace detection,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2021, pp. 1140–1145.
  • [22] C. Min, W. Jiang, D. Zhao, J. Xu, L. Xiao, Y. Nie, and B. Dai, “ORFD: A dataset and benchmark for off-road freespace detection,” in 2022 International Conference on Robotics and Automation (ICRA).   IEEE, 2022, pp. 2532–2538.
  • [23] J. Yang, B. Xue, Y. Feng, D. Wang, R. Fan, and Q. Chen, “Three-filters-to-normal+: Revisiting discontinuity discrimination in depth-to-normal translation,” IEEE Transactions on Automation Science and Engineering, 2024, DOI: 10.1109/TASE.2024.3355941.
  • [24] J. Li, Y. Zhang, P. Yun, G. Zhou, Q. Chen, and R. Fan, “RoadFormer: Duplex Transformer for RGB-normal semantic road scene parsing,” arXiv preprint arXiv:2309.10356, 2023.
  • [25] R. Fan, C. Bowd, N. Brye, M. Christopher, R. N. Weinreb, D. J. Kriegman, and L. M. Zangwill, “One-vote veto: Semi-supervised learning for low-shot glaucoma diagnosis,” IEEE Transactions on Medical Imaging, 2023, DOI: 10.1109/TMI.2023.3307689.
  • [26] R. Fan, X. Ai, and N. Dahnoun, “Road surface 3D reconstruction based on dense subpixel disparity map estimation,” IEEE Transactions on Image Processing, vol. 27, no. 6, pp. 3025–3035, 2018.
  • [27] Y. Cabon, N. Murray, and M. Humenberger, “Virtual KITTI 2,” arXiv preprint arXiv:2001.10773, 2020.
  • [28] M. Menze and A. Geiger, “Object scene flow for autonomous vehicles,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 3061–3070.
  • [29] U. Michieli, M. Biasetton, G. Agresti, and P. Zanuttigh, “Adversarial learning and self-teaching techniques for domain adaptation in semantic segmentation,” IEEE Transactions on Intelligent Vehicles, vol. 5, no. 3, pp. 508–518, 2020.
  • [30] Y. Yang, C. Shan, F. Zhao, W. Liang, and J. Han, “On exploring shape and semantic enhancements for RGB-X semantic segmentation,” IEEE Transactions on Intelligent Vehicles, 2023.
  • [31] J. Fan, F. Wang, H. Chu, X. Hu, Y. Cheng, and B. Gao, “MLFNet: Multi-level fusion network for real-time semantic segmentation of autonomous driving,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 1, pp. 756–767, 2022.
  • [32] R. Strudel, R. Garcia, I. Laptev, and C. Schmid, “Segmenter: Transformer for semantic segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 7262–7272.
  • [33] E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, “SegFormer: Simple and efficient design for semantic segmentation with transformers,” Advances in Neural Information Processing Systems (NeurIPS), vol. 34, pp. 12 077–12 090, 2021.
  • [34] B. Cheng, I. Misra, A. G. Schwing, A. Kirillov, and R. Girdhar, “Masked-attention mask transformer for universal image segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 1290–1299.
  • [35] Q. Ha, K. Watanabe, T. Karasawa, Y. Ushiku, and T. Harada, “MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2017, pp. 5108–5115.
  • [36] J.-R. Chang and Y.-S. Chen, “Pyramid stereo matching network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 5410–5418.
  • [37] X. Guo, K. Yang, W. Yang, X. Wang, and H. Li, “Group-wise correlation stereo network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 3273–3282.
  • [38] H. Xu and J. Zhang, “AANet: Adaptive aggregation network for efficient stereo matching,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 1959–1968.
  • [39] X. Cheng, Y. Zhong, M. Harandi, Y. Dai, X. Chang, H. Li, T. Drummond, and Z. Ge, “Hierarchical neural architecture search for deep stereo matching,” Advances in Neural Information Processing Systems (NeurIPS), vol. 33, pp. 22 158–22 169, 2020.
  • [40] L. Lipson, Z. Teed, and J. Deng, “RAFT-Stereo: Multilevel recurrent field transforms for stereo matching,” in 2021 International Conference on 3D Vision (3DV).   IEEE, 2021, pp. 218–227.
  • [41] J. Li, P. Wang, P. Xiong, T. Cai, Z. Yan, L. Yang, J. Liu, H. Fan, and S. Liu, “Practical stereo matching via cascaded recurrent network with adaptive correlation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 16 263–16 272.
  • [42] T. Elsken, J. H. Metzen, and F. Hutter, “Neural architecture search: A survey,” The Journal of Machine Learning Research, vol. 20, no. 1, pp. 1997–2017, 2019.
  • [43] Z. Teed and J. Deng, “RAFT: Recurrent all-pairs field transforms for optical flow,” in European Conference on Computer Vision (ECCV).   Springer, 2020, pp. 402–419.
  • [44] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele, “The CityScapes dataset for semantic urban scene understanding,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 3213–3223.
  • [45] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the KITTI vision benchmark suite,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).   IEEE, 2012, pp. 3354–3361.
  • [46] N. Mayer, E. Ilg, P. Hausser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox, “A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 4040–4048.
  • [47] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
  • [48] S. Okada, M. Ohzeki, and S. Taguchi, “Efficient partition of integer optimization problems with one-hot encoding,” Scientific Reports, vol. 9, no. 1, p. 13036, 2019.
  • [49] K. Sun, B. Xiao, D. Liu, and J. Wang, “Deep high-resolution representation learning for human pose estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 5693–5703.
  • [50] C. Yu, C. Gao, J. Wang, G. Yu, C. Shen, and N. Sang, “BiSeNet V2: Bilateral network with guided aggregation for real-time semantic segmentation,” International Journal of Computer Vision, vol. 129, pp. 3051–3068, 2021.
  • [51] Y. Hong, H. Pan, W. Sun, and Y. Jia, “Deep dual-resolution networks for real-time and accurate semantic segmentation of road scenes,” arXiv preprint arXiv:2101.06085, 2021.
  • [52] G. Xu, J. Cheng, P. Guo, and X. Yang, “Attention concatenation volume for accurate and efficient stereo matching,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 12 981–12 990.
  • [53] Z. Shen, Y. Dai, X. Song, Z. Rao, D. Zhou, and L. Zhang, “PCW-Net: Pyramid combination and warping cost volume for stereo matching,” in European Conference on Computer Vision (ECCV).   Springer, 2022, pp. 280–297.
  • [54] G. Xu, X. Wang, X. Ding, and X. Yang, “Iterative geometry encoding volume for stereo matching,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 21 919–21 928.
  • [55] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” arXiv preprint arXiv:1711.05101, 2017.
  • [56] H. Zhao, X. Puig, B. Zhou, S. Fidler, and A. Torralba, “Open vocabulary scene parsing,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 2002–2010.
  • [57] J. Wei, S. Wang, and Q. Huang, “F3Net: fusion, feedback and focus for salient object detection,” in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 2020, pp. 12 321–12 328.
  • [58] Y. Pang, L. Zhang, X. Zhao, and H. Lu, “Hierarchical dynamic filtering network for RGB-D salient object detection,” in European Conference on Computer Vision (ECCV).   Springer, 2020, pp. 235–252.
  • [59] X. Chen, K.-Y. Lin, J. Wang, W. Wu, C. Qian, H. Li, and G. Zeng, “Bi-directional cross-modality feature propagation with separation-and-aggregation gate for RGB-D semantic segmentation,” in European Conference on Computer Vision (ECCV).   Springer, 2020, pp. 561–577.
  • [60] Y. Wang, F. Sun, M. Lu, and A. Yao, “Learning deep multimodal feature representation with asymmetric multi-layer fusion,” in Proceedings of the 28th ACM International Conference on Multimedia (ACM MM), 2020, pp. 3902–3910.
  • [61] Y. Wang, W. Huang, F. Sun, T. Xu, Y. Rong, and J. Huang, “Deep multimodal fusion by channel exchanging,” Advances in Neural Information Processing Systems (NeurIPS), vol. 33, pp. 4835–4845, 2020.
  • [62] Y. Wang, X. Chen, L. Cao, W. Huang, F. Sun, and Y. Wang, “Multimodal token fusion for vision transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 12 186–12 195.