Airborne LiDAR Point Cloud Classification with Graph Attention Convolution Neural Network
Abstract
Airborne light detection and ranging (LiDAR) plays an increasingly significant role in urban planning, topographic mapping, environmental monitoring, power line detection and other fields thanks to its capability to quickly acquire large-scale and high-precision ground information. To achieve point cloud classification, previous studies proposed point cloud deep learning models that can directly process raw point clouds based on PointNet-like architectures. And some recent works proposed graph convolution neural network based on the inherent topology of point clouds. However, the above point cloud deep learning models only pay attention to exploring local geometric structures, yet ignore global contextual relationships among all points. In this paper, we present a graph attention convolution neural network (GACNN) that can be directly applied to the classification of unstructured 3D point clouds obtained by airborne LiDAR. Specifically, we first introduce a graph attention convolution module that incorporates global contextual information and local structural features. The global attention module examines spatial relationships among all points, while the local attention module can dynamically learn convolution weights with regard to the spatial position of the local neighboring points and reweight the convolution weights by inspecting the density of each local region. Based on the proposed graph attention convolution module, we further design an end-to-end encoder-decoder network, named GACNN, to capture multiscale features of the point clouds and therefore enable more accurate airborne point cloud classification. Experiments on the ISPRS 3D labeling dataset show that the proposed model achieves a new state-of-the-art performance in terms of average F1 score (71.5%) and a satisfying overall accuracy (83.2%). Additionally, experiments further conducted on the 2019 Data Fusion Contest Dataset by comparing with other prevalent point cloud deep learning models demonstrate the favorable generalization capability of the proposed model.
keywords:
Airborne LiDAR, Point cloud classification, Point cloud Deep learning, Graph Attention Convolution, ISPRS 3D labeling1 Introduction
Airborne light detection and ranging (LiDAR), as one of the most important techniques for data collection in earth observation (EO) systems, has the advantages of quickly acquiring large-scale and high-precision ground information, and plays an increasingly important role in urban planning (Yu et al.,, 2010), topographic mapping (Krabill et al.,, 1984; Liu,, 2008; Axelsson,, 2000), environmental monitoring (Huang et al.,, 2014; Bradbury et al.,, 2005), and power line detection (Sohn et al.,, 2012; Zhu and Hyyppä,, 2014), etc. By employing airborne LiDAR for city scanning, a massive and irregular spatially distributed 3D point cloud with coordinates (X, Y, Z) and certain properties (e.g. intensity) can be acquired directly, the classification of which is an important research direction in the field of photogrammetry and remote sensing. However, achieving automatic airborne LiDAR point cloud classification with high precision in real applications is challenging due to the high variability of object classes and complex object structure (Chen,, 2007; Niemeyer et al.,, 2012).
Early research has mainly focused on solving the problem of airborne LiDAR point cloud classification by applying traditional machine learning-based models. These methods usually start with designing hand-crafted features, such as geometry features, radiometry features, topology features, echo features, and full waveform features, and then conduct point cloud classification by employing machine learning-based classifiers, including Support Vector Machine (SVM) (Zhang et al.,, 2013), Adaboost (Lodha et al.,, 2007), Random Forest (Chehata et al.,, 2009), Markov Random Field (Munoz et al.,, 2009; Shapovalov et al.,, 2010) and Conditional Random Field (Niemeyer et al.,, 2011, 2014). Nevertheless, the calculation of these handcrafted features requires specific expert knowledge and has limited ability to extract effective features of the original point cloud data.
In recent years, deep learning models (LeCun et al.,, 2015) have drawn considerable attention from researchers due to their great success in various applications, such as natural language processing (Collobert and Weston,, 2008), speech recognition (Hinton et al.,, 2012), time series prediction (Wen et al.,, 2019), and image classification (Chan et al.,, 2015), etc. Convolutional neural networks (CNN), one of the most prevalent models in deep learning can receive only regular inputs; therefore, early studies focusing on point cloud classification mostly transform point clouds to regular 3D voxels or collections of 2D feature images (Maturana and Scherer,, 2015; Yang et al.,, 2017, 2018; Zhao et al.,, 2018). Nonetheless, this transformation process leads to inefficient computation and substantial memory consumption, and the transformation to 2D feature images causes spatial information loss.
A deep learning model called PointNet, which can directly consume raw point clouds by exploiting multilayer perceptron (MLP) and max pooling to obtain the global feature representation (Qi et al., 2017a, ), is recently proposed for point cloud classifcation. The PointNet++ model, which first generates the partitioning of the point set and then builds a hierarchical neural network that employs PointNet recursively on a nested partitioning of the point cloud (Qi et al., 2017b, ), is furhter presented. Similarly, various methods have been proposed to further improve the performance of point cloud classification by exploring local structure of point clouds (Li et al.,, 2018; Jiang et al.,, 2018; Thomas et al.,, 2019). More recently, a number of researchers introduce graph convolution neural network to classify point clouds based on the inherent topology of point clouds (Te et al.,, 2018; Wang et al., 2018a, ; Wang et al., 2019b, ). Inspired by the visual attention mechanism, some work further present graph attention convolution network to learn adaptive local geometric structures and conduct point cloud classification (Wang et al., 2019a, ; Chen et al.,, 2019).
However, the above point-based deep learning models in the field of computer vision only pay attention to explore local geometric structures, and ignore global contextual relationships among all points. In addition, few studies have investigated how to achieve airborne LiDAR point cloud classification by directly employing deep learning models to process raw point clouds (Yousefhussien et al.,, 2017; Wang et al., 2018b, ; Wen et al.,, 2020).
In this paper, we present a graph attention convolution neural network (GACNN) for airborne LiDAR point cloud classification. Specifically, the graph attention convolution module includes two types of attention mechanisms: a local attention module that combines edge attention and density attention, and a global attention module. The local edge attention module is designed to dynamically learn convolution weights in light of the spatial position relationships of neighboring points; thus, the receptive field of the convolution kernel can dynamically adjust to the structure of the point cloud. The local density attention module is devised to remedy the problem of the uneven density distribution of non-uniform sampled point cloud data. Moreover, to learn global contextual information of the point clouds, we implement a global attention module by calculating Euclidean distance between every two individual points and utilize an MLP network to learn their attention weights. Furthermore, based on the presented graph attention convolution module, we develop an encoder-decoder network that can accept arbitrary sizes of input points and be trained in an end-to-end manner for airborne LiDAR point cloud classification. The key contributions of our work are summed up as follows:
-
1.
We propose GACNN, a graph attention convolution neural network for airborne LiDAR point cloud classification that can be applied directly to raw point clouds to predict the semantic labels for arbitrarily sized input point clouds.
-
2.
We design a local graph attention module that combines edge attention and density attention. The proposed edge attention module can adapt to the structure of the point cloud by dynamically adjusting kernel weights via learning from the local spatial layouts of neighboring points. The proposed density attention module can overcome problem of the uneven density distribution of non-uniform sampled point cloud data.
-
3.
We introduce a global graph attention module by taking the global spatial distribution among all points into consideration to capture global contextual features, which can further improve the performance of the proposed network.
-
4.
The presented model realizes state-of-the-art classification performance on both ISPRS 3D Labeling Dataset and 2019 Data Fusion Contest Dataset.
The remaining part of this paper is structured as follows. In Section 2, a brief summary of point cloud classification methods is given with special regard to airborne LiDAR point clouds. We introduce the proposed GACNN in detail in Section 3. In Section 4, we conduct experiments to evaluate the performance of GACNN on the ISPRS 3D labeling benchmark dataset. The effect of the attention modules, and the superiority and generalization capability of the proposed model are discussed in Section 5. Finally, we conclude in Section 6.
2 Related Work
2.1 Non-deep learning method
Traditionally, studies accomplish airborne LiDAR point cloud classification by calculating handcrafted features and employing the classic machine learning models. Zhang et al., adopt the surface growing algorithm to cluster point clouds and utilize SVM to classify segments according to thirteen features related to radiometry, geometry, topology and echo characteristics (Zhang et al.,, 2013). Lodha et al., select five features, namely, image intensity, LiDAR return intensity, normal variation, height, and height variation, and implement AdaBoost to classify 3D airborne LiDAR data into four groups: buildings, trees, grass, and road (Lodha et al.,, 2007). Chehata et al., apply Random Forests to select the most important features from multi-echo and full-waveform features, and classify airborne LiDAR point clouds on urban scenes (Chehata et al.,, 2009). However, these models ignore the contextual information of point clouds and predict the semantic label for each point individually, which creates noises in the classification results and inconsistency in the labels.
To resolve this issue, Niemeyer et al., implement Conditional Random Field (CRF) method, which can be employed to incorporate contextual information and learn relations among objects, to classify airborne LiDAR point clouds based on geometrical and intensity features (Niemeyer et al.,, 2011). Besides, Niemeyer et al., exploit ten types of features obtained from LiDAR point clouds and integrate Random Forest classifier into CRF to achieve reliable classification results, even in complicated urban scenes (Niemeyer et al.,, 2014).
Although the above non-deep learning methods achieve satisfying classification performance, they require manual calculation of features in advance, and the results are sensitive to the choice of features.
2.2 Deep learning method on grids or collections of images
Deep learning, a novel machine learning method proposed in recent years, can automatically learn effective feature representations from a large amount of input data. CNNs, one of the most essential deep learning models, have made considerable progress in image classification tasks, such as object detection, semantic segmentation, and edge detection. However, because CNNs can handle only standard regular input format, direct application of CNNs to irregular and unordered 3D point clouds is infeasible.
To apply CNNs to airborne LiDAR point cloud classification, most researchers first convert the point cloud to regular 3D voxel grids or collections of images. Maturana and Scherer, develop the Voxnet model to transform point clouds to 3D voxel grids, followed by a 3D CNN that predicts the semantic label based on the occupancy grid (Maturana and Scherer,, 2015). However, 3D volumetric grids entail substantial memory consumption and computational cost.
Therefore, some researchers have aimed to transform point clouds into collections of feature images, and subsequently exert CNNs to extract high-level representations of features and perform point cloud classification. Yang et al., generate 2D feature images by extracting the local geometric features, global geometric features and full-waveform features of each point;the features are then input into 2D CNNs for point cloud classification (Yang et al.,, 2017). Moreover, Yang et al., change the previous image generation approach to a method that can implement the generation at different scales and design a multiscale CNN for final semantic classification based on five features, namely, intensity, eigenvalue features, planarity, sphericity, and variance of deviation angles (Yang et al.,, 2018). Zhao et al., propose a multiscale CNN that can automatically learn deep features of each LiDAR point by generating a set of contextual images from selected features of the LiDAR data, such as height, intensity and roughness (Zhao et al.,, 2018). However, these models cause spatial information loss and induce quantization error in the conversion process (Te et al.,, 2018).
2.3 Deep learning method on point cloud
As a pioneer work of directly applying deep learning models to raw point clouds, PointNet model (Qi et al., 2017a, ) employ MLP to learn the features of individual points and a symmetric function (e.g. max pooling) to encode global information. Although PointNet provide a unified and efficient approach to 3D recognition tasks, it cannot capture the local structure of point cloud. To solve this problem, PointNet++ (Qi et al., 2017b, ) is developed by constructing a hierarchical neural network that applies PointNet recursively on partitioning of point cloud generated via sampling layers and grouping layers. Following these two models, many researchers proposed various deep learning models for point cloud classification based on PointNet-like architectures (Li et al.,, 2018; Jiang et al.,, 2018; Thomas et al.,, 2019).
Considering the inherent topological information of point clouds, researchers propose graph convolution neural network (GCNN) for point cloud classification applied on unordered 3D point clouds (Te et al.,, 2018; Wang et al., 2018a, ). Similarly, Wang et al., 2019b design a dynamic graph convolution operations to capture local geometric structures by generating edge features between a point and its neighborhoods (Wang et al., 2019b, ). More recently, some studies try to introduce the attention mechanism to learn a more adaptive local summary of the neighborhood. For instance, Chen et al., propose the GAPNet model, which employs a multi-head mechanism to aggregate attention features for each point from its neighborhoods and applies stacked MLP layers to capture local geometric features from raw point cloud (Chen et al.,, 2019). Wang et al., 2019a introduce a graph attention convolution (GAC) to selectively focus on the most correlated part of the neighbors and train a graph attention convolution network for point cloud classification (Wang et al., 2019a, ). Note that these graph attention convolution neural networks just pay attention to local geometric structures, but our model also takes density distribution and global contextual relationships into consideration, which is the main difference of our graph attention convolution network from the previous ones in the computer vision field.
Specifically, in the field of airborne LiDAR point cloud classification, few studies apply deep learning models directly to raw point clouds. Yousefhussien et al., present a 1D-fully convolutional classification network that directly consumes 3D coordinates and three corresponding spectral features extracted from 2D georeferenced images for each point (Yousefhussien et al.,, 2017). Furthermore, Wang et al., 2018b design a deep neural network with spatial pooling (DNNSP) by adopting a max pooling layer to aggregate the point-based features into the cluster-based features, which are then input to another MLP for point cloud classification (Wang et al., 2018b, ). In addition, our previous work explores a directionally constrained point convolution (D-Conv) module to extract local features of 3D point sets from the projected 2D receptive fields and then introduces a multiscale fully convolutional neural network based on the above module (Wen et al.,, 2020).
However, these point cloud deep learning methods focus on extracting local features and ignore the global relationships between individual points. In addition, the above models employ standard convolution kernels with a regular receptive fields, which neglect the structural connections between points and fails to account for varying point density.
In this paper, we propose a GACNN for airborne LiDAR point cloud classification that can directly handle unstructured 3D point clouds by considering local structural features and global contextual information simultaneously. Specifically, our proposed graph attention convolution module dynamically learns convolution weights and adaptively adjusts the convolution kernel according to the local structural connection of the point cloud. Moreover, the unbalanced density distribution of the point cloud is taken into account by our module.
3 Methods
In this section, we first present the calculation of our proposed graph local attention (Section 3.1) and global attention (Section 3.2). Then, we describe how to combine these two attention mechanisms into our graph convolution module (Section 3.3.1). Based on the designed graph convolution module, we devise an encoder-decoder framework neural network (Section 3.3.2) that can enable learning multiscale features for airborne LiDAR point cloud classification.


3.1 Graph local attention
To further explore the local structure of point clouds, two attention modules are added to standard graph convolution in this study, i.e. edge attention and density attention, so the proposed graph convolution module can dynamically learn convolutional kernel shapes to adjust to the structure of point sets and take the varying density distribution of non-uniform sampled point clouds into account.
3.1.1 Edge attention
Given a point cloud , a KNN graph is constructed according to its spatial neighbors, where are nodes of points, are the edges connecting pairs of points, and is the neighborhood set of point . Edge features are defined as , where , , and represents the neighboring point of point .
To extract local structural features and learn the most related parts of the neighbors, the edge features are input into a nonlinear transformation function , which is implemented via MLP and can be defined by :
(1) |
where and , respectively, represent weight parameter and bias parameter of the MLP layer.
In order to keep the same scale for the neighbor attention coefficients of different vertices, a softmax function was used to normalize weights across all neighbors to the reference vertex as follows:
(2) |
where indicates the edge attention weight of vertex to reference vertex . Figure 1 illustrates the effect of edge attention on standard graph convolution for a subgraph of a point cloud. After adding an edge attention mechanism, the proposed model can learn to strengthen or weaken convolution weight and dynamically adjust the actual receptive field of the convolution kernel to the local structure of point cloud.
3.1.2 Density attention
To overcome nonuniform density of the raw point cloud across different locations, we add density attention to the graph convolution module. First, the density of each point is estimated by the kernel density estimation (KDE), which can be represented as follows:
(3) |
where K denotes the kernel function, n refers to the number of neighboring points, and h represents kernel window width. In this paper, we employ a Gaussian kernel function:
(4) |
Then the inverse density of neighboring point to reference point is computed and normalized by dividing the value by the maximum:
(5) |
(6) |
Finally, the density attention weight of point to reference point is calculated by MLP, as shown in Eq.7.
(7) |
3.2 Graph global attention
Previous works focus on extracting local features and obtaining global information by employing a symmetric function but ignore the spatial relationships among all points. To remedy this problem, we propose a graph global attention module to learn global contextual information of the point cloud. The Euclidean distance between every two individual points in each coordinate direction is calculated to obtain the distance matrix with dimension , which can be represented as:
(8) |
Similarly, we use a softmax layer to normalize the distance for each point, which can be expressed as:
(9) |
where and signify the distance and normalized distance between points and , respectively. Another MLP layer is employed as nonlinear transform function to obtain the final attention weight from the normalized distance, which is given by:
(10) |
3.3 Graph attention convolution neural network
In the last two sections, we introduce the local and global attention modules in detail. In this section, we first present how the two attention mechanisms are fused into the proposed graph attention convolution module. Then, we develop an end-to-end encoder-decoder network to learn multiscale features from raw point clouds.

3.3.1 Graph attention convolution module
As illustrated in Figure 2, a point cloud and its corresponding features are input into module. Note that these features are optional (represented by black dashed lines in Figure 2), and only the 3D coordinates are taken as input (i.e. ) when these additional features are not required. The KNN graph is built according to coordinates of the point cloud, and local neighboring point sets are obtained.
As discussed in Section 3.2, we first calculate the normalized Euclidean distance, and then learn a nonlinear transformation , which can be implemented via MLP to procure global attention weights. Meanwhile, another MLP is employed to learn a mapping: from the local neighboring point sets of the input point cloud. The global attention feature map is generated via matrix multiplication between global attention weights and the mapping result of the local neighboring point sets. Finally, after concatenation of the local neighboring point sets and , the feature map is obtained. The above process describes how global attention is added to our graph attention convolution module, which can be found in the blue dotted box in Figure 2.
Next, we introduce how local attention is added to the graph attention convolution module, as shown in the red dashed box in Figure 2. Based on the spatial coordinates of the local point sets, we calculate edge features , followed by two shared MLP layers (C1,C2) to output to the local edge attention feature map (see green dotted box in Figure 2). Then, the edge attention weights are acquired by applying the softmax layer to normalize . Thereafter, element-wise product of the edge attention weights and the feature map , which is the output of an MLP applied to , is performed to generate feature map . Regarding density attention (see the yellow dotted box in Figure 3), the density of each input raw point is calculated, and the density of the local neighboring point sets, which are selected through the KNN graph pursuant to spatial position, are aggregated to constitute a density neighboring matrix. Then, two shared MLP layers (C1,1) are used to acquire the density attention weights via element-wise product with . The output of the element-wise product is feature map , which is again input into another MLP layer and max pooling layer to secure the feature map , which is final output of our graph attention convolution module.

3.3.2 Overall architecture
Inspired by SegNet (Badrinarayanan et al.,, 2017) and PointNet++ (Qi et al., 2017b, ), we develop an encoder-decoder GACNN in an end-to-end manner based on the above graph attention convolution module for airborne LiDAR point cloud classification. Raw point clouds with 3D coordinates and optional features are directly input into our encoder network. Subsequently, the sampling layer implemented via the farthest point sampling algorithm, and graph attention convolution module are employed recursively four times to extract multiscale features in the encoder network. To propagate the learned features from the encoded sampled points to the original points, the interpolation is first attained through inverse distance weighting within the decoder network. More details of the sampling and interpolation can be found in the PointNet++ (Qi et al., 2017b, ). Then the interpolated features are concatenated in a skip manner with the point features from the corresponding encoder stages. Next the concatenated features are input into our graph attention convolution module to capture features from the coarse-level information. Note that the graph attention convolution module contains only two local attention mechanisms in each decoder. After the last interpolation, the feature collections encompass the same number of points as that of the original point sets and are fed into a 1 1 convolution to obtain the final semantic label for each point.
4 Experiments
In this section, the performance of our GACNN for airborne LiDAR point cloud classification is evaluated on a real-world dataset. We briefly describe the experimental dataset in Section 4.1 and introduce the evaluation metrics of the point cloud classification in Section 4.2. The implementation details of our GACNN model are illustrated in Section 4.3. In Section 4.4, we present the classification results of the GACNN model on the dataset.
4.1 Dataset
We assess the performance of our model on the International Society for Photogrammetry and Remote Sensing (ISPRS) 3D labeling dataset, which is composed of airborne laser scanning data acquired with a Leica ALS50 system from Vaihingen, Germany (Cramer,, 2010). The LiDAR data are categorized into nine semantic classes (Niemeyer et al.,, 2014), that is power line, low vegetation (low_veg), impervious surface (imp_surf), fence/hedge, car, roof, facade, shrub, and tree. Scenes from three different areas are provided on the ISPRS 3D labeling website: one scene with 753,876 points is employed as the training set and the other two scenes with 411,722 points are used as the test set. Table 1 describes the number of points of each category for each set. The spatial XYZ coordinates, intensity, the return of number, the number of returns and semantic labels are provided for each set. Figure 4 shows the three scenes in the dataset, and the rendering color of each category is referenced to (Blomley et al.,, 2016).
Categories | Training Set | Test Set |
---|---|---|
Powerline | 546 | 600 |
Low vegetation | 180,850 | 98,690 |
Impervious surfaces | 193,723 | 101,986 |
Car | 4,614 | 3,708 |
Fence/Hedge | 12,070 | 7,422 |
Roof | 152,045 | 109,048 |
Facade | 27,250 | 11,224 |
Shrub | 47,605 | 24,818 |
Tree | 135,173 | 54,226 |
Total | 753,876 | 411,722 |


Categories | power | low_veg | imp_surf | car | fence_hedge | roof | facade | shrub | tree |
---|---|---|---|---|---|---|---|---|---|
power | 0.775 | 0.000 | 0.000 | 0.000 | 0.000 | 0.152 | 0.002 | 0.005 | 0.067 |
low_veg | 0.000 | 0.780 | 0.082 | 0.001 | 0.006 | 0.008 | 0.004 | 0.102 | 0.017 |
imp_surf | 0.000 | 0.054 | 0.942 | 0.000 | 0.000 | 0.000 | 0.001 | 0.003 | 0.000 |
car | 0.000 | 0.040 | 0.012 | 0.709 | 0.039 | 0.020 | 0.010 | 0.152 | 0.019 |
fence_hedge | 0.000 | 0.091 | 0.015 | 0.009 | 0.290 | 0.026 | 0.026 | 0.440 | 0.103 |
roof | 0.001 | 0.019 | 0.000 | 0.000 | 0.001 | 0.912 | 0.013 | 0.013 | 0.041 |
fac | 0.002 | 0.056 | 0.005 | 0.003 | 0.005 | 0.151 | 0.540 | 0.134 | 0.104 |
shrub | 0.000 | 0.107 | 0.006 | 0.005 | 0.027 | 0.045 | 0.025 | 0.611 | 0.174 |
tree | 0.000 | 0.015 | 0.000 | 0.001 | 0.004 | 0.022 | 0.011 | 0.145 | 0.802 |
Precision/Correctness | 0.746 | 0.860 | 0.919 | 0.860 | 0.544 | 0.951 | 0.648 | 0.378 | 0.776 |
Recall/Completeness | 0.775 | 0.780 | 0.942 | 0.709 | 0.290 | 0.912 | 0.540 | 0.611 | 0.802 |
F1 score | 0.760 | 0.818 | 0.930 | 0.777 | 0.378 | 0.931 | 0.589 | 0.467 | 0.789 |
4.2 Evaluation metric
According to the standard evaluation metrics of the ISPRS 3D labeling contest, precision, recall, and F1 score, overall accuracy (OA) are used to evaluate the performance of point cloud classification. In general, OA, which is specified as the percentages of correctly classified points in the test set, examines the classification accuracy for all categories. F1 score measures the classification performance for each category based on the precision and recall of the classification model and is a better choice than OA when there is a large difference in the number of points in each category. The OA, precision, recall and F1 score are designated as follows:
(11) |
(12) |
(13) |
(14) |
where TP, TN, FP and FN, respectively, denotes true positive (the fraction of positives correctly classified), true negative (the fraction of negatives correctly classified), false positive (the fraction of positives misclassified), and false negative (the fraction of negatives misclassified).
4.3 Implementation Details
As stated in section 3.3, the raw 3D coordinates and optional features (intensity values acquired from the airborne LiDAR data and the height above ground features are adopted in this paper) of the point cloud are input into our model; however, the entire training point set cannot be directly fed into the network because of the limited GPU memory. To resolve this problem, we separated the whole scene into a set of 30 m*30 m*40 m cuboid regions. During the model training step, to make the model more robust and prevent overfitting, we first randomly select several cuboid regions, randomly choose 8,192 points from each cuboid region and indiscriminately drop 12.5% of these points. In terms of the model testing stage, all points from split cuboid regions in the test set (see Figure. 4 b)) are fed into the trained model to achieve point-to-point classification, although the number of points is different for each cuboid region. Note that the split small broken regions that occur at the edges of the scene are merged into the surrounding larger regions to ensure the integrity of each cuboid region.
The proposed model is implemented based on the Tensorflow framework. We employ the ADAM optimizer with a learning rate of 0.01 and divide the learning rate by 2 every 3,000 steps. The batch size is set to 8 for model training and to 1 for model testing. In the encoder network, the number of sampling points in each sampling layer is set to 1024, 512, 64, and 16, with the feature dimensions C1, C2, and C3 (see Figure 2) in each proposed graph attention convolution module set to (32, 32, 64), (64, 64, 128), (128, 128, 256) and (256, 256, 512), and the number of neighboring points K set to 32. In the decoder network, the feature dimensions C2 and C3 (without global attention in the decoder) in each proposed graph attention convolution module are (512, 512), (256, 256), (256, 128) and (128, 128), individually, and the number of neighboring points K is 16. Note that the model parameters are determined through a series of comparative experiments; we do not elaborate on the experimental details since that is not the focus of our paper.

4.4 Classification results
As covered in the last section, the model training is conducted on the ISPRS dataset for 1000 epochs, taking 10 hours on a Titan Xp GPU until convergence. Subsequently the point cloud of each cuboid region of the test set (Scene 2 and 3) is directly input into the trained model in turn to predict the semantic labels of each point. Figure 5 and Figure 6 show the classification results and error maps of our GACNN model which correctly labelled the majority of the points in the test set.
To quantitatively assess the classification performance of our model for the points in each category, the classification confusion matrix, precision, recall and F1 score of each category are all calculated and listed in Table 2. Table 2 shows that the proposed model achieves satisfactory classification (F1 score greater than 75%) on six categories, namely, power line, low vegetation, impervious surfaces, car, roof, and tree. Acceptable classification performance is achieved on facade categories due to the mingling of facade category with the roof, shrub and tree categories, which causes the misclassification of facade points.
However, our model demonstrates poor classification performance for the fence/hedge and shrub categories, most likely because of the similar geographical distribution and topological features of these two categories. In addition, Table 2 shows that some points of the shrub category were misclassified as tree, which may also result from the mixing and lack of obvious boundaries between the shrub and tree categories. As shown in Figure 7, points of the shrub category are mixed with points of the fence/hedge category (marked with the left circle and right circle) and points of the tree category (marked with the right circle).
5 Discussion
In this section, we examine the effect of the proposed global and local attention modules through a set of ablation studies in section 5.1. In Section 5.2, we compare the proposed GACNN with other state-of-the-art methods for airborne LiDAR point cloud classification on the ISPRS dataset. Moreover, we discuss the generalization capability of GACNN on 2019 Data Fusion Contest Dataset.
5.1 Ablation study for attention modules


To evaluate the effectiveness of the proposed graph attention convolution module (i.e. the global attention module and local attention module), we design an ablation experiment to compare six models: a) GACNN model without the global and the local attention modules (w/o global, w/o local), b) GACNN model with the global attention module but without the local attention modules (w global, w/o local), c) GACNN model without the global attention module but with the local attention modules (w/o global, w/ local), d) GACNN model with the global and the local edge attention module (w/ global, w/ local_edge), e) GACNN model with global and local density attention module (w/ global, w/ local_density), and f) GACNN model with the global and the two local attention modules (w/ global, w/ local).
Table 3 shows the classification results of the above six models. Take local attention module as an example, it can be found that adding local attention module helps to improve the classification performance by comparing a) GACNN with c) GACNN model, and b) GACNN with f) GACNN model. To be more specific, the comparison of classification results between b) GACNN and d) GACNN model, as well as e) GACNN and f) GACNN model demonstrate the effectiveness of the first local module, namely edge attention module. Similarly, the value of the second local module, density attention module, can be illustrated by comparing b) GACNN with e) GACNN model, and d) GACNN with f) GACNN. We can implement alike analysis to verify the effectiveness of global attention module. Trough these comparisons, it can be found that each attention module helps to improve the classification performance to some extent.
To more intuitively demonstrate the effects of each attention module, we randomly select a region and plot the classification results of the six models, as shown in Figure 8. The three models without the local edge attention module (a, b and e) misclassify the roof category points as tree category points, which validates our assumption that the addition of the local edge attention module promotes the classification performance by dynamically learning the convolution kernel according to the local structure of the point cloud.
Furthermore, Figure 9 visualizes the feature maps of the global and local attention modules in the first encoder layer of the GACNN model on a randomly selected region. The input point cloud on this region is first dowsampled into 1024 points and then fed into our graph attention convolution module. Three attention feature maps of the sampled points are procured by implementing the procedures illustrated in Figure 2. The black circles and arrows in Figure 9 show the partial enlargement of the local neighboring points of the selected point. The learned local features of the two local attention modules correspond to the label of the neighboring points in general. Moreover, the global attention feature map shows that the features of the neighboring points differ when the distance between the selected point and the surrounding points varies. Finally, the output feature maps (two channels) of the graph attention convolution module are shown on the right-most side of Figure 9. The features of impervious surfaces and roofs are substantially captured by our graph attention convolution module, which further authenticates the effect of the global and local attention modules.
Model | OA | Average F1 |
---|---|---|
a) GACNN(w/o global, w/o local) | 0.816 | 0.697 |
b) GACNN(w/ global, w/o local) | 0.828 | 0.707 |
c) GACNN(w/o global, w/ local) | 0.824 | 0.705 |
d) GACNN(w/ global, w/ local_edge) | 0.831 | 0.709 |
e) GACNN(w/ global, w/ local_density) | 0.830 | 0.710 |
f) GACNN(w/ global, w/ local) | 0.832 | 0.715 |
5.2 Comparisons with other methods

Categories | power | low_veg | imp_surf | car | fence_hedge | roof | facade | shrub | tree | OA | Average F1 |
---|---|---|---|---|---|---|---|---|---|---|---|
UM (Horvat et al.,, 2016) | 0.461 | 0.790 | 0.891 | 0.477 | 0.052 | 0.920 | 0.527 | 0.409 | 0.779 | 0.808 | 0.590 |
WhuY2 | 0.319 | 0.800 | 0.889 | 0.408 | 0.245 | 0.931 | 0.494 | 0.411 | 0.773 | 0.810 | 0.586 |
WhuY3 (Yang et al.,, 2017) | 0.371 | 0.814 | 0.901 | 0.634 | 0.239 | 0.934 | 0.475 | 0.399 | 0.780 | 0.823 | 0.616 |
LUH (Niemeyer et al.,, 2016) | 0.596 | 0.775 | 0.911 | 0.731 | 0.340 | 0.942 | 0.563 | 0.466 | 0.831 | 0.816 | 0.684 |
BIJ_W (Wang et al., 2018b, ) | 0.138 | 0.785 | 0.905 | 0.564 | 0.363 | 0.922 | 0.532 | 0.433 | 0.784 | 0.815 | 0.603 |
RIT_1 (Yousefhussien et al.,, 2018) | 0.375 | 0.779 | 0.915 | 0.734 | 0.180 | 0.940 | 0.493 | 0.459 | 0.825 | 0.816 | 0.633 |
NANJ2 (Zhao et al.,, 2018) | 0.620 | 0.888 | 0.912 | 0.667 | 0.407 | 0.936 | 0.426 | 0.559 | 0.826 | 0.852 | 0.693 |
WhuY4 (Yang et al.,, 2018) | 0.425 | 0.827 | 0.914 | 0.747 | 0.537 | 0.943 | 0.531 | 0.479 | 0.828 | 0.849 | 0.692 |
D-FCN (Wen et al.,, 2020) | 0.704 | 0.802 | 0.914 | 0.781 | 0.370 | 0.930 | 0.605 | 0.460 | 0.794 | 0.822 | 0.707 |
GACNN | 0.760 | 0.818 | 0.930 | 0.777 | 0.378 | 0.931 | 0.589 | 0.467 | 0.789 | 0.832 | 0.715 |
After illustrating the effectiveness of each attention module, we compare our model with other models submitted to the ISPRS 3D Labeling benchmark to exhibit the advantages of the proposed model. The top eight models with the highest performance on the benchmark, including UM (Horvat et al.,, 2016), WhuY2, WhuY3 (Yang et al.,, 2017), LUH (Niemeyer et al.,, 2016), BIJ_W (Wang et al., 2018b, ), RIT_1 (Yousefhussien et al.,, 2018), NANJ2 (Zhao et al.,, 2018) and WhuY4 (Yang et al.,, 2018)), are selected for performance comparisons. In addition, we also compare the proposed model with our previous D-FCN model (Wen et al.,, 2020), which has been shown to achieve better classification performance than that of the mentioned eight models. Table 4 displays OA and F1 score of our model and the comparison models on the ISPRS benchmark dataset. Our GACNN model achieves better classification performance in terms of average F1 score than the other models. Specifically, the proposed GACNN model achieves state-of-the-art classification performance for the power line, impervious surfaces categories.
In addition, Table 4 shows that the previous models, such as the NANJ2 model and D-FCN model,unilaterally achieves good results on a certain evaluation metric (overall accuracy or average F1 score), but performs poorly on the other indicator. By contrast, our model not only achieves the best performance on average F1 score, but also has a satisfying result on overall accuracy in comparison to other models. Meanwhile, it exhibits difficulty to simultaneously achieve the best overall accuracy and average F1 score on this benchmark dataset, since the number of points in various categories is extremely uneven (see Table 1). The categories with a large number of points receive more attention during model training, while the small categories are ignored, which leads to good OA performance but poor average F1 scores. On the other hand, performance enhancement on average F1 score will require the model to learn representative features on a small amount of data, which easily results in model overfitting and constrains on overall accuracy. Overall, our GACNN model achieves a satisfactory balance between OA and average F1 score.
Similarly, to further intuitively illustrate the classification results of different models, we plot the results of four models in Figure 10, including RIT_1, NANJ2, WhuY4, and D-FCN, which acquire the four best classification performance in terms of average F1 score . As shown in the Figure 10, the three models (a,b,and c) in the first row misclassify the points of facade category into the shrub and roof categories, which validates our assumption that facade category is difficult to distinguish from other categories (see Section 4.4). The main reason these models fail is that they cannot learn effective local neighboring features of the point cloud. Specifically, the RIT_1 model adopts a 1D fully convolutional network that cannot capture multiscale features, and NANJ2 and WhuY4 model lose 3D spatial information during the transformation of point cloud to 2D feature images. However, D-FCN and GACNN model learn effective local structural features by employing their core modules, thus obtaining better performance than that of the other models. Moreover, although most points in the low vegetation category are misclassified as impervious surfaces by all the models, our GACNN model still achieves an acceptable result.
Moreover, to fairly justify the superiority of our model, we compare our model with other prevalent point cloud deep learning models in the field of computer vision, including PointNet++ (Qi et al., 2017b, ), PointSIFT (Jiang et al.,, 2018), PointCNN (Li et al.,, 2018), PointCNN with A-XCRF model (Arief et al.,, 2019), KPConv(Thomas et al.,, 2019), DGCNN(Wang et al., 2019b, ) and GACNet(Wang et al., 2019a, ) (see Table 5). It can be seen from Table 5 that the proposed GACNN model have achieved better performance than these advanced point cloud deep learning models, which again demonstrate the superiority of our model. Please note that PointCNN with A-XCRF model is a post-processing method to address the overfitting of PointCNN, and our model has not adopted any post-processing technology. Furthermore, compared to DGCNN and GACNet, which are closely related to our work and recently proposed in the field of computer vision, our model achieves better accuracy and F1 score by a large margin. It demonstrates that adding attention mechanisms does improve model performance and the two attention mechanisms employed in this paper are more effective, obtaining superior classification performance than GACNet with only a local spatial attention mechanism.
Categories | power | low_veg | imp_surf | car | fence_hedge | roof | facade | shrub | tree | OA | Average F1 |
---|---|---|---|---|---|---|---|---|---|---|---|
PointNet++ (Qi et al., 2017b, ) | 0.579 | 0.796 | 0.906 | 0.661 | 0.315 | 0.916 | 0.543 | 0.416 | 0.770 | 0.812 | 0.656 |
PointSIFT (Jiang et al.,, 2018) | 0.557 | 0.807 | 0.909 | 0.778 | 0.305 | 0.925 | 0.569 | 0.444 | 0.796 | 0.822 | 0.677 |
PointCNN (Li et al.,, 2018) | 0.615 | 0.827 | 0.918 | 0.758 | 0.359 | 0.927 | 0.578 | 0.491 | 0.781 | 0.833 | 0.695 |
PointCNN+A-XCRF (Arief et al.,, 2019) | 0.630 | 0.826 | 0.919 | 0.749 | 0.399 | 0.945 | 0.593 | 0.508 | 0.827 | 0.850 | 0.711 |
KPConv (Thomas et al.,, 2019) | 0.631 | 0.823 | 0.914 | 0.725 | 0.252 | 0.944 | 0.603 | 0.449 | 0.812 | 0.837 | 0.684 |
DGCNN (Wang et al., 2019b, ) | 0.676 | 0.804 | 0.906 | 0.545 | 0.268 | 0.898 | 0.488 | 0.415 | 0.773 | 0.810 | 0.641 |
GACNet (Wang et al., 2019a, ) | 0.628 | 0.819 | 0.908 | 0.698 | 0.252 | 0.914 | 0.562 | 0.395 | 0.763 | 0.817 | 0.660 |
GACNN | 0.760 | 0.818 | 0.930 | 0.777 | 0.378 | 0.931 | 0.589 | 0.467 | 0.789 | 0.832 | 0.715 |
Furthermore, to analyze the computational cost of our model, we compare the per sample inference time of the proposed model with recent state-of-the-art models. As shown in Table 6, it can be noticed that the inference time of our model is close to the comparing models. Our model achieves new state-of-the-art classification performance with a promising computation efficiency.
Model | Inference Time |
---|---|
PointNet++(Qi et al., 2017b, ) | 0.140s |
PointSIFT(Jiang et al.,, 2018) | 0.389s |
PointCNN(Li et al.,, 2018) | 0.220s |
DGCNN(Wang et al., 2019b, ) | 0.379s |
GACNet(Wang et al., 2019a, ) | 0.952s |
GACNN | 0.379s |
5.3 The generalization capability of the proposed model

Categories | Training set | Test set | Total |
---|---|---|---|
Ground | 48,069,788 | 6,148,920 | 54,218,708 |
High Vegetation | 10,775,292 | 1,321,488 | 12,096,780 |
Building | 10,269,771 | 782,455 | 11,052,226 |
Water | 1,356,899 | 19,266 | 1,376,165 |
Bridge Deck | 934,545 | 25,034 | 959,579 |
Total | 71,406,295 | 8,297,163 | 79,703,458 |
Model | OA | Average F1 |
---|---|---|
a) GACNN(w/o global, w/o local) | 0.937 | 0.798 |
b) GACNN(w/ global, w/o local) | 0.945 | 0.820 |
c) GACNN(w/o global, w/ local) | 0.941 | 0.817 |
d) GACNN(w/ global, w/ local_edge) | 0.947 | 0.826 |
e) GACNN(w/ global, w/ local_density) | 0.948 | 0.823 |
f) GACNN(w/ global, w/ local) | 0.951 | 0.828 |
Categories | Ground | High Vegetation | Building | Water | Bridge Deck | OA | Average F1 |
---|---|---|---|---|---|---|---|
PointNet++ (Qi et al., 2017b, ) | 0.983 | 0.958 | 0.797 | 0.044 | 0.073 | 0.927 | 0.571 |
PointSIFT (Jiang et al.,, 2018) | 0.986 | 0.970 | 0.855 | 0.464 | 0.604 | 0.940 | 0.776 |
PointCNN (Li et al.,, 2018) | 0.987 | 0.972 | 0.849 | 0.441 | 0.653 | 0.938 | 0.780 |
KPConv (Thomas et al.,, 2019) | 0.984 | 0.942 | 0.874 | 0.430 | 0.775 | 0.945 | 0.801 |
DGCNN (Wang et al., 2019b, ) | 0.982 | 0.953 | 0.745 | 0.112 | 0.284 | 0.929 | 0.615 |
GACNet (Wang et al., 2019a, ) | 0.985 | 0.968 | 0.852 | 0.403 | 0.687 | 0.937 | 0.779 |
D-FCN (Wen et al.,, 2020) | 0.991 | 0.981 | 0.899 | 0.450 | 0.730 | 0.956 | 0.810 |
GACNN | 0.993 | 0.968 | 0.911 | 0.425 | 0.844 | 0.951 | 0.828 |
To demonstrate the generalization capability of the proposed model, we further conduct experiments on 2019 Data Fusion Contest Dataset. The 2019 Data Fusion Contest, organized by the Image Analysis and Data Fusion Technical Committee (IADF TC) of the IEEE Geoscience and Remote Sensing Society (GRSS), the Johns Hopkins University (JHU), and the Intelligence Advanced Research Projects Activity (IARPA), provides a large-scale Urban Semantic 3D (US3D) dataset covering approximately 100 square kilometers over Jacksonville, Florida and Omaha, Nebraska, United States. The point clouds in US3D dataset are labeled as five categories, including ground, trees, buildings, water and bridge, and are stored in ASCII text files with format x, y, z, intensity, return number.
This training set of Contest Dataset consists of 110 regions, from which the 100 regions are selected for the training dataset and the other 10 regions are used for testing dataset in this experiment. The number of points in this dataset and detailed category distribution is shown in Table 7. Similarly, we randomly select a 128m*128m*210m cuboid region from each region during training and arbitrarily choose 8192 points from the cuboid as model input. As for the testing stage, each region of the testing dataset is split into blocks of 128m*128m grids in the horizontal direction and all points is segmented blocks are input into the model for prediction. Note that the architecture and parameters of our model remain unchanged.
We first evaluate the effectiveness of the proposed graph attention convolution module and conduct an ablation experiment similar to what we have done in Section 5.1. The classification results of the six models which comprise of different attention modules are listed in Table 8, from which it can be found that the proposed local and global attention module do help to improve the classification performance to some extent as we discussed in Section 5.1.
The PointNet++ (Qi et al., 2017b, ), PointSIFT (Jiang et al.,, 2018), PointCNN (Li et al.,, 2018), KPConv (Thomas et al.,, 2019), DGCNN (Wang et al., 2019b, ), GACNet (Wang et al., 2019a, ) and DFCN model (Wen et al.,, 2020) are selected as comparison model, and the classification results on test regions are shown in Table 9. From Table 9, it can be seen that our model obtains the best classification performance on average F1 score and achieves the highest F1 score for three of the five categories, including ground, building and bridge deck. Moreover, Figure 11 displays the classification results of the four latest models among above prevalent point cloud deep learning models on a randomly selected test region, from which we can find that KPConv model and our model classify the points of bridge deck category more correctly. But there are some points of building category are misclassified by KPConv model. At the same time, from Table 9 and Figure 11, it can be found that our model, which incorporates local structural features and global contextual information simultaneously, performs better than DGCNN model and GACNet model, two closely related works in the field of computer vision, which further indicates the effectiveness of the proposed attention modules. As a result, the proposed model still outperforms other state-of-the-art models on the 2019 Data Fusion Contest Dataset and demonstrates a favorable generalization capability.
6 Conclusions
In this paper, a graph attention convolution neural network (GACNN) that is directly applied to unstructured 3D point clouds is proposed to conduct airborne LiDAR point cloud classification. The GACNN model is an end-to-end encoder-decoder network developed based on the proposed graph attention convolution module, which consists of a global attention module and a local attention module. These two attention modules take local structural features and global contextual information into consideration respectively, and their effectiveness is validated through a set of comparative experiments. Specifically, the proposed graph attention convolution module is capable of dynamically learning convolution weights according to the local structure of the point cloud, considering the unbalanced density distribution of the point cloud, and at the same time paying attention to spatial relationships among all points in global. Moreover, we compare our GACNN model with other state-the-of-art models on both ISPRS 3D Labeling Dataset and 2019 Data Fusion Contest Dataset, the results of which demonstrate the proposed model is superior to most of prevalent point cloud classification models, whether in the field of photogrammetry and remote sensing or computer vision, and achieves a new state-of-the-art classification performance in terms of average F1 score. Moreover, experiments on two different datasets show the favorable generalization capability of our model.
ACKNOWLEDGEMENTS (Optional)
This work was supported by the Beijing Municipal Science and technology Project (grant No. Z191100001419002). Additionally, we would like to gratefully acknowledge the ISPRS for providing airborne LiDAR data.
References
- Arief et al., (2019) Arief, H. A., Indahl, U. G., Strand, G.-H. and Tveite, H., 2019. Addressing overfitting on pointcloud classification using atrous xcrf. arXiv preprint arXiv:1902.03088.
- Axelsson, (2000) Axelsson, P., 2000. Dem generation from laser scanner data using adaptive tin models. International archives of photogrammetry and remote sensing 33(4), pp. 110–117.
- Badrinarayanan et al., (2017) Badrinarayanan, V., Kendall, A. and Cipolla, R., 2017. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine intelligence 39(12), pp. 2481–2495.
- Blomley et al., (2016) Blomley, R., Jutzi, B. and Weinmann, M., 2016. 3d semantic labeling of als point clouds by exploiting multi-scale, multi-type neighborhoods for feature extraction.
- Bradbury et al., (2005) Bradbury, R. B., Hill, R. A., Mason, D. C., Hinsley, S. A., Wilson, J. D., Balzter, H., Anderson, G. Q., Whittingham, M. J., Davenport, I. J. and Bellamy, P. E., 2005. Modelling relationships between birds and vegetation structure using airborne lidar data: a review with case studies from agricultural and woodland environments. Ibis 147(3), pp. 443–452.
- Chan et al., (2015) Chan, T.-H., Jia, K., Gao, S., Lu, J., Zeng, Z. and Ma, Y., 2015. Pcanet: A simple deep learning baseline for image classification? IEEE transactions on image processing 24(12), pp. 5017–5032.
- Chehata et al., (2009) Chehata, N., Guo, L. and Mallet, C., 2009. Airborne lidar feature selection for urban classification using random forests. International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences 38(Part 3), pp. W8.
- Chen et al., (2019) Chen, C., Fragonara, L. Z. and Tsourdos, A., 2019. Gapnet: Graph attention based point neural network for exploiting local feature of point cloud. arXiv preprint arXiv:1905.08705.
- Chen, (2007) Chen, Q., 2007. Airborne lidar data processing and information extraction. Photogrammetric engineering and remote sensing 73(2), pp. 109.
- Collobert and Weston, (2008) Collobert, R. and Weston, J., 2008. A unified architecture for natural language processing: Deep neural networks with multitask learning. In: Proceedings of the 25th international conference on Machine learning, ACM, pp. 160–167.
- Cramer, (2010) Cramer, M., 2010. The dgpf-test on digital airborne camera evaluation–overview and test design. Photogrammetrie-Fernerkundung-Geoinformation 2010(2), pp. 73–82.
- Hinton et al., (2012) Hinton, G., Deng, L., Yu, D., Dahl, G., Mohamed, A.-r., Jaitly, N., Senior, A., Vanhoucke, V., Nguyen, P., Kingsbury, B. et al., 2012. Deep neural networks for acoustic modeling in speech recognition. IEEE Signal processing magazine.
- Horvat et al., (2016) Horvat, D., Žalik, B. and Mongus, D., 2016. Context-dependent detection of non-linearly distributed points for vegetation classification in airborne lidar. ISPRS Journal of Photogrammetry and Remote Sensing 116, pp. 1–14.
- Huang et al., (2014) Huang, C., Peng, Y., Lang, M., Yeo, I.-Y. and McCarty, G., 2014. Wetland inundation mapping and change monitoring using landsat and airborne lidar data. Remote Sensing of Environment 141, pp. 231–242.
- Jiang et al., (2018) Jiang, M., Wu, Y., Zhao, T., Zhao, Z. and Lu, C., 2018. Pointsift: A sift-like network module for 3d point cloud semantic segmentation. arXiv preprint arXiv:1807.00652.
- Krabill et al., (1984) Krabill, W., Collins, J., Link, L., Swift, R. and Butler, M., 1984. Airborne laser topographic mapping results. Photogrammetric Engineering and Remote Sensing 50(6), pp. 685–694.
- LeCun et al., (2015) LeCun, Y., Bengio, Y. and Hinton, G., 2015. Deep learning. nature 521(7553), pp. 436.
- Li et al., (2018) Li, Y., Bu, R., Sun, M., Wu, W., Di, X. and Chen, B., 2018. Pointcnn: Convolution on x-transformed points. In: Advances in Neural Information Processing Systems, pp. 820–830.
- Liu, (2008) Liu, X., 2008. Airborne lidar for dem generation: some critical issues. Progress in Physical Geography 32(1), pp. 31–49.
- Lodha et al., (2007) Lodha, S. K., Fitzpatrick, D. M. and Helmbold, D. P., 2007. Aerial lidar data classification using adaboost. In: Sixth International Conference on 3-D Digital Imaging and Modeling (3DIM 2007), IEEE, pp. 435–442.
- Maturana and Scherer, (2015) Maturana, D. and Scherer, S., 2015. Voxnet: A 3d convolutional neural network for real-time object recognition. In: 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, pp. 922–928.
- Munoz et al., (2009) Munoz, D., Bagnell, J. A., Vandapel, N. and Hebert, M., 2009. Contextual classification with functional max-margin markov networks. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, pp. 975–982.
- Niemeyer et al., (2012) Niemeyer, J., Rottensteiner, F. and Soergel, U., 2012. Conditional random fields for lidar point cloud classification in complex urban areas. ISPRS annals of the photogrammetry, remote sensing and spatial information sciences 3, pp. 263–268.
- Niemeyer et al., (2014) Niemeyer, J., Rottensteiner, F. and Soergel, U., 2014. Contextual classification of lidar data and building object detection in urban areas. ISPRS journal of photogrammetry and remote sensing 87, pp. 152–165.
- Niemeyer et al., (2016) Niemeyer, J., Rottensteiner, F., Sörgel, U. and Heipke, C., 2016. Hierarchical higher order crf for the classification of airborne lidar point clouds in urban areas. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences-ISPRS Archives 41 (2016) 41, pp. 655–662.
- Niemeyer et al., (2011) Niemeyer, J., Wegner, J. D., Mallet, C., Rottensteiner, F. and Soergel, U., 2011. Conditional random fields for urban scene classification with full waveform lidar data. In: ISPRS Conference on Photogrammetric Image Analysis, Springer, pp. 233–244.
- (27) Qi, C. R., Su, H., Mo, K. and Guibas, L. J., 2017a. Pointnet: Deep learning on point sets for 3d classification and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 652–660.
- (28) Qi, C. R., Yi, L., Su, H. and Guibas, L. J., 2017b. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In: Advances in Neural Information Processing Systems, pp. 5099–5108.
- Shapovalov et al., (2010) Shapovalov, R., Velizhev, E. and Barinova, O., 2010. Nonassociative markov networks for 3d point cloud classification. the. In: International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XXXVIII, Part 3A, Citeseer.
- Sohn et al., (2012) Sohn, G., Jwa, Y. and Kim, H. B., 2012. Automatic powerline scene classification and reconstruction using airborne lidar data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci 13(167172), pp. 28.
- Te et al., (2018) Te, G., Hu, W., Zheng, A. and Guo, Z., 2018. Rgcnn: Regularized graph cnn for point cloud segmentation. In: 2018 ACM Multimedia Conference on Multimedia Conference, ACM, pp. 746–754.
- Thomas et al., (2019) Thomas, H., Qi, C. R., Deschaud, J.-E., Marcotegui, B., Goulette, F. and Guibas, L. J., 2019. Kpconv: Flexible and deformable convolution for point clouds. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 6411–6420.
- (33) Wang, C., Samari, B. and Siddiqi, K., 2018a. Local spectral graph convolution for point set feature learning. In: Proceedings of the European conference on computer vision (ECCV), pp. 52–66.
- (34) Wang, L., Huang, Y., Hou, Y., Zhang, S. and Shan, J., 2019a. Graph attention convolution for point cloud semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 10296–10305.
- (35) Wang, Y., Sun, Y., Liu, Z., Sarma, S. E., Bronstein, M. M. and Solomon, J. M., 2019b. Dynamic graph cnn for learning on point clouds. ACM Transactions on Graphics (TOG) 38(5), pp. 1–12.
- (36) Wang, Z., Zhang, L., Zhang, L., Li, R., Zheng, Y. and Zhu, Z., 2018b. A deep neural network with spatial pooling (dnnsp) for 3-d point cloud classification. IEEE Transactions on Geoscience and Remote Sensing 56(8), pp. 4594–4604.
- Wen et al., (2019) Wen, C., Liu, S., Yao, X., Peng, L., Li, X., Hu, Y. and Chi, T., 2019. A novel spatiotemporal convolutional long short-term neural network for air pollution prediction. Science of The Total Environment 654, pp. 1091–1099.
- Wen et al., (2020) Wen, C., Yang, L., Li, X., Peng, L. and Chi, T., 2020. Directionally constrained fully convolutional neural network for airborne lidar point cloud classification. ISPRS Journal of Photogrammetry and Remote Sensing 162, pp. 50–62.
- Yang et al., (2017) Yang, Z., Jiang, W., Xu, B., Zhu, Q., Jiang, S. and Huang, W., 2017. A convolutional neural network-based 3d semantic labeling method for als point clouds. Remote Sensing 9(9), pp. 936.
- Yang et al., (2018) Yang, Z., Tan, B., Pei, H. and Jiang, W., 2018. Segmentation and multi-scale convolutional neural network-based classification of airborne laser scanner data. Sensors 18(10), pp. 3347.
- Yousefhussien et al., (2017) Yousefhussien, M., Kelbe, D. J., Ientilucci, E. J. and Salvaggio, C., 2017. A fully convolutional network for semantic labeling of 3d point clouds. arXiv preprint arXiv:1710.01408.
- Yousefhussien et al., (2018) Yousefhussien, M., Kelbe, D. J., Ientilucci, E. J. and Salvaggio, C., 2018. A multi-scale fully convolutional network for semantic labeling of 3d point clouds. ISPRS journal of photogrammetry and remote sensing 143, pp. 191–204.
- Yu et al., (2010) Yu, B., Liu, H., Wu, J., Hu, Y. and Zhang, L., 2010. Automated derivation of urban building density information using airborne lidar data and object-based method. Landscape and Urban Planning 98(3-4), pp. 210–219.
- Zhang et al., (2013) Zhang, J., Lin, X. and Ning, X., 2013. Svm-based classification of segmented airborne lidar point clouds in urban areas. Remote Sensing 5(8), pp. 3749–3775.
- Zhao et al., (2018) Zhao, R., Pang, M. and Wang, J., 2018. Classifying airborne lidar point clouds via deep features learned by a multi-scale convolutional neural network. International Journal of Geographical Information Science 32(5), pp. 960–979.
- Zhu and Hyyppä, (2014) Zhu, L. and Hyyppä, J., 2014. Fully-automated power line extraction from airborne laser scanning point clouds in forest areas. Remote Sensing 6(11), pp. 11267–11282.