Figures
Abstract
Compared with other point clouds, the airborne LiDAR point cloud has its own characteristics. The deep learning network PointNet++ ignores the inherent properties of airborne LiDAR point, and the classification precision is low. Therefore, we propose a framework based on the PointNet++ network. In this work, we proposed an interpolation method that uses adaptive elevation weight to make full use of the objects in the airborne LiDAR point, which exhibits discrepancies in elevation distributions. The class-balanced loss function is used for the uneven density distribution of point cloud data. Moreover, the relationship between a point and its neighbours is captured, densely connecting point pairs in multiscale regions and adding centroid features to learn contextual information. Experiments are conducted on the Vaihingen 3D semantic labelling benchmark dataset and GML(B) benchmark dataset. The experiments show that the proposed method, which has additional contextual information and makes full use of the airborne LiDAR point cloud properties to support classification, achieves high accuracy and can be widely used in airborne LiDAR point classification.
Citation: Nong X, Bai W, Liu G (2023) Airborne LiDAR point cloud classification using PointNet++ network with full neighborhood features. PLoS ONE 18(2): e0280346. https://doi.org/10.1371/journal.pone.0280346
Editor: Wajid Mumtaz, National University of Sciences and Technology, PAKISTAN
Received: August 7, 2022; Accepted: December 23, 2022; Published: February 10, 2023
Copyright: © 2023 Nong et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The Vaihingen data are available from the Working Group (WG) III/4 of ISPRS from Vaihingen area of Germany, but the data need to be requested. The link is https://www.isprs.org/education/benchmarks/UrbanSemLab/default.aspx. The GML (B) dataset URLs are are available from https://github.com/bwf124565/data.
Funding: This research was funded by the National Natural Science Foundation of China (Grant No.:41904170).
Competing interests: The authors have declared that no competing interests exist.
Introduction
The airborne light detection and ranging (LiDAR) system provides a new technical approach for acquiring 3D spatial data, which provide reliable depth information. Airborne Laser Scanning (ALS) is one of the most import techniques for data collection for real-world scenes. The airborne laser scanning point cloud has been utilised in various fields, such as surveying and mapping, forestry survey [1], Underwater Navigation [2], 3D object detecting [3, 4], remote sensing [5], computer vision [6], disaster monitoring and cultural heritage protection [7].
The airborne LiDAR point has the following characteristics: (1) The categories in the airborne LiDAR point scenes have evident geometric properties. (2) The objects have extreme scale variations. (3) The objects have discrepancy distribution along the elevation [8]. Capturing high-resolution or fine-grained features for ALS point cloud classification is difficult because the geometric attributes are similar [9]. The early works for point cloud data processing focused on designing various handcrafted point descriptors derived from the surrounding neighbourhood [10, 11]. These point descriptors include density, roughness and curvature [12, 13]. The machine learning method is commonly used to achieve point cloud classification based on various handcrafted descriptors and various methods, e.g. support vector machine [14], random forest and AdaBoost [15]. Beside, Some probabilistic graphical models are used to take advantage of the contextual information [16, 17]. However, these methods heavily rely on handcrafted features and have limited generalisability for large-scale wild scenes. Nevertheless, the performance of the traditional method still heavily relies on the representation ability of the handcrafted features.
The deep learning method has recently achieved remarkable performance in scene classification, object detection and change detection [18]. Many researchers used deep learning–based methods to solve the problem for ALS point cloud classification. Some researchers projected row point clouds into 2D images [19] and then used convolutional neural networks (CNNs) for ALS classification to make full use of the advantages of CNNs [20]. The voxel-based method is another technique for ALS point cloud classification [21]. However, these projection methods usually require handcrafted features to enhance features and image representations, and the transformation from 3D to 2D inevitably cause information loss. In recent years, some studies have directly consumed raw point clouds and achieved state-of-the-art performance on some benchmarks.
The PointNet network [22] is a pioneering work that directly processes an irregular point. After the success of PointNet, many PointNet-like network architectures based on learned pointwise features have been proposed, such as PointNet++ [23], PointSift [24], PointCNN [25], PointWeb [26], DANCE-NET [27] and D-FCN [28]. Beside, the Graph-based method are develop rapidly [29, 30], such as LDGCNN [31] and Dynamic Graph CNN (DGCNN) [32]. ALS system has its own characteristics, including uneven point density, inexplicit structure and high redundancy. However, the works mentioned above have limitations and do not maximise the use of geometric structures. Thus, they lack the perception of geometry in encoded semantic features.
PointNet++ performs well on many tasks. However, its usage in large-scale airborne point clouds is not good. In PointNet++ network, the simple partitioning process cannot effectively capture a complicated relationship [33]. The recent advances in deep learning for point cloud processing are mainly focused on the designs of local aggregation operators. How to collect the information in the local area in the PointNet++ network still needs to be discussed.
In this study, we proposed a network modified from the PointNet++ network according to ALS point cloud characteristic. The main contributions of this study are shown as follows:
- The proposed network can be trained in an end-to-end manner. The feature on the centroid point and the neighbour relationship in the sample layer on the PointNet++ network is added to improve the representation ability to solve the uneven distribution problem.
- This study also utilises elevation information as weight on the up-sampling layer to support classification for the ALS point cloud, which exhibits discrepancies in elevation distributions.
- The class-balanced term is used on the 3D point cloud to solve the highly uneven category distribution problem.
Materials and methods
Feature learning in local region
The PointNet++ network processes irregular point clouds directly, it is highly robust to small perturbations and occlusion. Thus, the need for expensive manual feature computation is eliminated, and a new solution for the 3D scene process is provided. The core of the PointNet++ network is the sampling layer, grouping layer and PointNet layer. The PointNet++ network uses FPS to choose point part on the sampling layer, the KNN method to partition the grouping layer and the PointNet to collect features on the PointNet layer.
Pointnet++ treats individual points in local point sets independently. This approach lacks the perception of the whole neighbourhood structure. The feature on the centroid in the PointNet++ network is assigned by the neighbourhood points. This data processing method possesses the following drawbacks: (1) Directly replacing the centre point feature with the neighbourhood feature makes the selection centroid point very critical. (2) Information interaction amongst neighbours’ points is lacking.
One point is not isolated to other points, and the attributes are formed by plenty of points [34]. Many features exist in the local region: neighbourhood features, centre point features and relationship features between neighbourhood points. In this chapter, the PointNet++ network is modified to enrich the local neighbourhood information description. Moreover, the centroid point information, neighbourhood point relationship and normalised coordinate information are combined. The formula for feature learning is shown as follows:
(1)
where Δpij is the normalised coordinate information,
is the centroid i information, and Fjj′ is the relationship between neighbour point j and point j′. The Fjj′ value can be obtained in many ways. The simplest way involves the maximum and minimum sum and average. However, this method destroys the module structure. Thus, we use the adaptive method to learn contextual information in local regions.
(2)
where Wjj′ is the adaptive weight, which is calculated from the difference between the two feature vectors, i.e. features Fj and Fj′, after the MLP network. This adaptive weight uses convolution filters over neighbour point features and is shown as follows:
(3)
The strategy to calculate adaptive weight Wjj′ is also problematic because Wjj′ is equal to zero when j = j′, indicating that the impact of Fj on itself is zero. To increase more learned information, we set the impact of Fj on itself is estimated by its own feature Fj when j = j′.
The modified PointNet++ network has several abstraction layers and feature propagation layers with skip connections. The features on centroid and relationships amongst neighbours are also added. The architecture is shown in Fig 1.
This architecture added features on centroid and neighbour relationship. The input point clouds on the left are input points, and the output is the classification result.
Class-balanced loss function
The category distribution in ALS scene is highly uneven which can be deduced from the point number in each category. The training efficiency is low when most locations do not provide useful learning signals. The challenge of long-tailed training data can be alleviated using two strategies: resampling and reweighting. In resampling, the number of examples for the minor class is oversampling, and the major class is under-sampling. However, the resampling always introduces a large amount of duplicated samples. Thus, the model becomes susceptible to overfitting, and the training slows down. Reweighting the loss by inverse class frequency usually yields poor performance for highly imbalanced classes. This strategy has poor performance on large-scale datasets.
Class-balanced terms perform well on 2D images. However, their performance on ALS point cloud classification is unknown. The class-balanced loss function formula is shown as follows:
(4)
The main contribution of the class-balanced loss function is the added weighting factor , where ny is the number of samples in the ground truth class. When β = 0 corresponds to no reweighting, β→1 corresponds to reweighting by inverse class frequency. Therefore, adjusting the hyper-parameter β enables us to adjust the class-balanced term between no reweighting and reweighting via inverse class frequency.
The research shows that the class-balanced strategies perform remarkable improvements to existing commonly used loss functions, including sigmoid cross-entropy, softmax cross-entropy and focal loss [31]. The class-balanced softmax cross-entropy loss is
(5)
The class-balanced sigmoid cross-entropy loss is
(6)
The class-balanced focal loss is
(7)
The class-balanced term is designed to address imbalanced data training, and can be applied to various deep networks.
Adaptive elevation interpolation method
PointNet++ adopts a hierarchical propagation strategy with distance-based interpolation and across level skip links. The point feature propagation is achieved by interpolating feature values of Nl points at the coordinates of the Nl−1 points. The weight on PointNet++ is the inverse distance based on k nearest neighbours. The ALS point clouds are discrepancy distribution along the elevation. For example, the point elevation on the ground is the lowest, followed by low plants, roofs and trees. The elevation information can effectively help distinguish different objects.
In this part, we embed the elevation information to the interpolation function to improve the performance on ALS point cloud classification further. The weight coefficient of distance is calculated based on the elevation difference between two neighbour points.
The point i interpolated features f(l−1)(x) are calculated from the neighbouring j point feature f(l)(x), which is shown as follows:
(8)
where wij denotes the weight coefficient of distance, which is shown as follows:
(9)
where dij is the distance between points i and j, and w_zij is the weight coefficient of elevation, which is shown as follows:
(10)
where zij is the elevation diffident, and
is the average value of zij. Then, the interpolated features are concatenated with skip linked point features from the subsampled features.
Experimental result and analysis
The experiments are conducted on two airborne LiDAR point datasets to evaluate the effectiveness of the proposed method. The first one is the Vaihingen 3D labeling benchmark dataset released by International Society for Photogrammetry and Remote Sensing (ISPRS). This dataset was acquired by an airborne Leica ALS50 system with a 45° field of view and an average flying height of 500 m. The point density on the no-overlapping area and cross-flight belt area is approximately 4 and 8 point/m2, respectively, indicating that the point in this dataset is uneven. The ALS scene is complex and possesses numerous geometric instances, complicating the automatic recognition of the model. The distribution of training sets and test sets is shown in Fig 2. All points in the Vaihingen dataset are annotated in nine semantic categories: power line, low vegetation, impervious surfaces, car, fence, roof, façade, shrub and tree. The overview of the ISPRS Vaihingen 3D dataset is shown in Fig 2.
The legend at the bottom indicates the classification labels rendered in colours.
The Vaihingen benchmark dataset contains three blocks. Scene (I) is used for training, and scene (II) is utilised for evaluation. The point data number for training is 753876, and the number for testing is 411722. The proportion in different categories is shown in Table 1.
Table 1 shows that the proportion of the power line category accounts for only 0.07%, whereas the impervious surfaces category accounts for 25.70%. The proportions in different categories are extremely uneven. Training directly on this dataset can be misleading.
Following the evaluation metric of the ISPRS 3D benchmark dataset, we use three metrics to evaluate our method, i.e. F1 score, Overall Accuracy (OA) and mean intersection over union (MIoU). MIoU is used to measure the classification performance for all categories. F1 score is calculated from the precision and recall values. It is suitable for classification evaluation in uneven class distribution. The calculations of precision, recall, F1 score and MIoU are formulated as (11)-(14):
(11)
(12)
(13)
(14)
where TP, FN and FP are the true positives, false negatives and false positives, respectively. The average precision (AvgP), average recall (AvgR) and average F1 score (AvgF1) are also utilised.
Model training
We train our model on a single NVIDIA Tesla V100 GPU. The proposed method is implemented using the PyTorch framework. Given the limited GPU memory, the training scene is divided into small patches with regular blocks with a size of 40 m * 40 m in the horizontal direction. The empirical knowledge of the model design is from reference [14] and comparative experiments. The parameters on sampling number, batch size, decay rate, training epoch, learning rate and optimiser are set as 4096, 3, 0.007, 64, 0.001 and Adam. The parameters in the training are saved every five epochs. For convenience, the validation dataset is set in the same manner as the training dataset. The MIoU is calculated regularly to validate the model. The highest MIoU is regarded as the best model. During the model testing, scene (II) is also segmented into 40 * 40 blocks in the horizontal direction, with a stride of 20 m. Then, the blocks are input into the best model to test the performance of our proposed method. We also test the loss function, adaptive elevation, adding feature and generalisation ability to investigate whether the ALS point cloud processing strategy is feasible.
Test of the loss function
We first investigate the performance of our proposed method on the class-balanced term to select loss function. The ISPRS labelling dataset indicates that the point for each object category is quite different. Directly training on this unbalanced dataset may cause the classes with small numbers to become under fitted and misclassified. We use the modified loss function to address the issue mentioned by forcing our model to focus on the classes with few points. We experiment on the ISPRS dataset to investigate the performance of the class-balanced term and select the suitable function. The result is shown in Table 2.
From the value on Table 2 about AvgP, AvgR, AvgF1 and OA. the cross-entropy is low, and the focal loss function obtains a satisfying classification result. This result demonstrates that CB-focal loss obtains a quite stable performance. The CB-focal loss incorporates a category-specific weight factor for focal loss to reweight the classification loss. The class-balanced term combined with focal loss can outperform sigmoid cross-entropy and softmax cross-entropy. Therefore, we adopt CB-focal loss as the loss function for the following experiments.
Test of adaptive elevation
The ALS point clouds are discrepancy distribution along the elevation. We investigate the effect of the adaptive elevation interpolation method by embedding the elevation information to the interpolation function to improve the performance on ALS point cloud classification. The weight coefficient of the distance is calculated based on the elevation difference between two neighbour points. The performance of this adaptive elevation is shown in Table 3.
The AvgF1 and the OA improve by 0.4% and 0.8%, respectively. This finding indicates that the elevation information can effectively help distinguish different objects.
Test of added information in the local region
We develop three models to investigate the effect of adding features about centroid and neighbour relationship on the subsample in the PointNet++ network. Method (a) is the model without the centroid and neighbour features. Method (b) is the model with centroid point features. Method (c) is the model with a neighbour relationship. The model with the features of centroid and neighbour relationship is marked as ‘ours’. The classification performance of these four models is listed in Table 4.
Table 4 shows that the ALS point cloud classification performance is improved as the features become rich, and our proposed method has the highest OA and AvgF1. Method (a) and method (b) are limited for the feature extraction in distinguishing the classes in urban scenes because of scene complexity, unstructured nature of 3D point clouds, high sensor noise and incompleteness.
Our proposed method can effectively recognise most of the objects with an OA of 0.835 and AvgF1 0.718 by adding the feature about centroid point and neighbour relationship. The object class F1 scores that are higher than 60% are 7 out of 9, indicating that the proposed method can capture valuable features. Although the ISPRS dataset has a small number of points on the power line category, our proposed method achieves satisfying performance with an F1 score of 77.6%. Fig 3 shows the classification result of our proposed method and the error map.
Fig 3 shows that a large part of the test scenes is correctly classified. This observation is also validated by the error map. Table 5 shows the confusion matrix of the per-class accuracy of our proposed method.
The evaluation metrics about precision, recall and F1 score of each class are reported. The numbers in the confusion matrix are normalised along each row.
The confusion matrix shows that our proposed method obtains a quite satisfying performance on the impervious surfaces and power line categories as indicated by the F1 score. the proposed method also obtains acceptable performance on the roof and impervious surfaces categories. The shrub category and many shrub points are misclassified as low vegetation and tree points from the confusion matrix. The possible reason is that the shrubs have topological structures and elevation distribution similar to low vegetation and trees.
Discussion
Comparisons with other methods
We also compare our proposed method with other point-based models on the ISPRS 3D benchmark labelling. Table 6 lists the classification performance of our proposed method and all the methods compared.
The first nine columns in the table are the per-category F1 scores, and the last two columns are the OA and AvgF1.
Table 6 shows that the PointNet network has the lowest OA. The possible reason is that the PointNet network lacks neighbour information, and it does not collect enough features to represent the complex features on ALS point clouds. PointNet++ is a deep hierarchical network that recursively applies a unit PointNet on each grounded local region and makes full use of neighbour information. Compared with the baseline model (PointNet++ network), the proposed method increases by 2.5% in OA and 5.2% in AvgF1, which shows that the modified strategies are feasible.
DGCNN uses the dynamic graph convolutional neural to collect information. The collected neighbour point for centroid is the same. PointSift and D-FCN are the orientation-aware point feature learning methods. However, these methods ignore the relationship amongst neighbour points. PointCNN with A-XCRF [35] model obtains a high AvgF1. However, this model utilises a postprocessing step to refine the classification results. Our proposed method does not involve any postprocessing techniques.
Table 6 shows that the performance of our proposed method is better than the methods compared, as indicated by the AvgF1. Our proposed method also achieves remarkable higher performance on the powerline categories. The powerline category only takes a small proportion of the training data, which is difficult to predict, suggesting that the class-balanced term is feasible.
Validation of generalisation ability
The GML(B) dataset are also conduct to validate the generalisation ability of our model, which belong to GML dataset for the B part and was acquired by the airborne Leica ALTM 2050 system. All points in this dataset are annotated in four semantic categories, including ground, building, tree and low vegetation. Each point contains only coordinate features. We use x, y, z coordinates as the model inputs in our experiments. Moreover, the hyperparameters are the same as the hyperparameters for the experiments on the Vaihingen dataset, except for the size of the segmented region, which is set as 48 m in the horizontal direction for the point density is approximately 7 point/m2, which is large than ISPRS benchmark dataset at 4–8 point/m2. We also compare the performance of our proposed method with that of PointNet, PointNet++ and the Random Forest on GML(B) benchmark labelling dataset. Table 7 lists the classification performance of our proposed method and all the methods compared.
The first four columns are the F1 scores for different classes, and the last two columns are the AvgF1 and compute time.
Table 7 shows that the ground, building and tree are well recognised. The worst classification result lies in the low vegetation. Our proposed method achieves the best performance in AvgF1. As for computing time comparison, PointNet network has the higher processing efficiency and the lowest precision, which indicates that lacking neighborhood features is not well to the recognition of objects. PointNet++ network still has low precision in large-scale complex scenes classification for this network still considers each point in the local region independently. Our proposed method has the lowest computational efficiency, which is mainly related to the time in extracting neighborhood point relationships and center point features. The result on Random Forest models is come from reference [36], which do not have compute time. However, the Random Forest model take each point’s local geometry independently and ignore the spatial dependencies. Fig 4 shows the classification result of our proposed method and the error map.
In Fig 4, many points are classified correctly, and errors are mainly distributed over the object edges. The modified PointNet++ network makes full use of receptive field information at all levels on the ALS point classification task. Our model can produce an accurate classification for the majority of ALS point clouds.
Conclusion
This study proposes a modified PointNet++ network for airborne LiDAR point cloud classification based on their own characteristics, and verify the advantages of our proposed method through comprehensive ablation experiments on Vaihingen 3D semantic labelling benchmark dataset and the GML(B) dataset. Compared with the baseline model (PointNet++ network) from the Vaihingen datasets, the proposed method increases by 2.5% in OA and 5.2% in AvgF1, which shows that the modified strategies are feasible. Besides, the proposed method also achieves a new state-of-art performance for the power line category. The powerline category only takes a small proportion of the training data, which is difficult to predict, suggesting that the class-balanced term is feasible. The generalisation ability is verified by GML(B) dataset and our model can produce an accurate classification for the majority of ALS point clouds. The modified PointNet++ network makes full use of receptive field information at all levels and the elevation information can effectively help distinguish different objects on the ALS point classification task. Our method does not involve any postprocessing techniques and can operates unordered point sets with varying densities, which can be widely used in ALS point classification.
References
- 1. Polewski P.; Yao W.; Heurich M.; Krzystek P.; Stilla U. Detection of fallen trees in ALS point clouds using a Normalized Cut approach trained by simulation. ISPRS J. Photogramm. 2015, 105, 252–71.
- 2. Dong M.; Chou W.; Fang B. Underwater Matching Correction Navigation Based on Geometric Features Using Sonar Point Cloud Data. Sci. Programming-Neth. 2017, 2017, 1–10.
- 3. Liu M, Shao Y, Li R, et al. Method for extraction of airborne LiDAR point cloud buildings based on segmentation[J]. PloS one, 2020, 15(5): e0232778.
- 4. Wang L, Xu Y, Li Y, et al. Voxel segmentation-based 3D building detection algorithm for airborne LIDAR data[J]. Plos one, 2018, 13(12): e0208996.
- 5. Huang R.; Xu Y.; Stilla U. GraNet: Global relation-aware attentional network for semantic segmentation of ALS point clouds. ISPRS J. Photogramm. 2021, 177, 1–20.
- 6. Guo Y, Wang H, Hu Q, et al. Deep learning for 3d point clouds: A survey[J]. IEEE transactions on pattern analysis and machine intelligence, 2020, 43(12): 4338–4364.
- 7. Pan Y.; Dong Y.; Wang D.; Chen A.; Ye Z. Three-Dimensional Reconstruction of Structural Surface Model of Heritage Bridges Using UAV-Based Photogrammetric Point Clouds. Remote Sens.-Basel. 2019, 11, 1204.
- 8. Chen Y.; Liu G.; Xu Y.; Pan P.; Xing Y. PointNet++ Network Architecture with Individual Point Level and Global Features on Centroid for ALS Point Cloud Classification. Remote Sens.-Basel. 2021, 13, 472.
- 9. Li N.; Kahler O.; Pfeifer N. A Comparison of Deep Learning Methods for Airborne Lidar Point Clouds Classification. Ieee J.-Stars. 2021, 14, 6467–86.
- 10. Chu H, He Z, Liu S, et al. Deep Neural Network for Point Sets Based on Local Feature Integration[J]. Sensors-Basel, 2022, 22(9): 3209.
- 11. Miao Y, Hunter A, Georgilas I. An Occupancy Mapping Method Based on K-Nearest Neighbours[J].Sensors-Basel. 2021, 22(1): 139.
- 12. Tran G.; Nguyen D.; Milenkovic M.; Pfeifer N. Potential of full waveform airborne laser scanning data for urban area classification-Transfer of classification approaches between missions. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2015, XL-7/W3, 1317–23.
- 13. Lafarge F.; Mallet C. Creating Large-Scale City Models from 3D-Point Clouds: A Robust Approach with Hybrid Representation. Int. J. Comput. Vision. 2012, 99, 69–85.
- 14. Zhang J.; Lin X.; Ning X. SVM-Based Classification of Segmented Airborne LiDAR Point Clouds in Urban Areas. Remote Sens.-Basel. 2013, 5, 3749–75.
- 15. Ni H.; Lin X.; Zhang J. Classification of ALS Point Cloud with Improved Point Cloud Segmentation and Random Forests. Remote Sens.-Basel. 2017, 9, 288.
- 16. Niemeyer J.; Rottensteiner F.; Soergel U.; Heipke C. Hierarchical higher order crf for the classification of airborne lidar point clouds in urban areas. ISPRS—International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2016, XLI-B3, 655–62.
- 17. Niemeyer J.; Rottensteiner F.; Soergel U. Conditional random fields for lidar point cloud classification in complex urban areas. ISPRS annals of the photogrammetry, remote sensing and spatial information sciences. 2012, I-3, 263–8.
- 18. Kalogerakis E.; Averkiou M.; Maji S.; Chaudhuri S. 3D Shape Segmentation with Projective Convolutional Networks. 2016.
- 19. Zhao R.; Pang M.; Wang J. Classifying airborne LiDAR point clouds via deep features learned by a multi-scale convolutional neural network. International journal of geographical information science: IJGIS. 2018, 32, 960–79.
- 20. Qin N.; Hu X.; Dai H. Deep fusion of multi-view and multimodal representation of ALS point cloud for 3D terrain scene recognition. ISPRS J. Photogramm. 2018, 143, 205–12.
- 21. Qi C.R.; Su H.; Mo K.; Guibas L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition. 2017.
- 22. Liu Z.; Song W.; Tian Y.; Ji S.; Sung Y.; Wen L.; et al. VB-Net: Voxel-Based Broad Learning Network for 3D Object Classification. Applied Sciences. 2020, 10, 6735.
- 23. Qi C.R.; Yi L.; Su H.; Guibas L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Advances in Neural InformationProcessing Systems. 2017, 5099–108.
- 24. Jiang M.; Wu Y.; Zhao T.; Zhao Z.; Lu C. PointSIFT: A SIFT-like Network Module for 3D Point Cloud Semantic Segmentation. Computer Vision and Pattern RecognitionComputer Science—Computer Vision and Pattern Recognition, 2018.
- 25. Li Y, Bu R, Sun M, et al. Pointcnn: Convolution on x-transformed points[J]. arXiv preprint arXiv:1801.07791 (2018).
- 26. Zhao H.; Jiang L.; Fu C.; Jia J. PointWeb: Enhancing Local Neighborhood Features for Point Cloud Processing. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019,5560–8.
- 27. Li W.; Wang F.; Xia G. A geometry-attentional network for ALS point cloud classification. ISPRS J. Photogramm. 2020, 164, 26–40.
- 28. Wen C.; Yang L.; Li X.; Peng L.; Chi T. Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification. ISPRS J. Photogramm. 2020, 162, 50–62.
- 29. Huang R.; Xu Y.; Hong D.; Yao W.; Ghamisi P.; Stilla U. Deep point embedding for urban classification using ALS point clouds: A new perspective from local to global. ISPRS J. Photogramm. 2020, 163, 62–81.
- 30. Simonovsky M.; Komodakis N. Dynamic Edge-Conditioned Filters in Convolutional Neural Networks on Graphs. Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- 31. Zhang K, Hao M, Wang J, et al. Linked dynamic graph cnn: Learning on point cloud via linking hierarchical features[J]. arXiv preprint arXiv:1904.10014, 2019.
- 32. Wang Y.; Sun Y.; Liu Z.; Sarma S.E.; Bronstein M.M.; Solomon J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM T. Graphic. 2019, 38, 1–12.
- 33. Wang X, He J, Ma L. Exploiting local and global structure for point cloud semantic segmentation with contextual point representations[J]. Advances in Neural Information Processing Systems, 2019, 32.
- 34. Bazazian D.; Parés M.E. EDC-Net: Edge Detection Capsule Network for 3D Point Clouds. Applied Sciences. 2021, 11, 1833.
- 35. Arief H.A.A.; Indahl U.G.; Strand G.; Tveite H. Addressing overfitting on point cloud classification using Atrous XCRF. ISPRS J. Photogramm. 2019, 155, 90–101.
- 36. Shapovalov R.; Velizhev A.; Barinova O. Non-Associative Markov Networks For 3d Point Cloud Classification. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 2010,38,103–108.