Figures
Abstract
Low-end LiDAR sensor provides an alternative for depth measurement and object recognition for lightweight devices. However due to low computing capacity, complicated algorithms are incompatible to be performed on the device, with sparse information further limits the feature available for extraction. Therefore, a classification method which could receive sparse input, while providing ample leverage for the classification process to accurately differentiate objects within limited computing capability is required. To achieve reliable feature extraction from a sparse LiDAR point cloud, this paper proposes a novel Clustered Extraction and Centroid Based Clustered Extraction Method (CE-CBCE) method for feature extraction followed by a convolutional neural network (CNN) object classifier. The integration of the CE-CBCE and CNN methods enable us to utilize lightweight actuated LiDAR input and provides low computing means of classification while maintaining accurate detection. Based on genuine LiDAR data, the final result shows reliable accuracy of 97% through the method proposed.
Citation: Mohd Romlay MR, Mohd Ibrahim A, Toha SF, De Wilde P, Venkat I (2021) Novel CE-CBCE feature extraction method for object classification using a low-density LiDAR point cloud. PLoS ONE 16(8): e0256665. https://doi.org/10.1371/journal.pone.0256665
Editor: Jie Zhang, Newcastle University, UNITED KINGDOM
Received: May 24, 2021; Accepted: August 11, 2021; Published: August 25, 2021
Copyright: © 2021 Mohd Romlay et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting Information files.
Funding: This research is supported by the FRGS 2019 Grant: FRGS/1/2019/ICT02/UIAM/02/2 awarded by the Ministry of Education Malaysia.
Competing interests: The authors have declared that no competing interests exist.
Introduction
A LiDAR sensor provides a solution for mobile applications where the system needs to be compact, lightweight and handy [1, 2]. It has a 360° degree field of view [3], possesses high accuracy of distance measurement and in contrast to a camera, it does not depend on the light intensity of the surroundings [4, 5]. The LiDAR sensor is robust to illumination variation [6], and can be used to obtain the transformation matrix between 2D coordinate system and 3D model of the scene [7]. Its detection range is also comparatively higher in accuracy and provides better reliability when compared to stereo methods [8, 9].
In some applications where portability and mobility are of prime importance, a single sensor which acts as the detection system is required [10, 11]. Examples of such applications are mobile monitoring robots [12], robotic navigation aids [13], skateboarder near-crash identification [14], smart canes for blind people and traffic monitoring for electric-assisted bicycles (e-bikes) [15]. Hence, there is a demand for system navigation by using a single portable 3D sensor [16].
Related works
An integral parts of robotics [17–20], object recognition with LiDAR sensor commonly depends on feature extraction for classification of objects. Accurate object detection and classification allows object tracking, road signs detection [21], scene understanding [22] and behaviour recognition [23]. 3D LiDAR traits based on local surface and key-points are amongst the main features for extraction within object recognition [24]. These key-point detection algorithms detect variances, normal vector, curvature or any other spatial geometric attributes as its feature extraction objectives. Following transformation, these key-points are examined for matches between predicted object and newly identified object. However, this does result in a significant increase in computational cost and time as the feature extraction solely depends on finding neighboring surrounding and determining adjacent points [25].
Other researchers suggested usage of histograms based on key-points to gather spatial features of point clouds in varying dimensions. The authors of [26] measured orientation angles in between points and their neighbors to create a histogram which is based on feature descriptors; namely fast point feature histogram (FPFH). Despite the FPFH showing promising results in terms of processing timing, the neighboring space of the FPFH descriptor is still considerably broad, resulting in a large execution time.
Yamada et al. [27] present a gait-based human identification using a real-time multi-line LiDAR. The author combines LiDAR data with long short-term memory for gait recognition with different appearances. Even though it performs well for on face recognition and human identification, gait-based feature extraction is mostly suitable for biometric recognition. Hence, limiting its object of detection from other objects in the environment.
The authors of [28] proposed a 3D convolutional kernel at varying scales to derive the features of targeted objects with distinct resolutions. Hence, the positional structure of the point clouds can be assembled in a more particular manner in which object classification accuracy, semantic recognition and other applications in the sequence are enhanced. However, the requirements of revising new points in a tree-based storage construction are high, specifically where point clouds were collected incrementally along the process.
In contrast to spatial feature extraction from individual points, clusters of points can be considered as voxels to be processed as coarse feature extraction. In [29], the authors suggested a semantic classification, partitioning spatial space into organized fixed-sized voxels. This solves the issue of processing difficulties within uneven density distributions. However, voxels with fixed size segregation may be inadequate for object recognition because of the unsymmetrical framework of the point clouds and its densities.
The authors of [19] computed a feature vector for pedestrian recognition with 195 collected features including 2D covariance matrix, normalize 2D histogram and slice feature. Tian et al. [30] proposed multiple feature extraction which consists of 9 features including density, centroid and variance. However, the features are from a global voxel are made from the entire LiDAR’s point of view thus contributing to heavy computation. This is mainly due to calculation cost is proportion to the number of voxels rather than the number of points [31]. Choi et al. [32] proposed using the Region of Interest (ROI) method with smaller number of features extracted. Even though only basic features are extracted such as width length and height, it does require an eagles-eye point of view by an airborne LiDAR or integration of numerous numbers of LiDAR’s for accurate scanning.
A few researchers chose bottom-up or top-down approaches to extract information and distinguish objects, however they do not consider the variation of point densities which will cause significant change of accuracy [33]. This is due to density changes with the distance from the LiDAR [34]. Especially in a very sparse LiDAR data recognition performance drastically decrease as distance between human and LiDAR increases, due to the number of points being inversely proportional to the square of the distance between a human to LiDAR [35].
An object detection network which views data in the form of a matrix is proposed by [36], with continuous properties from channel views as its extracted features and 2D convolution network. However, the approach requires LiDAR’s with higher density to feed channels of scanning for extraction. Due to disparity in object detection using sparse LiDAR point cloud, there is a necessity to develop a method which could perform well with such limited input. Existing works mainly provide solutions for high density point clouds, often involving heavy computational cost and computing load. Alternative methods which work on low density data are commonly limited to binary classification and insubstantial when dealing with multiple class detection. Table 1 shows the findings of main comparison feature extraction methods.
Contributions
It is vital to develop a classification method which can work with a sparse 3D point cloud, while providing enough leverage for the classification process to accurately recognize objects within limited computing capacity. Thus, this paper proposes an object recognition system with multiple feature extraction based on segregated clusters from LiDAR point clouds. To extract geometry features, we rasterize each point cloud of the object in a local voxel slice model based on its centroid.
The method proposed introduces local abscissa, ordinate and applicate (z-axis) voxels to reduce the computational cost of computing global voxels in the spatial domain while removing uncertainties of varying densities, discarding rigid transformation and unsymmetrical structure of point clouds associated with global voxel with arranged fixed-sized segregation. This research further introduces a novel feature extraction technique which considers a collection of features viz., density to centroid height ratio and density to volume ratio. These features capture the point cloud disparity and achieves a higher detection rate when compared to other state-of-the-art feature extraction methods. The proposed feature extraction helps to overcome the inconsistent point cloud detection due to the single point of view in scanning, allowing accurate object recognition from using a single actuated LiDAR sensor.
Employing machine learning for object classifiers has been a major interest of researchers as a means to train extracted features including for LiDAR point cloud classification [39, 40]. Bobkov et al. [41] implement a convolutional neural network (CNN) with 5 filters and pooling for layer extraction. Whereas Tian et al. [30] implemented multiple object features with annotated labels incorporated with an initialized neural network. Considering the success of machine learning algorithms in various areas including feature-based object classification [42], this research further optimizes the features extracted from the proposed method to be trained with selected machine learning optimizers. The algorithms selected are the k-nearest neighbor (k-NN), decision tree (DT) and convolutional neural network (CNN).
For the class of object detection, static object classification has been shown with target-level and with low level data [43]. In this paper, we will focus on methods addressing moving road users in conventional streets. We selected three important class of objects detection, categorically pedestrians, motorcyclists and cars. These three objects are typical on-road scene [44, 45]. Thus, its detection provide critical information for security and surveillance, law enforcement monitoring, search and rescue team [46, 47].
To the best knowledge of the authors, no existing work in the field of machine vision explicitly exploits information from a single unit, sparse LiDAR sensor for 3D scanning to achieve object recognition with high accuracy rates. We prove that object recognition can be obtained using a single unit, single stripe, actuated LiDAR with low computing necessity via the fusion of Clustered Extraction and Centroid Based Clustered Extraction (CE-CBCE) methods to accomplish high-reliability object recognition from sparse LiDAR point cloud data. To summarize, our main contributions are
- State-of-the-art combination of the Clustered Extraction (CE) and Centroid Based Clustered Extraction (CBCE) method which includes features extracted from the abscissa, ordinate and applicate voxels, a novel density to centroid height interval ratio and density to volume ratio. These features of sparse LiDAR point cloud data allow accurate classification from a single detection sensor.
- Result analysis and comparison of the CE-CBCE method trained by using k-NN, DT and CNN classification methods. The CE-CBCE optimized with the CNN classification recorded the best accuracy, excellent and consistent scores in terms of recall, precision and F1-score. The results show that the proposed method outperforms other state of the art feature extraction methods.
- Genuine 3D LiDAR point cloud data taken from a custom-built mobile robot with a detection system from a single LiDAR sensor. The data consist of 1200 scans of 3 main objects classes with 4 pose orientation headings. The data have been made public and can be accessed accordingly [48].
The rest of the paper is organized as follows. Section II introduces the technical issues and proposed methods for gathering and processing data from the LiDAR sensor. Here, the description of each step is explained in detail. Section III presents the results and analysis of the output from our experiment. Finally, conclusions and future recommendations are discussed in Section IV.
Proposed method
Primarily the research intends to propose a novel clustering-based feature extraction technique to exploit discriminative features from the scarce LiDAR point cloud data. Initially, the process starts with background filtering and clustering the raw point cloud data. Then the proposed method extracts features from the clustered object point clouds. The proposed method is divided into two parts namely Clustered Extraction (CE) and the Centroid Based Clustered Extraction (CBCE) method. Through this method, there are less computational power required when compared to using global voxels within the spatial domain. The final stored elements were taken from sparsely distributed values in the LiDAR point cloud, which will then be trained by selected classification methods (k-NN, DT and CNN) for human detection. A written consent of this research which involves detection of human subjects have been approved by IIUM Research Ethics Committee (IREC) with ID No: IREC 2017–066.
The following section explains the details of each procedure step by step. The sequence starts with data collection, filtering, clustering and finally object classification.
Data collection
For genuine data collection purposes, we have constructed a mobile robot with LiDAR based sensor. The hardware components include Garmin LiDAR Lite v3, Arduino Uno, FS5109 servo motor, L298N motor driver. For wireless communication, XBee which comes readily with TX and RX communication modules, allowing wireless data transmission [49] and Li-Po external battery are required, providing mobility to the system. The scanning degree is fixed at 130°, resembling a human’s point of view [50]. The mobile robot scanning can be seen in Fig 1.
Over 1200 scans have been collected which contained 400 objects for each of the three cluster categories viz., human, motorcyclist and cars. These three classes of objects are the most commonly found for on-road scene. For comparison purposes, we have selected the same number of samples for each class and orientation. The scenes are recorded within indoor and outdoor environments, during day and night for better reflection of the real-world environment. The position of the object detected varies with distance up to 40 meters from the mobile robot. This is the effective distance of the proposed method and the range of detection for the LiDAR sensor. The clustered point clouds are classified into 3 categories as mentioned before.
However even for the same object, different poses (rotations as well as translations) with respect to the LiDAR could result in different coordinates (ex: varying x/y/z minimum and maximum, object centroid and volume size). Therefore, we have taken the sample with various distance from 1 m to 40 m, with different pose orientation of the object from front, right left and back side of the targeted sample. Fig 2 below shows the total number of samples across the 3 class of object detection and its orientation facing the mobile robot.
Filtering
From the raw point cloud data collected with the mobile robot, unnecessary noises are removed from the scene. A threshold value of 500 z coordinate is fixed, approximately 5 meters from the ground. The points above the threshold are considered as non-disturbance. It does not pose as an obstacle for the mobile robot movement and does not represent any classes from the targeted class of object detection. Therefore, all points which surpass the set threshold value are subtracted from the point cloud, before entering the clustering process.
Clustering
Following filtering, the remaining point cloud goes through the process of clustering with the k-means clustering algorithm. Fig 3 shows raw, filtered and clustered data of all subjects of recognition.
Up until this point all classification methods go through the same scanning, filtering and clustering procedures. In the next step of feature extraction, the proposed technique will be compared with other state-of-the-art classification method. To solely compare the performance of each feature extraction method, we have designed the process to be non-end-to-end classifiers with pre-processing procedures (filtering & clustering) and post-processing steps (classification methods using DT, k-NN and CNN).
Clustered Extraction (CE) method
The LiDAR point cloud gives output in Cartesian coordinate system with the x, y & z origin set to be the position of the LiDAR on top of the mobile robot.
The first part of our proposed clustered extraction (CE) method is the feature extraction of α, β, γ. Given a point cloud,
(1)
the CE method proposed is explained below as,
(2)
where P is the clustered LiDAR point cloud data, extracted into three parts. The first part is denoted by alpha α, which stores the values of width (w), length (l), height (h) of the object, and the number of points in the cluster (N). The second part is represented by the array beta β which stores the number of elements within the segregated intervals represented as xdataset, ydataset and zdataset. The unique feature of β is the number of points derived from abscissa, ordinate and applicate voxels. The third part is the minimum and maximum value of each axis in the clusters, denoted by gamma γ.
From Eq (1), we compute the centroid of the cluster c,
(6)
Centroid c acts as the origin of the local voxels for each cluster. From the centroid, addition for voxel borders is constructed with a predetermined increment value. These will act as the abscissa, ordinate and applicate voxels. Therefore,
and
(7)
where |.| indicates cardinality. The initial value of the dataset interval is denoted as xfloor; where xfloor presents the value of xmin rounded down to increment value Δ (in this case we set it to be 50),
(8)
Finally, the end value of the dataset is defined as xceiling, which is the xmax value rounded up to the nearest hundred,
(9)
The stored elements of maxima and minima of the coordinate together with the number of elements within a determined interval serve as the input for the human detection classifier. The same procedures are done to acquire ceiling and floor value of ydataset and zdataset.
Centroid Based Clustered Extraction (CBCE) method
From here onwards, two additional collective features are extracted from the point cloud denoted by delta δ (for features related to density to centroid height ratio () and epsilon ε (for features related to density to volume ratio (
). First, the collective features of δ are discussed.
Initially, the height of the object h is determined and it is divided by the total number of parts t set to be 10 as default to acquire height of i-th part of n . For each part, the centroid height is calculated as a reference point.
For i = 1,2,…t. Height of cluster n, is calculated in (12).
For . Then density between interval is denoted by
acquired through subtracting current cluster density
to previous cluster density
(13)
From here we obtain the density to height ratio.
Next, the collective features related to density to volume ratio (denoted by ε) is defined as:
(15)
For i = 1,2,…t. The total number of parts t set to be 10 as default as shown in Eq (11)
(16)
where i represents the number of parts. The volume difference
is acquired by subtracting current cluster volume
to previous cluster volume
(17)
The density difference is acquired as shown in (13). Therefore, the density to volume ratio is denoted by:
(18)
The flow chart of the proposed CE-CBCE feature extraction method can be seen in Fig 4.
Summary of features extracted are shown in Table 2 with its dimensional count and feature description.
From the extracted features, classifications are done with 75%-25% split of training and testing data. Training aimed to decrease the model loss function value against training data as each step was processed. Model performance was indicated and measured through improvements in accuracy of the model against the test dataset [51].
The accuracy of the classifications is calculated as follows:
(19)
With k representing the total number of class, TPi as the true positive, TNi as the true negative, FPi as false positive and FNi as false negative; for i = 1,2,3,…k.True positive is when the model correctly predicts the positive class, and a false negative is recorded when a class is incorrectly predicted to be negative. False positive occurs when a class is incorrectly predicted to be positive, and true negative is considered when the model correctly predicts the negative class. However for multiclass classification such as our case [48], true positive occurs only when the right class is correctly predicted, and similarly for false negative etc.: they all depend on the class.
Results and discussions
Post CE and CBCE extraction, the collective features are optimized with kNN, DT and CNN. The method proposed is compared with three feature extraction methods which are the region of interest (ROI) [32, 52], feature vector (FV) [19] and multiple feature extraction (MFE) [30].
The chosen comparative method is considered as it similarly handles sparse point cloud, has low computing cost, employs geometrical features and runs on real-time execution. For ROI, the authors proposed taking width (w), length (l), height (h), width difference (Δw) and length difference (Δl) as the extracted features for classification. So, the complete geometric feature is ΔG = [w,l,h,Δw,Δl].
The second comparison method is the feature vector (FV) introduced by Wang et al. [19]. The feature extracted includes 2D covariance matrix in 3 zones, 2D histogram for x-y plane and 2D histogram for y-z plane. The total dimensional count amounts to 175 features.
The third comparison method is proposed by Tian et al. [30], with multiple feature extraction which includes point count (N), point density (ρ), voxel centroid (μ), point variance (σ2), point covariance (), point eigenvector (ν), point eigenvalue (γ), surface curvature (k) and divergence degree (F). The final feature extraction of MFE includes 27-dimensional count, with 9 counts from point eigenvector, 3 counts each from voxel centroid, point variance, point covariance, point eigenvalue and divergence degree, and finally a single dimension count from point count, point density and surface curvature.
The results obtained are then compared in terms of accuracy, precision, recall and F1-score. k-Fold Cross-validation is performed on the classifiers to select model parameters which best fit our data. Table 3 shows the complete hardware and software configurations for the experiments conducted. The simulation was done on an Intel (R) Core (TM) i7-5500U CPU @ 2.40GHz, 8Gb RAM and 64-bit operating system.
Classification experiments
The experiment was done on an Intel(R) Core (TM) i7-5500U CPU @ 2.40GHz, 8Gb RAM and 64-bit operating system.
The results of our experiment are represented in Table 4. The "Parameter" column shows the varying parameters to be determined to achieve the best configuration for each optimization algorithm. The parameters refer to the number of nearest neighbors for k-NN, the maximum depth for DT and the number of layers for CNN. For DT and k-NN the parameters are set in the range of 1 until 30. For CNN, the number of hidden layers is ranged between 1 to 10, with the batch size of 10, 1000 the number of epochs, and an activation function of Rectified Linear Units (ReLUs) and Softmax function. For the proposed method, initially the CE and CBCE methods are implemented separately, before combining both collective features to show the improved performance.
Table 4 shows the comparison of the results in terms of accuracy for ROI, FV, MFE, CE, CBCE and fusion of CE-CBCE combined. The selected results displayed are with the increment of 5 for k-NN and DT parameters and 2 for CNN parameter.
From Table 4, the general accuracy overview results in over 65% accuracy for all feature extraction methods. In terms of consistency, the CE method scores the highest mean accuracy at 88.3%, followed by CE-CBCE (85.4%), CBCE (84%), MFE (83.2%), ROI (79%) and finally FV with 69.3%.
A complete comparison in terms of mean accuracy, precision, recall, and F1-score can be seen in Fig 5. In general, the CE method prevails in terms of accuracy, precision, recall and F1-score across all optimization methods of k-NN, DT and CNN.
After analysis of the statistics across all optimization methods, a particular examination of each best performing optimization parameter is done. Specifically, the CE-CBCE method optimized by CNN with 6 hidden layers recorded best accuracy for object recognition and classification at 97% detection. This is followed by both the CE and CBCE method optimized with k-NN with the k-value of 1 and CNN with 8 hidden layers, respectively. Both methods recorded accuracy of 93% detection. The rest of the feature extraction methods achieved 91% for MFE, 87% for ROI and 79% for FV. Fig 6 shows the best optimization results in terms of precision, recall, F1-score and accuracy for each feature extraction method.
The best parameter choice for each method can be seen in Table 5. The precision, recall and F1-score for each individual class of human, motorcyclists and cars are presented with its final accuracy.
Full results of CE-CBCE across all optimization techniques are shown in Fig 7 on the following page. Results achieved across all parameters are in the graphs within said figure.
As satisfying results are achieved for object recognition amongst 3 classes of objects, another experiment with added difficulty is conducted. The aim of classifying within 3 classes of objects remains the same, however this time around the output of the prediction includes the pose of subject. Therefore, for each detected object, the direction pose needs to be predicted whether it is facing front, right, back of left side towards the mobile robot. The same number of samples have been provided as the input, with 300 samples for each class and 100 samples for each orientation. Table 6 shows the best results obtained from each feature extraction methods.
In the Table 6, excellent results for the object’s orientation are recorded. The combination of CE-CBCE method achieved accuracy of 82%, followed by CE, CBCE, MFE, ROI and finally FV methods. Especially for easily perceived poses for human subjects facing right or left, CE-CBCE and CE methods both recorded 100% detection rate. The radar chart in Fig 8 shows the accuracy of CE + CBCE method for each object with varying pose orientation.
Conclusions
In this research, we proposed a novel feature extraction for sparse LiDAR point cloud object recognition. Indoor and outdoor data were collected, with different backgrounds for better simulation of varying surroundings. We also analysed the performance of our feature extraction method with different classes of objects in varying pose orientations. The result shows a promising achievement with a sparse LiDAR point cloud. The flow of the research proposed can be seen in Fig 9. The process started form hardware development, before moving into genuine data collection, preprocessing, proposed feature extraction method, classification algorithms and finally object classification.
As the proposed method targeted sparse LiDAR point cloud input, its performance on high density data remains to be explored. High density, compact point clouds such as autonomous vehicle and airborne LiDAR are often associated with large scale mapping and varying elevation scanning. A larger number of classes for object detection could also pose a challenge as they tend to be less distinctive in terms of density and point cloud distribution.
For future research, position tracking can be implemented on top of the object recognition and classification. Especially for a safety-critical system where accuracy is of utmost importance, safety features should be a main concerning issue. A fail-safe mechanism in place to override the controls in case of a malfunction could be implemented, with a warning system which alerts the user during detection of a faulty device to allow human intervention. Additional elements or subjects of detection can also be added to further test and improve the reliability of the system.
References
- 1.
Ogawa H, Tobita K, Sagayama K,Tomizuka M. A Guidance Robot for the Visually Impaired: System Description and Velocity Reference Generation. In: Computational Intelligence in Robotic Rehabilitation and Assistive Technologies (CIR2AT). 2014. p. 9–15.
- 2. Kim M, Lee K, Han Y, Lee J, Nam B. Generating 3D texture models of vessel pipes using 2D texture transferred by object recognition. Journal of Computational Design and Engineering. 2021;8(1):1–13.
- 3. Benedek C. 3D People Surveillance on Range Data Sequences of a Rotating Lidar. Pattern Recognition Letters [Internet]. 2014;50:149–58. Available from: http://dx.doi.org/10.1016/j.patrec.2014.04.010
- 4.
Li Q, Chen S, Wang C, Li X, Wen C, Cheng M, et al. LO-Net: Deep Real-time Lidar Odometry. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2019. p. 8473–82.
- 5. Tu X, Xu C, Liu S, Lin S, Chen L, Xie G, et al. LiDAR Point Cloud Recognition and Visualization with Deep Learning for Overhead Contact Inspection. Sensors. 2020;20(21):1–18. pmid:33182360
- 6. Zou C, He B, Zhu M, Zhang L, Zhang J. Learning Motion Field of LiDAR Point Cloud with Convolutional. Pattern Recognition Letters [Internet]. 2019;125:514–20. Available from: https://doi.org/10.1016/j.patrec.2019.06.009
- 7. Li T, Suzuki H, Ohtake Y. Visualization of user ‘s attention on objects in 3D environment using only eye tracking glasses. Journal of Computational Design and Engineering. 2020;7(March):228–37.
- 8.
Qiu J, Cui Z, Zhang Y, Zhang X, Liu S, Zeng B, et al. DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene from Sparse LiDAR Data and Single Color Image. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2019. p. 3313–22.
- 9.
Ibrahim AM, Saifullah M, Romlay MRM, Venkat I, Ibrahim I. Hybrid Social Force-Fuzzy Logic Evacuation Simulation Model For Multiple Exits. In: 2019 7th International Conference on Mechatronics Engineering (ICOM). IEEE; 2019. p. 1–5.
- 10. Wang Y, Peng C, Ravankar AA, Ravankar A. A Single LiDAR-Based Feature Fusion Indoor Localization Algorithm. Sensors. 2018;18(4). pmid:29690624
- 11. Premebida C, Ludwig O, Nunes U. LIDAR and Vision-Based Pedestrian Detection System. 2009;26(Iv):696–711.
- 12. Romlay MRM, Azhar MI, Toha SF, Rashid MM. Two-wheel Balancing Robot; Review on Control Methods and Experiment. International Journal of Recent Technology and Engineering (IJRTE). 2019;7(68):106–12. pmid:25983678
- 13. Rabani M, Romlay M, Toha SF, Ibrahim AM, Venkat I. Methodologies and evaluation of electronic travel aids for the visually impaired people: a review. Bulletin of Electrical Engineering and Infromatics. 2021;10(3):1747–58.
- 14. Wu J, Zhang Y, Xu H. A novel skateboarder-related near-crash identi fi cation method with roadside LiDAR data. Accident Analysis and Prevention [Internet]. 2020;137(July 2019):105438. Available from: pmid:32004863
- 15. Blankenau I, Zolotor D, Choate M, Jorns A, Homann Q, Depcik C. Development of a Low-Cost LIDAR System for Bicycles. SAE Technical Paper. 2018;1–9.
- 16. Pereira M, Silva D, Santos V, Dias P. Self calibration of multiple LIDARs and cameras on autonomous vehicles. Robotics and Autonomous Systems [Internet]. 2016;83:1–12. Available from: http://dx.doi.org/10.1016/j.robot.2016.05.010
- 17.
Wang Y, Chao W, Garg D, Hariharan B, Campbell M, Weinberger KQ. Pseudo-LiDAR from Visual Depth Estimation: Bridging the Gap in 3D Object Detection for Autonomous Driving. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2019. p. 8445–53.
- 18.
Li P, Chen X, Shen S. Stereo R-CNN based 3D Object Detection for Autonomous Driving. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2019. p. 7644–52.
- 19. Wang H, Wang B, Liu B, Meng X, Yang G. Pedestrian recognition and tracking using 3D LiDAR for autonomous. Robotics and Autonomous Systems [Internet]. 2017;88:71–8. Available from: http://dx.doi.org/10.1016/j.robot.2016.11.014
- 20. Asvadi A, Garrote L, Premebida C, Peixoto P, Nunes UJ. Multimodal Vehicle Detection: Fusing 3D-LIDAR and Colour Camera Data. Pattern Recognition Letters [Internet]. 2018;115:20–19. Available from: https://doi.org/10.1016/j.patrec.2017.09.038
- 21. Ashokkumar P, Siva Shankar G, Gautam Srivastava, Praveen Kumar Reddy Maddikunta and TRG. A Two-stage Text Feature Selection Algorithm for Improving Text Classification. ACM Trans Asian Low-Resour Lang Inf Process. 2021;20(3):19.
- 22. Xu F, Chen L, Lou J, Id MR. A real-time road detection method based on reorganized lidar data. PLOS ONE. 2019;14(4):1–17. pmid:30990825
- 23. Li Y, Li S, Du H, Chen L, Zhang D. YOLO-ACN: Focusing on Small Target and Occluded Object Detection. IEEE Access. 2020;8:227288–303.
- 24. Guo Y, Bennamoun M, Sohel F, Lu M, Wan J. 3D Object Recognition in Cluttered Scenes with Local Surface Features: A Survey. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2014;36(11):1–20.
- 25. Aouf N, Gray G, Chermak L, Richardson M, Oudyi F. Local feature based automatic target recognition for future 3D active homing seeker missiles. Aerospace Science and Technology [Internet]. 2017;1:1–9. Available from: https://doi.org/10.1016/j.ast.2017.12.011
- 26. Weber T, Hänsch R, Hellwich O. Automatic registration of unordered point clouds acquired by Kinect sensors using an overlap heuristic. ISPRS Journal of Photogrammetry and Remote Sensing [Internet]. 2015;102:96–109. Available from: http://dx.doi.org/10.1016/j.isprsjprs.2014.12.014
- 27. Yamada H, Ahn J, Mozos OM, Iwashita Y. Gait-based person identification using 3D LiDAR and long short-term memory deep networks. Advanced Robotics [Internet]. 2020;34(18):1201–11. Available from: https://doi.org/01691864.2020.1793812
- 28. Lei H, Jiang G, Quan L. Fast Descriptors and Correspondence Propagation for Robust Global Point Cloud Registration. IEEE Transactions on Image Processing. 2015;14(8):1–10.
- 29. Zhu Q, Li Y, Hu H, Wu B. Robust point cloud classification based on multi-level semantic relationships for urban scenes. ISPRS Journal of Photogrammetry and Remote Sensing. 2017;129:86–102.
- 30. Tian Y, Song W, Sun S, Fong S, Zou S. 3D object recognition method with multiple feature extraction from LiDAR point clouds. The Journal of Supercomputing. 2019;(75):4430–4442.
- 31. Zhao X, Sun P, Xu Z, Min H, Yu H. Fusion of 3D LIDAR and Camera Data for Object Detection in Autonomous Vehicle Applications. IEEE Sensors Journal. 2020;20(9):4901–13.
- 32.
Choi J, Ulbrich S, Lichte B, Maurer M. Multi-Target Tracking using a 3D-Lidar Sensor for Autonomous Vehicles. In: IEEE Annual Conference on Intelligent Transportaion Systems (ITSC 2013). 2013. p. 881–6.
- 33. Wang Yu, Cheng Liang, Chen Yanming, Yang wu ML. Building Point Detection from Vehicle-Borne LiDAR Data Based on Voxel Group and Horizontal Hollow Analysis. Remote Sensing. 2016;8(5).
- 34. Donoghue DNM, Watt PJ, Cox NJ, Wilson J. Remote sensing of species mixtures in conifer plantations using LiDAR height and intensity data. Remote Sensing of Environment. 2007;110(4Fhe):509–22.
- 35. Kim W, Tanaka M, Okutomi M, Sasaki Y. Learning-Based Human Segmentation and Velocity Estimation Using Automatic Labeled LiDAR Sequence for Training. IEEE Access. 2020;8:88443–52.
- 36. Kwon S, Park T. Channel-Based Network for Fast Object Detection of. Electronics. 2020;8(7):1–10.
- 37.
Melotti G, Premebida C, Goncalves N. Multimodal Deep-Learning for Object Recognition Combining Camera and LIDAR Data. In: 2020 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC). 2020. p. 177–82.
- 38.
Engels G, Aranjuelo N, Arganda-carreras I, Nieto M, Otaegui O. 3D Object Detection From LiDAR Data Using Distance Dependent. In: 6th International Conference on Vehicle Technology and Intelligent Transport Systems (VEHITS 2020). 2020.
- 39. Thippa Reddy Gadekallu Mamoun Alazab, Kaluri Rajesh, Praveen Kumar Reddy Maddikunta, Sweta Bhattacharya, Kuruva Lakshmanna PM. Hand gesture classification using a novel CNN-crow search algorithm. Complex and Intellignet Systems [Internet]. 2021; Available from:
- 40. Shankar GS, Ashokkumar P, Vinayakumar R, Ghosh U, Mansoor W, Alnumay WS. An Embedded-Based Weighted Feature Selection Algorithm for Classifying Web Document. 2020;2020.
- 41. Bobkov D, Chen S, Jian R, Iqbal MZ, Steinbach E. Noise-resistant Deep Learning for Object Classification in 3D Point Clouds Using a Point Pair Descriptor. IEEE Robotics and Automation Letters. 2018;3766(c):1–8.
- 42. Romlay MRM, Rashid MM, Toha SF. Rainfall-Runoff Model Based on ANN with LM, BR and PSO as Learning Algorithms. International Journal of Recent Technology and Engineering(IJRTE). 2019;8(September):971–9.
- 43. Palffy A, Dong J, Kooij JFP, Gavrila DM. CNN based Road User Detection using the 3D Radar Cube. IEEE Robotics and Automation Letters. 2020;5(2):1263–70.
- 44. Hu Q, Paisitkriangkrai S, Shen C, Hengel A Van Den, Porikli F. Fast Detection of Multiple Objects in Traffic Scenes With a Common Detection Framework. IEEE Transactions on Intelligent Transportation Systems. 2015;17(4):1002–14.
- 45. Id QZ, Hu X, Su Z, Song Z. 3D car-detection based on a Mobile Deep Sensor Fusion Model and real-scene applications. PLOS ONE [Internet]. 2020;15(9):1–18. Available from: http://dx.doi.org/10.1371/journal.pone.0236947
- 46. Kim Y, Member S, Alnujaim I, You S, Jeong BJ. Human Detection Based on Time-Varying Signature on Range-Doppler Diagram Using Deep Neural Networks. IEEE Geoscience and Remote Sensing Letters. 2020;18(3):426–30.
- 47. Lindenbergh R, Rodrı P. Automatic tree parameter extraction by a Mobile LiDAR System in an urban context. PLOS ONE. 2018;13(4):1–23.
- 48. Romlay MRM, Ibrahim AM, Toha SF. LiDAR Point Cloud Object Classification [Internet]. IEEE Dataport. 2021. Available from: https://dx.doi.org/10.21227/f61n-5e90.
- 49. Rashid MM, Romlay MRM, Ferdaus MM. Development of Electronic Rain Gauge System. International Journal of Electronics and Electrical Engineering. 2015;3(4):245–9.
- 50.
Raja Kushaslnagar, Ludi S, Kushalnagar P. Multi-View Platform: An Accessible Live Classroom Viewing Approach for Low Vision Students. In: The Proceedings of the 13th International ACM SIGACCESS Conference on Computers and Accessibility. 2011. p. 267–8.
- 51. Mcilwaine B, Casado MR. JellyNet: The convolutional neural network jellyfish bloom detector. International Journal of Applied Earth Observation and Geoinformation. 2021;97(November 2020).
- 52. Sun W, Zhao H, Jin Z. A Visual based ROI Detection Method for Facial Expression Recognition. Neurocomputing [Internet]. 2018;296:12–22. Available from: https://doi.org/10.1016/j.neucom.2018.03.034