Next Article in Journal
Erratum: Martins, M.S., et al. Wideband and Wide Beam Polyvinylidene Difluoride (PVDF) Acoustic Transducer for Broadband Underwater. Sensors 2019, 19, 3991
Next Article in Special Issue
A Visual and Persuasive Energy Conservation System Based on BIM and IoT Technology
Previous Article in Journal
Mounting Behaviour Recognition for Pigs Based on Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Dual Neural Architecture Combined SqueezeNet with OctConv for LiDAR Data Classification

1
The Higher Educational Key Laboratory for Measuring & Control Technology and Instrumentations of Heilongjiang, Harbin University of Science and Technology, Harbin 150080, China
2
Department of Computer Science, Chubu University, Aichi 487-8501, Japan
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(22), 4927; https://doi.org/10.3390/s19224927
Submission received: 7 October 2019 / Revised: 6 November 2019 / Accepted: 9 November 2019 / Published: 12 November 2019
(This article belongs to the Special Issue Environmental Sensors and Their Applications)

Abstract

:
Light detection and ranging (LiDAR) is a frequently used technique of data acquisition and it is widely used in diverse practical applications. In recent years, deep convolutional neural networks (CNNs) have shown their effectiveness for LiDAR-derived rasterized digital surface models (LiDAR-DSM) data classification. However, many excellent CNNs have too many parameters due to depth and complexity. Meanwhile, traditional CNNs have spatial redundancy because different convolution kernels scan and store information independently. SqueezeNet replaces a part of 3 × 3 convolution kernels in CNNs with 1 × 1 convolution kernels, decomposes the original one convolution layer into two layers, and encapsulates them into a Fire module. This structure can reduce the parameters of network. Octave Convolution (OctConv) pools some feature maps firstly and stores them separately from the feature maps with the original size. It can reduce spatial redundancy by sharing information between the two groups. In this article, in order to improve the accuracy and efficiency of the network simultaneously, Fire modules of SqueezeNet are used to replace the traditional convolution layers in OctConv to form a new dual neural architecture: OctSqueezeNet. Our experiments, conducted using two well-known LiDAR datasets and several classical state-of-the-art classification methods, revealed that our proposed classification approach based on OctSqueezeNet is able to provide competitive advantages in terms of both classification accuracy and computational amount.

1. Introduction

Light detection and ranging (LiDAR) technology is an active remote sensing measurement technology which can obtain ground object information by emitting laser to the target [1]. Compared with optical imaging and infrared remote sensing, LiDAR can obtain high-resolution three-dimensional spatial point clouds, elevation models, and other raster-derived data independent of weather conditions through laser beam [2,3]. The data adopted in this article are LiDAR-derived rasterized digital surface models (LiDAR-DSM) which are obtained by rasterizing the point cloud data acquired by the LiDAR system [4]. There is a lot of research on LiDAR-DSM data classification. Priestnall et al. examined methods for extracting surface features from DSM produced by LiDAR [5]. Song et al. modified the crown shape in spectrum by using DSM [6]. Zhou et al. used minimum description length (MDL) and morphology to recognize the buildings [7]. Zhou combined LiDAR height and intensity data to accurately map urban land cover [8].
LiDAR-DSM mainly includes terrain changes and feature heights of the target area and it is suitable for classification tasks that distinguish targets with different heights. Thus, the accurate classification of DSM plays an important role in distinguishing different land cover categories. The classification task of LiDAR-DSM data is usually based on the classification of pixels, that is, the interpretation process of remote sensing images [9]. For example, Lodha et al. used the support vector machine (SVM) algorithm to classify LiDAR data and obtained higher accuracy and convincing visual results [10]. Sasaki et al. adopted a decision tree analysis to investigate the average height of each land class [11]. Naidoo et al. used an automated Random Forest modelling approach to classify eight common savanna tree species [12]. Khodadadzadeh et al. developed a new efficient strategy for fusion and classification of hyperspectral and LiDAR data to integrate multiple types of features [13]. Ghamisi et al. proposed a joint classification method which is based on extinction profiles (EPs) features and convolutional neural network (CNN) to improve the classification accuracy [14]. Then Ghamisi et al. proposed to extract the spatial and background information of DSM data in an unsupervised manner to obtain higher classification precision [15]. Wang et al. combined morphological profiles (MPs) and CNN to provide more features for LiDAR-DSM classification [16]. He et al. used spatial transformer networks (STN) to identify the best input of CNN for LiDAR classification [17]. Xia et al. used an ensemble classifier to process morphological features which fuses hyperspectral (HS) images and LiDAR-DSM [18]. Ge et al. proposed a new framework for fusion of hyperspectral image and LiDAR data based on the extinction profiles, local binary pattern (LBP), and kernel collaborative representation classification [19].
AlexNet [20] won the ImageNet [21] competition in 2012, making the convolutional neural network (CNN) become the focus again. It is a breakthrough achievement in the history of CNN development. Since then, convolutional neural networks have been widely used in various image processing fields [22,23,24,25,26,27,28,29,30,31,32]. Compared with the traditional machine learning algorithms, the three characteristics of CNN (receptive field, weights sharing, and subsampling) make it learn characteristics from the images directly. Researchers have proposed various CNNs with different structures and obtained good classification results [33,34,35,36]. These excellent CNNs are designed to achieve higher classification accuracy by designing deeper and more complex structures. However, they have relatively many structural parameters and may require large storage space and computing resources.
SqueezeNet is a lightweight structure with a smaller number of structural parameters and less calculations whose structure and classification accuracy satisfy the application requirements as well [37]. SqueezeNet has only 1 × 1 and 3 × 3 convolution kernels and its purpose is not to obtain the best classification accuracy but to simplify the complexity of the network and achieve the classification accuracy similar to a public network. Octave Convolution (OctConv) is mainly used to deal with feature mapping of multiple spatial frequencies and reduce spatial redundancy. It is a single, generic, and plug-and-play convolution unit that can replace ordinary convolution directly and it does not need to adjust the network structure. In addition, the features of adjacent pixels in one image have similarities; however, the convolution kernel in traditional CNNs sweeps each location and stores its own feature description independently. It ignores spatial consistency, so that the feature maps have a large amount of redundancy in the spatial dimension. OctConv scales a portion of the feature maps and then communicates with the unscaled portion to reduce spatial redundancy and enhance feature utilization [38]. In this article, we propose a novel dual neural architecture, OctSqueezeNet, which combines SqueezeNet with OctConv, which improves LiDAR-DSM data classification accuracy with less structural parameter memory.
In this article, OctSqueezeNet was applied to LiDAR data classification. The main contributions are written as follows:
SqueezeNet combined with OctConv was used to LiDAR data classification for the first time, which improved the classification accuracy and reduced the storage space occupied by the network. We replaced the original convolutional layer in OctConv with the Fire modules in SqueezeNet to compensate for the shortcomings of SqueezeNet, which had less extracted feature information due to lots of 1 × 1 convolution kernel.

2. SqueezeNet Design Architecture

The main objective of SqueezeNet is maintaining competitive accuracy with few parameters. To achieve this goal, three main strategies were adopted. Firstly, the 3 × 3 filter was replaced by the 1 × 1 filter for which has fewer parameters. Secondly, we reduced the number of input channels to the 3 × 3 filters. Finally, we performed subsampled operations in the later stages of the network to make the convolution layer with large activation. SqueezeNet drew on the idea of the Inception module [23] to design a Fire module with a squeeze layer and an expand layer. The structure of Fire module is shown in Figure 1. In order to reduce the number of channels for input elements, the squeeze layer used a 1 × 1 convolution kernel to compress the input elements. The expansion layer used the 1 × 1 and 3 × 3 convolution kernels for multi-scale learning and concatenating.
The operation process of the Fire module is shown in Figure 2. The size of the input feature maps is h × w × n. Firstly, the input feature maps pass through the squeeze layer and obtain the output feature maps with a size of h × w × s1. The size of the feature maps is unchanged but the channels reduce from n to s1. The output feature maps of squeeze layer are sent into 1 × 1 and 3 × 3 convolution kernels in expand layer, respectively. Then concatenate the result of convolution. Finally, only the number of channels changes to e1 + e3. In order to enable the output activation of the 1 × 1 and 3 × 3 filters of the extension module to have the same height and width, the boundary zero filling operation with 1 pixel is performed for the input of the 3 × 3 filters in the extension module. Rectified Linear Unit (ReLU) is applied to the activation of squeeze layer and an expand layer. Meanwhile, there is no full-connection layer in SqueezeNet.

3. Octave Convolution

Natural images can be decomposed into low spatial frequency components and high spatial frequency components. Similarly, the feature maps of the convolutional layers can also be decomposed into feature components of different spatial frequencies. The low frequency components are used to describe the structure with the smooth changes, and the high frequency components are used to describe the fine details of the fast changes.
As shown in Figure 3, the components of high and low frequency of the image are written as XH and XL, and their corresponding outputs are YH and YL after the convolution operation, where α ∈ (0,1) represents the ratio of the low frequency channels. The arrows above and below represent that the feature maps with the same frequency self-update their information by convolution operation, and the crossed arrows help to exchange information between the two frequencies by pooling, upsample, and add operations. In the convolution operation, WH responses are both from XH to YH and from XL to YH, that is WH = [WH→H,WL→H]. WH→H is traditional convolution, and the size of input and output images are the same. WLH upsamples the input image first and then performs traditional convolution, while WHL pools the input image first. As shown as Equations (1) and (2), the high and low frequency feature maps are stored in different groups. Sharing information between adjacent locations can reduce the spatial resolution of the low-frequency group and spatial redundancy.
Y H = Y H H + Y L H
Y L = Y L L + Y H L
The specific calculation process is written as Equations (3) and (4).
Y H = f ( X H ; W H H ) + u p s a m p l e   ( f ( X L ; W L H ) , 2 )
Y L = f ( X L ; W L L ) + f ( p o o l ( X H , 2 ) ; W H L )

4. OctSqueezeNet for LiDAR Classification

As shown in Figure 4, we rescaled the feature maps with different resolutions to the same spatial resolution and concatenated their feature channels, forming a dual neural architecture similar to multi-layers called OctSqueezeNet [39] for LiDAR data classification processing.
First, random selection of 20% from the input feature map performed a 2 × 2 maxpooling operation to halve the size to 16 × 16, and the remaining feature maps maintained the original size, 32 × 32. The two parts were sent separately to different Fire modules to obtain their respective output feature maps. The size of feature maps did not change through the Fire module. We repeated the above operation for the 32 × 32 output feature maps. At the same time, 80% of the 16 × 16 output feature maps were sent to a Fire module and a 2 × 2 upsampling operation was performed on the corresponding output to restore the size to 32 × 32; the remaining 20% of these were sent to another Fire module. Feature maps with the same size in different sections were combined to form the dual outputs, then sent them to the average pooling layer to obtain output feature maps with size 16 × 16 and 8 × 8.
As shown in the lower part of Figure 4, the above operation was repeated for these 16 × 16 output feature maps. Meanwhile, 80% of the these 8 × 8 output feature maps were only sent to a Fire module; the remaining 20% were sent to a Fire module and then the outputs the outputs after this Fire module were sent to a 2 × 2 upsampled layer to restore to the size of 16 × 16.
The feature maps with the same size in different parts were combined to obtain 8 × 8 and 16 × 16 and, respectively, sent to the Fire module, shown as the lower part of Figure 4. Then we upsampled the 8 × 8 feature maps to 16 × 16. The 16 × 16 feature maps were merged to obtain a single output and downsampled to 8 × 8. Finally, softmax was used for data classification.

4.1. Adaptive Learning Optimization Algorithm

Adam optimization algorithm is an extension of the stochastic gradient descent (SGD) algorithm. It is a step-optimization algorithm based on the gradient stochastic objective function and low-order moment adaptive estimation based on training data. It can replace the SGD and update the weight of the neural network iteratively based on the training data. The method estimates the different parameters by the first order matrix, mt, and the second order matrix, nt, of the gradient to calculate the adaptive learning rate. The process is shown as Equations (5)–(7):
m t = μ × m t 1 + ( 1 μ ) × g t
n t = V × n t 1 + ( 1 V ) × g t 2
m t = m t 1 μ t
n t = n t 1 v t
The mt′ and nt′ are the correction for mt and nt, which approximate the expected unbiased estimate, have no additional requirements for memory, and can be adjusted dynamically according to the gradient, where - m t / n t + ε forms a dynamic constraint on the learning rate and it has a clear scope.

4.2. Loss and Activate Function

The activate function of the structure in this article is ReLU. As shown as Equation (9), it causes a part of the output of neuron to zero, which results in sparseness of the network and reduces the interdependence of parameters, alleviating the problem of overfitting. The calculation of the whole process saves a lot of time.
g ( x ) = max ( 0 , x )
Because our dataset is a multi-category dataset, softmax is adopted as the final classifier of the network model. Shown in Equation (10), softmax is used as the exponential operation, which can increase the comparison of the large values and small values to improve learning efficiency.
a j L = e Z j L K e Z K L
where Z j L represents the input of the jth neuron of the Lth layer (usually the last layer), a j L represents the output of the jth neuron in the Lth layer, and e represents the natural constant. K e Z K L represents the sum of the inputs of all neurons in the Lth layer. Therefore, as Equation (11), the corresponding loss function is a combination of softmax and cross-entropy loss.
L o s s i = log y i = log e Z j L K e Z K L

5. Experimental Results and Analysis

5.1. Datasets Description

This article conducted experiments on two different datasets, Bayview Park and Recology, to evaluate the performance of the proposed classification method. They are the public datasets of the 2012 IEEE International Remote Sensing Image Convergence Competition and collected in the city of San Francisco, CA, USA. The Bayview Park dataset has a size of 300 × 200 pixels, the spatial resolution of 1.8 meters and marks 7 land classes. The Recology dataset consists of 200 × 250 pixels with a spatial resolution of 1.8 meters and contains 11 land classes. Figure 5 shows the DSM maps and the groundtruth maps, respectively, for Bayview Park and Recology datasets.

5.2. Experimental Set-Up

Experiments used the TensorFlow under windows as the backend, encoded with Keras and Python. We divided the datasets into two parts: Training samples and test samples. The number of training samples (i.e., 400, 500, 600, and 700) was selected randomly and the rest were used as test sets. And we adopted overall accuracy (OA), average accuracy (AA), and Kappa as objective evaluation criteria. In the experiments, SqueezeNet only used convolutions size with 1 × 1 and 3 × 3, the input features of both datasets were 32 × 32 pixels. DSM data were linearly mapped to (−1 1) and the gradient optimization algorithm selected Adam. The kernel function of SVM was set to radial basis function (rbf) and the coefficient of rbf was default to auto. The penalty parameter of the error terms was 100. The initial learning rate for Bayview Park and Recology datasets both were 0.001 in CNN, and they were set to 0.0005 and 0.001, respectively, in OctConv. In SqueezeNet and OctSqueezeNet, the initial learning rates for two datasets were both set to 0.0005. For the two datasets, the ratio of pooling for feature maps in the first input layer and middle layers were all set to 0.2. The last output layer did not perform pooling. In this article, the initial learning rate for Bayview Park and Recology datasets both were 0.001 in CNN, and they were set to 0.0005 and 0.001, respectively, in OctConv; in SqueezeNet and OctSqueezeNet, the initial learning rates for the two datasets were both set to 0.0005.

5.2.1. Bayview Park Dataset

The experiments ran on a 3.2-GHz CPU with a GTX 1060 GPU card. As shown in Table 1, OctSqueezeNet achieved the best results on OA, AA, and Kappa when different numbers of training samples were selected. When 700 samples were selected, the best OA was 95.42%, increasing 1.91%, 2.21%, 6.32%, and 19.17% compared to OctConv, SqueezeNet, CNN, and SVM, respectively. Figure 6 uses the bar chart to show the accuracy of each comparison method when different numbers of samples were selected. It can be seen intuitively that the accuracy of OctSqueezeNet proposed by us was always higher than that of other methods.
Table 2 and Figure 7 show the classification accuracy of each class. OctSqueezeNet had a good effect on classification of different land classes. Figure 8 shows the classification results of different networks for Bayview Park visually through the false color maps. It can be seen from the classification results, as the precision increased, the mis-division area of the feature gradually decreased.

5.2.2. Recology Dataset

As shown in Table 3, OctSqueezeNet also achieved the best results for Recology dataset on OA, AA, and Kappa when selecting different numbers of training samples. When selecting 700 samples, the best OA was 95.91%, increasing 0.63%, 0.97%, 3.22%, and 17.98% compared to OctConv, SqueezeNet, CNN, and SVM, respectively. Figure 9 shows that the accuracy of the five methods grew steadily with the number of samples increasing. The accuracy of OctSqueezeNet was always the highest.
Table 4 and Figure 10 also prove that OctSqueezeNet had better classification effects on different classes. Figure 11 shows the classification results of the land classes and the area of the wrong division was also less and less with the increase of precision.

5.3. Selection of Experimental Parameters

The number of parameters and the model size of the proposed method in this article were compared with the classical methods. In order to adapt to the Bayview Park and Recology datasets, we made some adjustments to the structure of these classic algorithms. Next, the adjustments made in each architecture are described.
(1)
SqueezeNet: The first change was the addition of batch normalization, which was not present in the original architecture. Then dropout was not used in the last convolution layer. Finally, we changed the size of the kernel on the first convolution layer from original 7 × 7 to 3 × 3.
(2)
AlexNet: Its original architecture contained eight weights layers; the first five layers were convolution and the rest of the layers were fully connected. In comparison experiments, the first convolution layer changed the stride from 4 to 1. The output of the last fully connected layer was fed into a 7-way softmax for the Bayview Park dataset and an 11-way softmax for the Recology dataset.
(3)
ResNet-34: Original Resnet-34 had 4 sections which were composed of 3, 4, 6, and 3 identity blocks, respectively. The number of filters was 64, 128, 256, and 512, respectively, in each identity block of the four sections. In our experiments, the kernel of the first convolution layer was changed from 7 to 3. The number of filters was modified to 16, 28, 40, and 52, respectively. The output of the last fully connected layer was set the same as AlexNet.
As shown in Table 5, compared with some classical CNN algorithms, OctSqueezeNet proposed by us had higher precision, and reduced the number of parameters and model size significantly. When 700 training sets were selected, the OA of OctSqueezeNet on the Bayview Park dataset reached 95.42%, increasing 2.33%, 1.67%, and 2.21% compared with AlexNet, ResNet-34, and SqueezeNet, respectively; but the number of parameters of OcSqueezeNet was only 0.32M, which is 29.67M, 0.1M, and 0.92M less than the number of parameters of AlexNet, ResNet-34, and SqueezeNet, respectively. The model size was only 1.38M, which is 113M, 0.45M, and 3.45M less than the model size of AlexNet, ResNet-34, and SqueezeNet, respectively. The best OA of OcSqueezeNet on the Recology dataset was 95.91% and it was 1.64%, 0.86%, and 0.97% higher than AlexNet, ResNet-34, and SqueezeNet. Its number of parameters was the least (0.33M) compared with the other three methods. The number of parameters of AlexNet, ResNet-34, and SqueezeNet was 114.51M, 1.84M, and 4.85M, respectively. The model size of OctSqueezeNet was also the smallest (1.37M) of these four methods, which decreased 113.14M, 0.47M, and 3.48M compare to the model size of AlexNet, ResNet-34, and SqueezeNet.
Additionally, the training and test time are shown in Table 5 where 700 samples were selected as training sets and experiments ran 150 epochs. For Bayview Park data set, the training time of AlexNet, Resnet-34, SqueezeNet and OctSqueezeNet are 212.13 s, 116.35 s, 75.82 s and 108.82 s respectively. The test time of them are 223.12 s, 160.17 s, 65.43 s and 151.07 s respectively. For Recology data set, the training time of the four methods are 222.94 s, 169.94 s, 38.04 s and 119.62 s respectively. The test time of them are 230.52 s, 207.63 s, 44.12 s and 191.54 s respectively. The time of OctSqueezeNet was longer than SqueezeNet. Although the parameters were reduced, the structural branches were more complicated, which affected the transfer time. However, though the time was slower than SqueezeNet itself, it was acceptable compared to the greatly improved accuracy and the reduced number of parameters and model size. OctSqueezeNet was much faster than AlexNet and Resnet-34.

6. Conclusions

This article designed a dual neural architecture OctSqueezeNet to classify LiDAR-DSM data. Using SqueezeNet alone without combining with OctConv, the network parameters were more and the model size was larger. Most importantly, the classification accuracy of the dataset was reduced significantly. Because SqueezeNet replaced the traditional 3 × 3 convolution kernels with a large number of 1 × 1 convolution kernels, which lost more extracted features relatively. The OctConv did not change the size of the convolution kernel. By reducing the size of a part of the feature maps, it was equivalent to expanding the receptive field to obtain more global feature information. The feature maps with the original size were equivalent to extracting more detailed feature information, because the receptive field had no change. This can compensate for the loss of SqueezeNet in feature extraction. Because there were only 1 × 1 and 3 × 3 convolution kernels, it also made parameters of a network fewer and the model size smaller.
Overall, combining SqueezeNet with OctConv gave better classification precision of datasets and less space for the storage model. The experiment results indicate that OctSqueezeNet achieved 95.42% and 95.66%, respectively, in terms of OA on the Bayview Park and Recology datasets when the number of training samples was 700, which were better than other classification methods. At the same time, for the two datasets, the number of parameters of OctSqueezeNet achieved 0.32M and 0.33M, respectively, and the model size of OctSqueezeNet achieved 1.38M and 1.37M, respectively. They were lower than some other methods. The combination of SqueezeNet and OctConv opens a new window for LiDAR data classification by fully extracting its spatial information.

Author Contributions

This article was completed by all authors. A.W. and M.W. designed and implemented the classification algorithm. K.J. and M.C. made an experimental analysis of the algorithm. Y.I. participated in the writing of the paper.

Funding

Supported by National Science Foundation of China (NSFC-61671190), the University Nursing Program for Young Scholars with Creative Talents in Heilongjiang Province (UNPYSCT-2017086) and Fundamental Research Fundation for Universities of Heilongjiang Province (LGYC2018JQ014).

Acknowledgments

The authors would like to thank the support of the laboratory and university.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. You, S.Y.; Hu, J.H.; Neumann, U.; Fox, P. Urban Site Modeling from LiDAR. In Proceedings of the Computational Science and Its Applications—ICCSA, Montreal, QC, Canada, 18–21 May 2003; Volume 2669, pp. 579–588. [Google Scholar]
  2. Hackel, T.; Wegner, J.D.; Schindler, K. Fast Semantic Segmentation of 3D Point Clouds with Strongly Varying Density. In Proceedings of the ISPRS Annals of Photogrammetry Remote Sensing and Spatial Information Science, Prague, Czech Republic, 12–19 July 2016; Volume 3, pp. 177–184. [Google Scholar]
  3. Weidner, U. Digital Surface Models for Building Extraction. In Automatic Extraction of Man-Made Objects from Aerial and Space Images, 2nd Ed.; Gruen, A., Baltsavias, E.P., Henricsson, O., Eds.; Birkhäuser: Basel, Switzerland, 1997; pp. 193–202. [Google Scholar] [CrossRef]
  4. Lo, C.S.; Lin, C. Growth-Competition-Based Stem Diameter and Volume Modeling for Tree-Level Forest Inventory Using Airborne LiDAR Data. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2216–2226. [Google Scholar] [CrossRef]
  5. Priestnall, G.; Jaafar, J.; Duncan, A. Extracting Urban Features from LiDAR Digital Surface Models. Comput. Environ. Urban 2000, 24, 65–78. [Google Scholar] [CrossRef]
  6. Song, Y.; Imanishi, J.; Hashimoto, H.; Morimura, A.; Morimoto, Y.; Kitada, K. Spectral Correction for the Effect of Crown Shape at the Single-Tree Level: An Approach Using a Lidar-Derived Digital Surface Model for Broad-Leaved Canopy. IEEE Trans. Geosci. Remote Sens. 2013, 21, 755–764. [Google Scholar] [CrossRef]
  7. Zhou, S.; Mi, L.; Chen, H. Building Detection in Digital Surface Model. In Proceedings of the IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 22–23 October 2013; pp. 194–199. [Google Scholar] [CrossRef]
  8. Zhou, W.Q. An Object-Based Approach for Urban Land Cover Classification: Integrating LiDAR Height and Intensity Data. IEEE Geosci. Remote Sens. Lett. 2013, 8, 928–931. [Google Scholar] [CrossRef]
  9. Liu, Y.; Ren, Y.; Hu, L.; Liu, Z. Study on Highway Geological Disasters Knowledge base for Remote Sensing Images Interpretation. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012. [Google Scholar] [CrossRef]
  10. Lodha, S.K.; Kreps, E.J.; Helmbold, D.P.; Fitzpatrick, D. Aerial LiDAR Data Classification Using Support Vector Machines (SVM). In Proceedings of the 3rd International Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT’06), Chapel Hill, NC, USA, 14–16 June 2006; pp. 567–574. [Google Scholar] [CrossRef]
  11. Sasaki, T.; Imanishi, J.; Ioki, K.; Morimoto, Y.; Kitada, K. Object-based Classification of Land Cover and Tree Species by Integrating Airborne LiDAR and High Spatial Resolution Imagery Data. Landsc. Ecol. Eng. 2012, 8, 157–171. [Google Scholar] [CrossRef]
  12. Naidoo, L.; Cho, M.A.; Mathieu, R.; Asner, G. Classification of Savanna Tree Species, in the Greater Kruger National Park Region, by Integrating Hyperspectral and LiDAR Data in a Random Forest Data Mining Environment. ISPRS J. Photogramm. Remote Sens. 2012, 69, 167–179. [Google Scholar] [CrossRef]
  13. Khodadadzadeh, M.; Li, J.; Prasad, S.; Plaza, A. Fusion of Hyperspectral and LiDAR Remote Sensing Data Using Multiple Feature Learning. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2971–2983. [Google Scholar] [CrossRef]
  14. Ghamisi, P.; Höfle, B.; Zhu, X.X. Hyperspectral and LiDAR Data Fusion Using Extinction Profiles and Deep Convolutional Neural Network. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2017, 10, 3011–3024. [Google Scholar] [CrossRef]
  15. Ghamisi, P.; Höfle, B. LiDAR Data Classification Using Extinction Profiles and a Composite Kernel Support Vector Machine. IEEE Geosci. Remote Sens. Lett. 2017, 14, 659–663. [Google Scholar] [CrossRef]
  16. Wang, A.L.; He, X.; Ghamisi, P.; Chen, Y.S. LiDAR Data Classification Using Morphological Profiles and Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 774–778. [Google Scholar] [CrossRef]
  17. He, X.; Wang, A.L.; Ghamisi, P.; Li, G.; Chen, Y.S. LiDAR Data Classification Using Spatial Transformation and CNN. IEEE Geosci. Remote Sens. Lett. 2018, 16, 125–129. [Google Scholar] [CrossRef]
  18. Xia, J.S.; Yokoya, N.; Iwasaki, A. Fusion of Hyperspectral and LiDAR Data with a Novel Ensemble Classifier. IEEE Geosci. Remote Sens. Lett. 2018, 15, 957–961. [Google Scholar] [CrossRef]
  19. Ge, C.; Du, Q.; Li, W.; Li, Y.S.; Sun, W.W. Hyperspectral and LiDAR Data Classification Using Kernel Collaborative Representation Based Residual Fusion. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2019, 12, 1963–1973. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  21. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.F. ImageNet: A Large-scale Hierarchical Image Database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  22. Ito, S.; Hiratsuka, S.; Ohta, M.; Matsubara, H.; Ogawa, M. Small Imaging Depth LIDAR and DCNN-Based Localization for Automated Guided Vehicle. Sensors 2018, 18, 177. [Google Scholar] [CrossRef] [PubMed]
  23. Kwon, S.K.; Jung, H.S.; Baek, W.K.; Kim, D. Classification of Forest Vertical Structure in South Korea from Aerial Orthophoto and Lidar Data Using an Artificial Neural Network. Appl. Sci. 2017, 7, 1046. [Google Scholar] [CrossRef] [Green Version]
  24. Shao, J.; Qu, C.; Li, J.; Peng, S. A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification. Sensors 2018, 18, 3039. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Gao, F.; Huang, T.; Wang, J.; Sun, J.; Hussain, A.; Yang, E. Dual-Branch Deep Convolution Neural Network for Polarimetric SAR Image Classification. Appl. Sci. 2017, 7, 447. [Google Scholar] [CrossRef] [Green Version]
  26. Gao, Q.; Lim, S.; Jia, X. Hyperspectral Image Classification Using Convolutional Neural Networks and Multiple Feature Learning. Remote Sens. 2018, 10, 299. [Google Scholar] [CrossRef] [Green Version]
  27. Zhang, X.Y.; Shi, H.C.; Zhu, X.B.; Li, P. Active semi-supervised learning based on self-expressive correlation with generative adversarial networks. Neurocomputing 2019, 345, 103–113. [Google Scholar] [CrossRef]
  28. Zhang, X.Y.; Shi, H.C.; Li, C.S.; Zheng, K.; Zhu, X.B.; Li, D. Learning transferable self-attentive representations for action recognition in untrimmed videos with weak supervision. In Proceedings of the 33th AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 1–8. [Google Scholar]
  29. Zhu, X.B.; Li, Z.Z.; Zhang, X.Y.; Li, P.; Xue, Z.; Wang, L. Deep convolutional representations and kernel extreme learning machines for image classification. Multimed. Tools Appl. 2018, 78, 29271–29290. [Google Scholar] [CrossRef]
  30. Jiang, Y.G.; Wu, Z.X.; Tang, J.H.; Li, Z.C.; Xue, X.Y.; Chang, S.H. Modeling Multimodal Clues in a Hybrid Deep Learning Framework for Video Classification. IEEE Trans. Multimedia 2018, 20, 3137–3147. [Google Scholar] [CrossRef] [Green Version]
  31. Jiang, Y.G.; Wu, Z.X.; Wang, J.; Xue, X.Y.; Chang, S.H. Exploiting Feature and Class Relationships in Video Categorization with Regularized Deep Neural Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 352–364. [Google Scholar] [CrossRef] [PubMed]
  32. Yang, P.; Zhao, P.; Gao, X.; Liu, Y. Robust cost-sensitive learning for recommendation with implicit feedback. In Proceedings of the SIAM International Conference on Data Mining, San Diego, CA, USA, 3–5 May 2018; pp. 621–629. [Google Scholar]
  33. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference of Learning Representation, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  34. Szegedy, C.; Liu, W.; Jia, Y.Q.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 1, pp. 770–778. [Google Scholar]
  36. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  37. SqueezeNet: Alex-Net-level Accuracy with 50x Fewer Parameters and <0.5 MB Model Size. Available online: https://arxiv.org/abs/1602.07360 (accessed on 4 September 2019).
  38. Drop an Octave: Reducing Spatial Redundancy in Convolutional Neural Networks with Octave Convolution. Available online: https://export.arxiv.org/abs/1904.05049 (accessed on 4 September 2019).
  39. Ke, T.W.; Maire, M.; Yu, S.X. Multigrid Neural Architectures. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1063–6919. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Structure of Fire module.
Figure 1. Structure of Fire module.
Sensors 19 04927 g001
Figure 2. Operation process of Fire module.
Figure 2. Operation process of Fire module.
Sensors 19 04927 g002
Figure 3. Structure of OctConv.
Figure 3. Structure of OctConv.
Sensors 19 04927 g003
Figure 4. Architecture of OctSqueezeNet.
Figure 4. Architecture of OctSqueezeNet.
Sensors 19 04927 g004
Figure 5. Bayview Park (up) and Recology (down) datasets: (a) Digital surface models (DSM) of Bayview Park, (b) groundtruth of Bayview Park, (c) DSM of Recology, (d) groundtruth of Recology.
Figure 5. Bayview Park (up) and Recology (down) datasets: (a) Digital surface models (DSM) of Bayview Park, (b) groundtruth of Bayview Park, (c) DSM of Recology, (d) groundtruth of Recology.
Sensors 19 04927 g005
Figure 6. The classification results of different methods on Bayview Park.
Figure 6. The classification results of different methods on Bayview Park.
Sensors 19 04927 g006
Figure 7. Classification results of different methods for each class on Bayview Park.
Figure 7. Classification results of different methods for each class on Bayview Park.
Sensors 19 04927 g007
Figure 8. Classification results on Bayview Park: (a) Groundtruth map, (b) support vector machine (SVM), (c) convolutional neural network (CNN), (d) SqueezeNet, (e) Octave Convolution (OctConv), and (f) OctSqueezeNet.
Figure 8. Classification results on Bayview Park: (a) Groundtruth map, (b) support vector machine (SVM), (c) convolutional neural network (CNN), (d) SqueezeNet, (e) Octave Convolution (OctConv), and (f) OctSqueezeNet.
Sensors 19 04927 g008
Figure 9. The classification results of different methods on Recology.
Figure 9. The classification results of different methods on Recology.
Sensors 19 04927 g009
Figure 10. Classification results of different methods for each class on Recology.
Figure 10. Classification results of different methods for each class on Recology.
Sensors 19 04927 g010
Figure 11. Classification results on Recology: (a) Groundtruth map, (b) SVM, (c) CNN, (d) SqueezeNet, (e) OctConv, and (f) OctSqueezeNet.
Figure 11. Classification results on Recology: (a) Groundtruth map, (b) SVM, (c) CNN, (d) SqueezeNet, (e) OctConv, and (f) OctSqueezeNet.
Sensors 19 04927 g011
Table 1. Classification results on Bayview Park dataset.
Table 1. Classification results on Bayview Park dataset.
Training of Different SamplesIndexSVMCNNSqueezeNetOctConvOctSqueezeNet
400OA%72.11 ± 2.0286.77 ± 2.1188.60 ± 1.5691.32 ± 0.4191.99 ± 0.81
AA%77.08 ± 1.4388.01 ± 1.3289.04 ± 0.4793.02 ± 0.3293.21 ± 0.43
K×10063.11 ± 1.8982.92 ± 1.7785.42 ± 2.0188.30 ± 0.6689.48 ± 1.00
500OA%73.75 ± 2.4087.37 ± 1.0791.31 ± 2.8191.61 ± 0.6392.79 ± 0.41
AA%78.45 ± 2.0789.21 ± 2.9990.59 ± 2.0593.37 ± 1.2195.02 ± 0.90
K×10065.41 ± 0.8484.62 ± 1.5288.44 ± 1.6588.47 ± 1.4390.48 ± 0.47
600OA%74.72 ± 2.0388.27 ± 0.7191.32 ± 1.0392.32 ± 0.4994.09 ± 1.23
AA%78.87 ± 1.1089.44 ± 2.6492.11 ± 1.5794.30 ± 0.5195.75 ± 1.25
K×10066.74 ± 1.8885.88 ± 2.0189.63 ± 1.4389.89 ± 0.7292.23 ± 1.64
700OA%76.25 ± 0.7489.10 ± 2.0993.21 ± 0.6393.51 ± 0.7795.42 ± 0.91
AA%81.21 ± 2.2590.11 ± 0.8893.33 ± 0.7995.64 ± 1.5696.43 ± 1.37
K×10068.70 ± 2.1486.52 ± 2.7091.11 ± 2.1491.38 ± 1.0493.99 ± 1.97
Table 2. Classification results of each class on Bayview Park dataset.
Table 2. Classification results of each class on Bayview Park dataset.
NO.ClassesSVMCNNSqueezeNetOctConvOctSqueezeNet
1 Sensors 19 04927 i00182.00 ± 4.2193.78 ± 1.7694.89 ± 1.3399.40 ± 0.4199.52 ± 0.09
2 Sensors 19 04927 i00283.88 ± 3.3292.71 ± 1.0498.16 ± 1.8498.08 ± 1.5399.93 ± 0.07
3 Sensors 19 04927 i00391.01 ± 5.2092.53 ± 1.8296.12 ± 1.1498.84 ± 1.3699.54 ± 0.46
4 Sensors 19 04927 i00481.70 ± 4.3191.05 ± 1.2492.37 ± 0.3396.29 ± 0.8896.29 ± 2.77
5 Sensors 19 04927 i00583.85 ± 1.9986.03 ± 1.8197.45 ± 3.1195.54 ± 0.6798.67 ± 0.88
6 Sensors 19 04927 i00660.53 ± 3.3985.37 ± 1.7986.29 ± 3.0283.51 ± 0.5388.75 ± 3.26
7 Sensors 19 04927 i00785.53 ± 2.7490.98 ± 2.6492.25 ± 1.9897.01 ± 2.5192.47 ± 2.30
Table 3. Classification results on Recology dataset.
Table 3. Classification results on Recology dataset.
Training of Different SamplesIndexSVMCNNSqueezeNetOctConvOctSqueezeNet
400OA%71.86 ± 1.0585.57 ± 1.3790.39 ± 0.6592.88 ± 1.0793.93 ± 0.61
AA%72.69 ± 2.4388.97 ± 2.1389.02 ± 0.1891.17 ± 1.5693.63 ± 0.17
K×10066.62 ± 0.8882.35 ± 1.7188.67 ± 3.2191.28 ± 0.7392.79 ± 0.74
500OA%74.51 ± 0.7787.63 ± 1.4392.54 ± 2.0193.24 ± 1.1494.77 ± 0.83
AA%77.64 ± 3.2691.06 ± 0.3991.28 ± 1.2191.41 ± 0.5793.72 ± 0.60
K×10069.73 ± 1.7585.98 ± 0.2987.30 ± 1.4392.06 ± 1.2093.79 ± 0.99
600OA%76.35 ± 2.3390.12 ± 0.5293.13 ± 0.7194.53 ± 1.2495.07 ± 0.48
AA%79.51 ± 1.3188.77 ± 1.2292.38 ± 1.7294.02 ± 0.6295.36 ± 1.15
K×10071.85 ± 2.0186.89 ± 0.7291.94 ± 2.1393.51 ± 1.4494.13 ± 0.63
700OA%77.93 ± 0.9392.69 ± 1.6994.94 ± 2.6395.28 ± 1.3895.91 ± 0.73
AA%81.00 ± 1.3292.22 ± 1.4794.29 ± 0.7994.82 ± 1.1695.89 ± 0.17
K×10073.70 ± 1.0189.53 ± 2.1893.19 ± 2.1494.40 ± 1.0995.13 ± 0.11
Table 4. Classification results of each class on Recology dataset.
Table 4. Classification results of each class on Recology dataset.
NO.ClassesSVMCNNSqueezeNetOctConvOctSqueezeNet
1 Sensors 19 04927 i00871.66 ± 0.9498.42 ± 1.0996.87 ± 1.2096.05 ± 1.0199.06 ± 0.94
2 Sensors 19 04927 i00965.04 ± 0.7495.53 ± 1.2198.89 ± 1.1098.67 ± 1.6999.56 ± 0.44
3 Sensors 19 04927 i01093.04 ± 1.0494.09 ± 1.1298.38 ± 1.9999.11 ± 0.5498.12 ± 1.43
4 Sensors 19 04927 i01189.54 ± 2.3397.25 ± 1.1395.48 ± 1.2598.80 ± 0.5899.55 ± 0.45
5 Sensors 19 04927 i01286.08 ± 1.2096.01 ± 1.9596.72 ± 2.1699.59 ± 0.41100
6 Sensors 19 04927 i01369.11 ± 1.1294.87 ± 1.4195.54 ± 1.4296.21 ± 0.1195.35 ± 1.81
7 Sensors 19 04927 i01487.94 ± 3.1196.83 ± 1.7995.87 ± 2.0197.74 ± 2.4696.94 ± 1.64
8 Sensors 19 04927 i01586.66 ± 1.1195.36 ± 0.6197.64 ± 0.6197.36 ± 0.3497.54 ± 2.01
9 Sensors 19 04927 i01663.61 ± 2.1277.14 ± 0.7989.99 ± 1.6486.99 ± 1.2587.48 ± 1.27
10 Sensors 19 04927 i01780.31 ± 4.3472.62 ±1.4678.53 ± 2.3382.28 ± 2.1289.68 ± 0.53
11 Sensors 19 04927 i01898.09 ± 1.8796.31 ± 1.2793.31 ± 0.4894.09 ± 1.2691.68 ± 1.61
Table 5. Performance comparison of 700 samples for different networks.
Table 5. Performance comparison of 700 samples for different networks.
DataMethodParams
(Million)
OA
(%)
Model Size
(M)
Training
Time (s)
Test
Time (s)
Bayview ParkAlexNet29.9993.09 ± 0.66114.38212.13223.12
ResNet-340.4293.75 ± 1.291.83125.15160.17
SqueezeNet1.2493.21 ± 0.634.8375.8265.43
OctSqueezeNet0.3295.42 ± 0.911.38108.82151.07
RecologyAlexNet29.9994.27 ± 0.74114.51222.94230.52
ResNet-340.4195.05 ± 0.311.84169.94207.63
SqueezeNet1.2594.94 ± 2.634.8538.0444.12
OctSqueezeNet0.3395.91 ± 0.731.37119.62191.54

Share and Cite

MDPI and ACS Style

Wang, A.; Wang, M.; Jiang, K.; Cao, M.; Iwahori, Y. A Dual Neural Architecture Combined SqueezeNet with OctConv for LiDAR Data Classification. Sensors 2019, 19, 4927. https://doi.org/10.3390/s19224927

AMA Style

Wang A, Wang M, Jiang K, Cao M, Iwahori Y. A Dual Neural Architecture Combined SqueezeNet with OctConv for LiDAR Data Classification. Sensors. 2019; 19(22):4927. https://doi.org/10.3390/s19224927

Chicago/Turabian Style

Wang, Aili, Minhui Wang, Kaiyuan Jiang, Mengqing Cao, and Yuji Iwahori. 2019. "A Dual Neural Architecture Combined SqueezeNet with OctConv for LiDAR Data Classification" Sensors 19, no. 22: 4927. https://doi.org/10.3390/s19224927

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop