ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences (Aug 2020)

EVALUATING A CONVOLUTIONAL NEURAL NETWORK FOR FEATURE EXTRACTION AND TREE SPECIES CLASSIFICATION USING UAV-HYPERSPECTRAL IMAGES

  • C. Sothe,
  • L. E. C. la Rosa,
  • C. M. de Almeida,
  • A. Gonsamo,
  • M. B. Schimalski,
  • J. D. B. Castro,
  • R. Q. Feitosa,
  • M. Dalponte,
  • C. L. Lima,
  • V. Liesenberg,
  • G. T. Miyoshi,
  • A. M. G. Tommaselli

DOI
https://doi.org/10.5194/isprs-annals-V-3-2020-193-2020
Journal volume & issue
Vol. V-3-2020
pp. 193 – 199

Abstract

Read online

The classification of tree species can significantly benefit from high spatial and spectral information acquired by unmanned aerial vehicles (UAVs) associated with advanced feature extraction and classification methods. Different from the traditional feature extraction methods, that highly depend on user’s knowledge, the convolutional neural network (CNN)-based method can automatically learn and extract the spatial-related features layer by layer. However, in order to capture significant features of the data, the CNN classifier requires a large number of training samples, which are hardly available when dealing with tree species in tropical forests. This study investigated the following topics concerning the classification of 14 tree species in a subtropical forest area of Southern Brazil: i) the performance of the CNN method associated with a previous step to increase and balance the sample set (data augmentation) for tree species classification as compared to the conventional machine learning methods support vector machine (SVM) and random forest (RF) using the original training data; ii) the performance of the SVM and RF classifiers when associated with a data augmentation step and spatial features extracted from a CNN. Results showed that the CNN classifier outperformed the conventional SVM and RF classifiers, reaching an overall accuracy (OA) of 84.37% and Kappa of 0.82. The SVM and RF had a poor accuracy with the original spectral bands (OA 62.67% and 59.24%) but presented an increase between 14% and 21% in OA when associated with a data augmentation and spatial features extracted from a CNN.