26 June 2017 Lidar-based individual tree species classification using convolutional neural network
Author Affiliations +
Terrestrial lidar is commonly used for detailed documentation in the field of forest inventory investigation. Recent improvements of point cloud processing techniques enabled efficient and precise computation of an individual tree shape parameters, such as breast-height diameter, height, and volume. However, tree species are manually specified by skilled workers to date. Previous works for automatic tree species classification mainly focused on aerial or satellite images, and few works have been reported for classification techniques using ground-based sensor data. Several candidate sensors can be considered for classification, such as RGB or multi/hyper spectral cameras. Above all candidates, we use terrestrial lidar because it can obtain high resolution point cloud in the dark forest. We selected bark texture for the classification criteria, since they clearly represent unique characteristics of each tree and do not change their appearance under seasonable variation and aged deterioration. In this paper, we propose a new method for automatic individual tree species classification based on terrestrial lidar using Convolutional Neural Network (CNN). The key component is the creation step of a depth image which well describe the characteristics of each species from a point cloud. We focus on Japanese cedar and cypress which cover the large part of domestic forest. Our experimental results demonstrate the effectiveness of our proposed method.
© (2017) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Tomohiro Mizoguchi, Tomohiro Mizoguchi, Akira Ishii, Akira Ishii, Hiroyuki Nakamura, Hiroyuki Nakamura, Tsuyoshi Inoue, Tsuyoshi Inoue, Hisashi Takamatsu, Hisashi Takamatsu, } "Lidar-based individual tree species classification using convolutional neural network", Proc. SPIE 10332, Videometrics, Range Imaging, and Applications XIV, 103320O (26 June 2017); doi: 10.1117/12.2270123; https://doi.org/10.1117/12.2270123

Back to Top