8 May 2018 Landing zone identification for autonomous UAV applications using fused hyperspectral imagery and LIDAR point clouds
Author Affiliations +
Abstract
Multi-modal data fusion for situational awareness is of interest because fusion of data can provide more information than the individual modalities alone. However, many questions remain, including what data is beneficial, what algorithms work the best or are fastest, and where in the processing pipeline should data be fused? In this paper, we explore some of these questions through a processing pipeline designed for multi-modal data fusion in an autonomous UAV landing scenario. In this paper, we assess landing zone identification methods using two data modalities: hyperspectral imagery and LIDAR point clouds. Using hyperspectral image and LIDAR data from two datasets of Maui and a university campus, we assess the accuracies of different landing zone identification methods, compare rule-based and machine learning based classifications, and show that depending on the dataset, fusion does not always increase performance. However, we show that machine learning methods can be used to ascertain the usefulness of individual modalities and their resulting attributes when used to perform classification.
© (2018) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Sarah Lane, Sarah Lane, Zsolt Kira, Zsolt Kira, Ryan James, Ryan James, Domenic Carr, Domenic Carr, Grady Tuell, Grady Tuell, } "Landing zone identification for autonomous UAV applications using fused hyperspectral imagery and LIDAR point clouds", Proc. SPIE 10644, Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XXIV, 106440D (8 May 2018); doi: 10.1117/12.2305136; https://doi.org/10.1117/12.2305136
PROCEEDINGS
12 PAGES


SHARE
Back to Top