15 November 2017 Multi-model convolutional extreme learning machine with kernel for RGB-D object recognition
Author Affiliations +
Proceedings Volume 10605, LIDAR Imaging Detection and Target Recognition 2017; 106051Z (2017) https://doi.org/10.1117/12.2292428
Event: LIDAR Imaging Detection and Target Recognition 2017, 2017, Changchun, China
Abstract
With new depth sensing technology such as Kinect providing high quality synchronized RGB and depth images (RGB-D data), learning rich representations efficiently plays an important role in multi-modal recognition task, which is crucial to achieve high generalization performance. To address this problem, in this paper, we propose an effective multi-modal convolutional extreme learning machine with kernel (MMC-KELM) structure, which combines advantages both the power of CNN and fast training of ELM. In this model, CNN uses multiple alternate convolution layers and stochastic pooling layers to effectively abstract high level features from each modality (RGB and depth) separately without adjusting parameters. And then, the shared layer is developed by combining these features from each modality. Finally, the abstracted features are fed to the extreme learning machine with kernel (KELM), which leads to better generalization performance with faster learning speed. Experimental results on Washington RGB-D Object Dataset show that the proposed multiple modality fusion method achieves state-of-the-art performance with much less complexity.
© (2017) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Yunhua Yin, Huifang Li, Xinling Wen, "Multi-model convolutional extreme learning machine with kernel for RGB-D object recognition", Proc. SPIE 10605, LIDAR Imaging Detection and Target Recognition 2017, 106051Z (15 November 2017); doi: 10.1117/12.2292428; https://doi.org/10.1117/12.2292428
PROCEEDINGS
10 PAGES


SHARE
Back to Top