Paper
13 May 2016 Vehicle classification in WAMI imagery using deep network
Meng Yi, Fan Yang, Erik Blasch, Carolyn Sheaff, Kui Liu, Genshe Chen, Haibin Ling
Author Affiliations +
Abstract
Humans have always had a keen interest in understanding activities and the surrounding environment for mobility, communication, and survival. Thanks to recent progress in photography and breakthroughs in aviation, we are now able to capture tens of megapixels of ground imagery, namely Wide Area Motion Imagery (WAMI), at multiple frames per second from unmanned aerial vehicles (UAVs). WAMI serves as a great source for many applications, including security, urban planning and route planning. These applications require fast and accurate image understanding which is time consuming for humans, due to the large data volume and city-scale area coverage. Therefore, automatic processing and understanding of WAMI imagery has been gaining attention in both industry and the research community. This paper focuses on an essential step in WAMI imagery analysis, namely vehicle classification. That is, deciding whether a certain image patch contains a vehicle or not. We collect a set of positive and negative sample image patches, for training and testing the detector. Positive samples are 64 × 64 image patches centered on annotated vehicles. We generate two sets of negative images. The first set is generated from positive images with some location shift. The second set of negative patches is generated from randomly sampled patches. We also discard those patches if a vehicle accidentally locates at the center. Both positive and negative samples are randomly divided into 9000 training images and 3000 testing images. We propose to train a deep convolution network for classifying these patches. The classifier is based on a pre-trained AlexNet Model in the Caffe library, with an adapted loss function for vehicle classification. The performance of our classifier is compared to several traditional image classifier methods using Support Vector Machine (SVM) and Histogram of Oriented Gradient (HOG) features. While the SVM+HOG method achieves an accuracy of 91.2%, the accuracy of our deep network-based classifier reaches 97.9%.
© (2016) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Meng Yi, Fan Yang, Erik Blasch, Carolyn Sheaff, Kui Liu, Genshe Chen, and Haibin Ling "Vehicle classification in WAMI imagery using deep network", Proc. SPIE 9838, Sensors and Systems for Space Applications IX, 98380E (13 May 2016); https://doi.org/10.1117/12.2224916
Lens.org Logo
CITATIONS
Cited by 5 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Analytical research

Cameras

Image analysis

Sensors

Unmanned aerial vehicles

Convolution

Data modeling

RELATED CONTENT

SA-UNet for face anti-spoofing with depth estimation
Proceedings of SPIE (February 16 2022)
UAV imagery analysis: challenges and opportunities
Proceedings of SPIE (May 01 2017)
Reconnaissance Applications Of Image Understanding
Proceedings of SPIE (November 12 1981)
RADIUS: the government viewpoint
Proceedings of SPIE (April 01 1992)

Back to Top