16 April 2014 Visual attention based bag-of-words model for image classification
Author Affiliations +
Proceedings Volume 9159, Sixth International Conference on Digital Image Processing (ICDIP 2014); 91591P (2014) https://doi.org/10.1117/12.2064432
Event: Sixth International Conference on Digital Image Processing, 2014, Athens, Greece
Abstract
Bag-of-words is a classical method for image classification. The core problem is how to count the frequency of the visual words and what visual words to select. In this paper, we propose a visual attention based bag-of-words model (VABOW model) for image classification task. The VABOW model utilizes visual attention method to generate a saliency map, and uses the saliency map as a weighted matrix to instruct the statistic process for the frequency of the visual words. On the other hand, the VABOW model combines shape, color and texture cues and uses L1 regularization logistic regression method to select the most relevant and most efficient features. We compare our approach with traditional bag-of-words based method on two datasets, and the result shows that our VABOW model outperforms the state-of-the-art method for image classification.
© (2014) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Qiwei Wang, Qiwei Wang, Shouhong Wan, Shouhong Wan, Lihua Yue, Lihua Yue, Che Wang, Che Wang, } "Visual attention based bag-of-words model for image classification", Proc. SPIE 9159, Sixth International Conference on Digital Image Processing (ICDIP 2014), 91591P (16 April 2014); doi: 10.1117/12.2064432; https://doi.org/10.1117/12.2064432
PROCEEDINGS
7 PAGES


SHARE
Back to Top