Paper
16 April 2014 Visual attention based bag-of-words model for image classification
Qiwei Wang, Shouhong Wan, Lihua Yue, Che Wang
Author Affiliations +
Proceedings Volume 9159, Sixth International Conference on Digital Image Processing (ICDIP 2014); 91591P (2014) https://doi.org/10.1117/12.2064432
Event: Sixth International Conference on Digital Image Processing, 2014, Athens, Greece
Abstract
Bag-of-words is a classical method for image classification. The core problem is how to count the frequency of the visual words and what visual words to select. In this paper, we propose a visual attention based bag-of-words model (VABOW model) for image classification task. The VABOW model utilizes visual attention method to generate a saliency map, and uses the saliency map as a weighted matrix to instruct the statistic process for the frequency of the visual words. On the other hand, the VABOW model combines shape, color and texture cues and uses L1 regularization logistic regression method to select the most relevant and most efficient features. We compare our approach with traditional bag-of-words based method on two datasets, and the result shows that our VABOW model outperforms the state-of-the-art method for image classification.
© (2014) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Qiwei Wang, Shouhong Wan, Lihua Yue, and Che Wang "Visual attention based bag-of-words model for image classification", Proc. SPIE 9159, Sixth International Conference on Digital Image Processing (ICDIP 2014), 91591P (16 April 2014); https://doi.org/10.1117/12.2064432
Lens.org Logo
CITATIONS
Cited by 5 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Visualization

Visual process modeling

Image classification

Databases

Feature selection

Feature extraction

Data modeling

Back to Top