TY - GEN
T1 - Object-aware dictionary learning with deep features
AU - Xie, Yurui
AU - Porikli, Fatih
AU - He, Xuming
N1 - Publisher Copyright:
© Springer International Publishing AG 2017.
PY - 2017
Y1 - 2017
N2 - Visual dictionary learning has the capacity to determine sparse representations of input images in a data-driven manner using over-complete bases. Sparsity allows robustness to distractors and resistance against overfitting, two valuable attributes of a competent classification solution. Its data-driven nature is comparable to deep convolutional neural networks, which elegantly blend global and local information through progressively more specific filter layers with increasingly extending receptive fields. One shortcoming of dictionary learning is that it does not explicitly select and focus on important regions, instead it generates responses on uniform grid of patches or entire image. To address this, we present an Object-aware dictionary learning framework that systematically incorporates region proposals and deep features in order to improve the discriminative power of the combined classifier. Rather than extracting a dictionary from all fixed sized image windows, our methods concentrates on a small set of object candidates, which enables consolidation of semantic information. We formulate this as an optimization problem on a new objective function and propose an iterative solver. Our results on benchmark datasets demonstrate the effectiveness of our method, which is shown to be superior to the stateoftheart dictionary learning and deep learning based image classification approaches.
AB - Visual dictionary learning has the capacity to determine sparse representations of input images in a data-driven manner using over-complete bases. Sparsity allows robustness to distractors and resistance against overfitting, two valuable attributes of a competent classification solution. Its data-driven nature is comparable to deep convolutional neural networks, which elegantly blend global and local information through progressively more specific filter layers with increasingly extending receptive fields. One shortcoming of dictionary learning is that it does not explicitly select and focus on important regions, instead it generates responses on uniform grid of patches or entire image. To address this, we present an Object-aware dictionary learning framework that systematically incorporates region proposals and deep features in order to improve the discriminative power of the combined classifier. Rather than extracting a dictionary from all fixed sized image windows, our methods concentrates on a small set of object candidates, which enables consolidation of semantic information. We formulate this as an optimization problem on a new objective function and propose an iterative solver. Our results on benchmark datasets demonstrate the effectiveness of our method, which is shown to be superior to the stateoftheart dictionary learning and deep learning based image classification approaches.
UR - http://www.scopus.com/inward/record.url?scp=85016180403&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-54184-6_15
DO - 10.1007/978-3-319-54184-6_15
M3 - Conference contribution
SN - 9783319541839
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 237
EP - 253
BT - Computer Vision - ACCV 2016 - 13th Asian Conference on Computer Vision, Revised Selected Papers
A2 - Sato, Yoichi
A2 - Lai, Shang-Hong
A2 - Lepetit, Vincent
A2 - Nishino, Ko
PB - Springer Verlag
T2 - 13th Asian Conference on Computer Vision, ACCV 2016
Y2 - 20 November 2016 through 24 November 2016
ER -