TY - GEN
T1 - Kernel Support Vector Machines and Convolutional Neural Networks
AU - Jiang, Shihao
AU - Hartley, Richard
AU - Fernando, Basura
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2019/1/16
Y1 - 2019/1/16
N2 - Convolutional Neural Networks (CNN) have achieved great success in various computer vision tasks due to their strong ability in feature extraction. The trend of development of CNN architectures is to increase their depth so as to increase their feature extraction ability. Kernel Support Vector Machines (SVM), on the other hand, are known to give optimal separating surfaces by their ability to automatically select support vectors and perform classification in higher dimensional spaces. We investigate the idea of combining the two such that best of both worlds can be achieved and a more compact model can perform as well as deeper CNNs. In the past, attempts have been made to use CNNs to extract features from images and then classify with a kernel SVM, but this process was performed in two separate steps. In this paper, we propose one single model where a CNN and a kernel SVM are integrated together and can be trained end-to-end. In particular, we propose a fully-differentiable Radial Basis Function (RBF) layer, where it can be seamless adapted to a CNN environment and forms a better classifier compared to the normal linear classifier. Due to end-to-end training, our approach allows the initial layers of the CNN to extract features more adapted to the kernel SVM classifier. Our experiments demonstrate that the hybrid CNN-kSVM model gives superior results to a plain CNN model, and also performs better than the method where feature extraction and classification are performed in separate stages, by a CNN and a kernel SVM respectively.
AB - Convolutional Neural Networks (CNN) have achieved great success in various computer vision tasks due to their strong ability in feature extraction. The trend of development of CNN architectures is to increase their depth so as to increase their feature extraction ability. Kernel Support Vector Machines (SVM), on the other hand, are known to give optimal separating surfaces by their ability to automatically select support vectors and perform classification in higher dimensional spaces. We investigate the idea of combining the two such that best of both worlds can be achieved and a more compact model can perform as well as deeper CNNs. In the past, attempts have been made to use CNNs to extract features from images and then classify with a kernel SVM, but this process was performed in two separate steps. In this paper, we propose one single model where a CNN and a kernel SVM are integrated together and can be trained end-to-end. In particular, we propose a fully-differentiable Radial Basis Function (RBF) layer, where it can be seamless adapted to a CNN environment and forms a better classifier compared to the normal linear classifier. Due to end-to-end training, our approach allows the initial layers of the CNN to extract features more adapted to the kernel SVM classifier. Our experiments demonstrate that the hybrid CNN-kSVM model gives superior results to a plain CNN model, and also performs better than the method where feature extraction and classification are performed in separate stages, by a CNN and a kernel SVM respectively.
UR - http://www.scopus.com/inward/record.url?scp=85062240046&partnerID=8YFLogxK
U2 - 10.1109/DICTA.2018.8615840
DO - 10.1109/DICTA.2018.8615840
M3 - Conference contribution
T3 - 2018 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2018
BT - 2018 International Conference on Digital Image Computing
A2 - Pickering, Mark
A2 - Zheng, Lihong
A2 - You, Shaodi
A2 - Rahman, Ashfaqur
A2 - Murshed, Manzur
A2 - Asikuzzaman, Md
A2 - Natu, Ambarish
A2 - Robles-Kelly, Antonio
A2 - Paul, Manoranjan
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2018 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2018
Y2 - 10 December 2018 through 13 December 2018
ER -