TY - JOUR
T1 - Mixing linear SVMs for nonlinear classification
AU - Fu, Zhouyu
AU - Robles-Kelly, Antonio
AU - Zhou, Jun
PY - 2010/12
Y1 - 2010/12
N2 - In this paper, we address the problem of combining linear support vector machines (SVMs) for classification of large-scale nonlinear datasets. The motivation is to exploit both the efficiency of linear SVMs (LSVMs) in learning and prediction and the power of nonlinear SVMs in classification. To this end, we develop a LSVM mixture model that exploits a divide-and-conquer strategy by partitioning the feature space into subregions of linearly separable datapoints and learning a LSVM for each of these regions. We do this implicitly by deriving a generative model over the joint data and label distributions. Consequently, we can impose priors on the mixing coefficients and do implicit model selection in a top-down manner during the parameter estimation process. This guarantees the sparsity of the learned model. Experimental results show that the proposed method can achieve the efficiency of LSVMs in the prediction phase while still providing a classification performance comparable to nonlinear SVMs.
AB - In this paper, we address the problem of combining linear support vector machines (SVMs) for classification of large-scale nonlinear datasets. The motivation is to exploit both the efficiency of linear SVMs (LSVMs) in learning and prediction and the power of nonlinear SVMs in classification. To this end, we develop a LSVM mixture model that exploits a divide-and-conquer strategy by partitioning the feature space into subregions of linearly separable datapoints and learning a LSVM for each of these regions. We do this implicitly by deriving a generative model over the joint data and label distributions. Consequently, we can impose priors on the mixing coefficients and do implicit model selection in a top-down manner during the parameter estimation process. This guarantees the sparsity of the learned model. Experimental results show that the proposed method can achieve the efficiency of LSVMs in the prediction phase while still providing a classification performance comparable to nonlinear SVMs.
KW - Classification
KW - expectation-maximization algorithm
KW - mixture of experts
KW - model selection
KW - support vector machines
UR - http://www.scopus.com/inward/record.url?scp=78650056777&partnerID=8YFLogxK
U2 - 10.1109/TNN.2010.2080319
DO - 10.1109/TNN.2010.2080319
M3 - Article
SN - 1045-9227
VL - 21
SP - 1963
EP - 1975
JO - IEEE Transactions on Neural Networks
JF - IEEE Transactions on Neural Networks
IS - 12
M1 - 5634127
ER -