Smoothing multivariate performance measures

Xinhua Zhang*, Ankan Saha, S. V.N. Vishwanathan

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

31 Citations (Scopus)

Abstract

Optimizing multivariate performance measure is an important task in Machine Learning. Joachims (2005) introduced a Support Vector Method whose underlying optimization problem is commonly solved by cutting plane methods (CPMs) such as SVM-Perf and BMRM. It can be shown that CPMs converge to an e accurate solution iterations, where l is the trade-off parameter between the regularizer and the loss function. Motivated by the impressive convergence rate of CPM on a number of practical problems, it was conjectured that these rates can be further improved. We disprove this conjecture in this paper by constructing counter examples. However, surprisingly, we further discover that these problems are not inherently hard, and we develop a novel smoothing strategy, which in conjunction with Nesterov's accelerated gradient method, can find an ε accurate solution iterations. Computationally, our smoothing technicue is also particularly advantageous for optimizing multivariate performance scores such as precision/recall break-even point and ROCArea; the cost per iteration remains the same as that of CPMs. Empirical evaluation on some of the largest publicly available data sets shows that our method converges significantly faster than CPMs without sacrificing generalization ability.

Original languageEnglish
Pages (from-to)3623-3680
Number of pages58
JournalJournal of Machine Learning Research
Volume13
Publication statusPublished - Dec 2012
Externally publishedYes

Fingerprint

Dive into the research topics of 'Smoothing multivariate performance measures'. Together they form a unique fingerprint.

Cite this