Abstract
We propose a family of clustering algorithms based on the maximization of dependence between the input variables and their cluster labels, as expressed by the Hilbert-Schmidt Independence Criterion (HSIC). Under this framework, we unify the geometric, spectral, and statistical dependence views of clustering, and subsume many existing algorithms as special cases (e.g. k-means and spectral clustering). Distinctive to our framework is that kernels can also be applied on the labels, which can endow them with particular structures. We also obtain a perturbation bound on the change in k-means clustering.
Original language | English |
---|---|
Pages | 815-822 |
Number of pages | 8 |
DOIs | |
Publication status | Published - 2007 |
Externally published | Yes |
Event | 24th International Conference on Machine Learning, ICML 2007 - Corvalis, OR, United States Duration: 20 Jun 2007 → 24 Jun 2007 |
Conference
Conference | 24th International Conference on Machine Learning, ICML 2007 |
---|---|
Country/Territory | United States |
City | Corvalis, OR |
Period | 20/06/07 → 24/06/07 |