Abstract
We propose a family of clustering algorithms based on the maximization of dependence between the input variables and their cluster labels, as expressed by the Hilbert-Schmidt Independence Criterion (HSIC). Under this framework, we unify the geometric, spectral, and statistical dependence views of clustering, and subsume many existing algorithms as special cases (e.g. k-means and spectral clustering). Distinctive to our framework is that kernels can also be applied on the labels, which can endow them with particular structures. We also obtain a perturbation bound on the change in k-means clustering.
| Original language | English |
|---|---|
| Pages | 815-822 |
| Number of pages | 8 |
| DOIs | |
| Publication status | Published - 2007 |
| Externally published | Yes |
| Event | 24th International Conference on Machine Learning, ICML 2007 - Corvalis, OR, United States Duration: 20 Jun 2007 → 24 Jun 2007 |
Conference
| Conference | 24th International Conference on Machine Learning, ICML 2007 |
|---|---|
| Country/Territory | United States |
| City | Corvalis, OR |
| Period | 20/06/07 → 24/06/07 |