Consistency of Multiple Kernel Clustering

Abstract

Consistency plays an important role in learning theory. However, in multiple kernel clustering (MKC), the consistency of kernel weights has not been sufficiently investigated. In this work, we fill this gap with a non-asymptotic analysis on the consistency of kernel weights of a novel method termed SimpleMKKM. Under the assumptions of the eigenvalue gap, we give an infinity norm bound as $\widetilde{\mathcal{O}}(k/\sqrt{n})$, where $k$ is the number of clusters and $n$ is the number of samples. On this basis, we establish an upper bound for the excess clustering risk. Moreover, we study the difference of the kernel weights learned from $n$ samples and $r$ points sampled without replacement, and derive its upper bound as $\widetilde{\mathcal{O}}(k\cdot\sqrt{1/r-1/n})$. Based on the above results, we propose a novel strategy with Nyström method to enable SimpleMKKM to handle large-scale datasets with a theoretical learning guarantee. Finally, extensive experiments are conducted to verify the theoretical results and the effectiveness of the proposed large-scale strategy.

Cite

Text

Liang et al. "Consistency of Multiple Kernel Clustering." International Conference on Machine Learning, 2023.

Markdown

[Liang et al. "Consistency of Multiple Kernel Clustering." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/liang2023icml-consistency/)

BibTeX

@inproceedings{liang2023icml-consistency,
  title     = {{Consistency of Multiple Kernel Clustering}},
  author    = {Liang, Weixuan and Liu, Xinwang and Liu, Yong and Ma, Chuan and Zhao, Yunping and Liu, Zhe and Zhu, En},
  booktitle = {International Conference on Machine Learning},
  year      = {2023},
  pages     = {20650-20676},
  volume    = {202},
  url       = {https://mlanthology.org/icml/2023/liang2023icml-consistency/}
}