Modified K-Means Algorithm with Local Optimality Guarantees

Abstract

The K-means algorithm is one of the most widely studied clustering algorithms in machine learning. While extensive research has focused on its ability to achieve a globally optimal solution, there still lacks a rigorous analysis of its local optimality guarantees. In this paper, we first present conditions under which the K-means algorithm converges to a locally optimal solution. Based on this, we propose simple modifications to the K-means algorithm which ensure local optimality in both the continuous and discrete sense, with the same computational complexity as the original K-means algorithm. As the dissimilarity measure, we consider a general Bregman divergence, which is an extension of the squared Euclidean distance often used in the K-means algorithm. Numerical experiments confirm that the K-means algorithm does not always find a locally optimal solution in practice, while our proposed methods provide improved locally optimal solutions with reduced clustering loss. Our code is available at https://github.com/lmingyi/LO-K-means.

Cite

Text

Li et al. "Modified K-Means Algorithm with Local Optimality Guarantees." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Li et al. "Modified K-Means Algorithm with Local Optimality Guarantees." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/li2025icml-modified/)

BibTeX

@inproceedings{li2025icml-modified,
  title     = {{Modified K-Means Algorithm with Local Optimality Guarantees}},
  author    = {Li, Mingyi and Metel, Michael R. and Takeda, Akiko},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {35656-35684},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/li2025icml-modified/}
}