Modified K-Means Algorithm with Local Optimality Guarantees
Abstract
The K-means algorithm is one of the most widely studied clustering algorithms in machine learning. While extensive research has focused on its ability to achieve a globally optimal solution, there still lacks a rigorous analysis of its local optimality guarantees. In this paper, we first present conditions under which the K-means algorithm converges to a locally optimal solution. Based on this, we propose simple modifications to the K-means algorithm which ensure local optimality in both the continuous and discrete sense, with the same computational complexity as the original K-means algorithm. As the dissimilarity measure, we consider a general Bregman divergence, which is an extension of the squared Euclidean distance often used in the K-means algorithm. Numerical experiments confirm that the K-means algorithm does not always find a locally optimal solution in practice, while our proposed methods provide improved locally optimal solutions with reduced clustering loss. Our code is available at https://github.com/lmingyi/LO-K-means.
Cite
Text
Li et al. "Modified K-Means Algorithm with Local Optimality Guarantees." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Li et al. "Modified K-Means Algorithm with Local Optimality Guarantees." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/li2025icml-modified/)BibTeX
@inproceedings{li2025icml-modified,
title = {{Modified K-Means Algorithm with Local Optimality Guarantees}},
author = {Li, Mingyi and Metel, Michael R. and Takeda, Akiko},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {35656-35684},
volume = {267},
url = {https://mlanthology.org/icml/2025/li2025icml-modified/}
}