Discrete Continuous Optimization Framework for Simultaneous Clustering and Training in Mixture Models
Abstract
We study a new framework of learning mixture models via automatic clustering called PRESTO, wherein we optimize a joint objective function on the model parameters and the partitioning, with each model tailored to perform well on its specific cluster. In contrast to prior work, we do not assume any generative model for the data. We convert our training problem to a joint parameter estimation cum a subset selection problem, subject to a matroid span constraint. This allows us to reduce our problem into a constrained set function minimization problem, where the underlying objective is monotone and approximately submodular. We then propose a new joint discrete-continuous optimization algorithm that achieves a bounded approximation guarantee for our problem. We show that PRESTO outperforms several alternative methods. Finally, we study PRESTO in the context of resource-efficient deep learning, where we train smaller resource-constrained models on each partition and show that it outperforms existing data partitioning and model pruning/knowledge distillation approaches, which in contrast to PRESTO, require large initial (teacher) models.
Cite
Text
Sangani et al. "Discrete Continuous Optimization Framework for Simultaneous Clustering and Training in Mixture Models." International Conference on Machine Learning, 2023.Markdown
[Sangani et al. "Discrete Continuous Optimization Framework for Simultaneous Clustering and Training in Mixture Models." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/sangani2023icml-discrete/)BibTeX
@inproceedings{sangani2023icml-discrete,
title = {{Discrete Continuous Optimization Framework for Simultaneous Clustering and Training in Mixture Models}},
author = {Sangani, Parth Vipul and Kashettiwar, Arjun Shashank and Chakraborty, Pritish and Gangula, Bhuvan Reddy and S, Durga and Ramakrishnan, Ganesh and Iyer, Rishabh K and De, Abir},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {29950-29970},
volume = {202},
url = {https://mlanthology.org/icml/2023/sangani2023icml-discrete/}
}