Joker: Joint Optimization Framework for Lightweight Kernel Machines
Abstract
Kernel methods are powerful tools for nonlinear learning with well-established theory. The scalability issue has been their long-standing challenge. Despite the existing success, there are two limitations in large-scale kernel methods: (i) The memory overhead is too high for users to afford; (ii) existing efforts mainly focus on kernel ridge regression (KRR), while other models lack study. In this paper, we propose Joker, a joint optimization framework for diverse kernel models, including KRR, logistic regression, and support vector machines. We design a dual block coordinate descent method with trust region (DBCD-TR) and adopt kernel approximation with randomized features, leading to low memory costs and high efficiency in large-scale learning. Experiments show that Joker saves up to 90% memory but achieves comparable training time and performance (or even better) than the state-of-the-art methods.
Cite
Text
Zhang and Lai. "Joker: Joint Optimization Framework for Lightweight Kernel Machines." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Zhang and Lai. "Joker: Joint Optimization Framework for Lightweight Kernel Machines." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/zhang2025icml-joker/)BibTeX
@inproceedings{zhang2025icml-joker,
title = {{Joker: Joint Optimization Framework for Lightweight Kernel Machines}},
author = {Zhang, Junhong and Lai, Zhihui},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {75297-75313},
volume = {267},
url = {https://mlanthology.org/icml/2025/zhang2025icml-joker/}
}