Reweighting Improves Conditional Risk Bounds
Abstract
In this work, we study the weighted empirical risk minimization (weighted ERM) schema, in which an additional data-dependent weight function is incorporated when the empirical risk function is being minimized. We show that under a general ``balanceable" Bernstein condition, one can design a weighted ERM estimator to achieve superior performance in certain sub-regions over the one obtained from standard ERM, and the superiority manifests itself through a data-dependent constant term in the error bound. These sub-regions correspond to large-margin ones in classification settings and low-variance ones in heteroscedastic regression settings, respectively. Our findings are supported by evidence from synthetic data experiments.
Cite
Text
Zhang et al. "Reweighting Improves Conditional Risk Bounds." Transactions on Machine Learning Research, 2025.Markdown
[Zhang et al. "Reweighting Improves Conditional Risk Bounds." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/zhang2025tmlr-reweighting/)BibTeX
@article{zhang2025tmlr-reweighting,
title = {{Reweighting Improves Conditional Risk Bounds}},
author = {Zhang, Yikai and Lin, Jiahe and Li, Fengpei and Zheng, Songzhu and Raj, Anant and Schneider, Anderson and Nevmyvaka, Yuriy},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/zhang2025tmlr-reweighting/}
}