Bias/Variance Is Not the Same as Approximation/Estimation
Abstract
We study the relation between two classical results: the bias-variance decomposition, and the approximation-estimation decomposition. Both are important conceptual tools in Machine Learning, helping us describe the nature of model fitting. It is commonly stated that they are “closely related”, or “similar in spirit”. However, sometimes it is said they are equivalent. In fact they are different, but have subtle connections cutting across learning theory, classical statistics, and information geometry, that (very surprisingly) have not been previously observed. We present several results for losses expressible as Bregman divergences: a broad family with a known bias-variance decomposition. Discussion and future directions are presented for more general losses, including the 0/1 classification loss.
Cite
Text
Brown and Ali. "Bias/Variance Is Not the Same as Approximation/Estimation." Transactions on Machine Learning Research, 2024.Markdown
[Brown and Ali. "Bias/Variance Is Not the Same as Approximation/Estimation." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/brown2024tmlr-bias/)BibTeX
@article{brown2024tmlr-bias,
title = {{Bias/Variance Is Not the Same as Approximation/Estimation}},
author = {Brown, Gavin and Ali, Riccardo},
journal = {Transactions on Machine Learning Research},
year = {2024},
url = {https://mlanthology.org/tmlr/2024/brown2024tmlr-bias/}
}