Cross-Regularization: Adaptive Model Complexity Through Validation Gradients
Abstract
Model regularization requires extensive manual tuning to balance complexity against overfitting. Cross-regularization resolves this tradeoff by computing validation gradients that directly adapt regularization parameters during training. The method splits parameter optimization - training data guides feature learning while validation data shapes complexity controls - converging provably to cross-validation optima with computational cost scaling only in regularization dimension. When implemented through noise injection in neural networks, this approach reveals striking patterns: unexpectedly high noise tolerance and architecture-specific regularization that emerges organically during training. Beyond complexity control, the framework integrates seamlessly with data augmentation and uncertainty calibration while maintaining single-run efficiency through a simple gradient-based approach.
Cite
Text
Brito. "Cross-Regularization: Adaptive Model Complexity Through Validation Gradients." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Brito. "Cross-Regularization: Adaptive Model Complexity Through Validation Gradients." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/brito2025icml-crossregularization/)BibTeX
@inproceedings{brito2025icml-crossregularization,
title = {{Cross-Regularization: Adaptive Model Complexity Through Validation Gradients}},
author = {Brito, Carlos Stein},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {5558-5577},
volume = {267},
url = {https://mlanthology.org/icml/2025/brito2025icml-crossregularization/}
}