Cross-Regularization: Adaptive Model Complexity Through Validation Gradients

Abstract

Model regularization requires extensive manual tuning to balance complexity against overfitting. Cross-regularization resolves this tradeoff by computing validation gradients that directly adapt regularization parameters during training. The method splits parameter optimization - training data guides feature learning while validation data shapes complexity controls - converging provably to cross-validation optima with computational cost scaling only in regularization dimension. When implemented through noise injection in neural networks, this approach reveals striking patterns: unexpectedly high noise tolerance and architecture-specific regularization that emerges organically during training. Beyond complexity control, the framework integrates seamlessly with data augmentation and uncertainty calibration while maintaining single-run efficiency through a simple gradient-based approach.

Cite

Text

Brito. "Cross-Regularization: Adaptive Model Complexity Through Validation Gradients." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Brito. "Cross-Regularization: Adaptive Model Complexity Through Validation Gradients." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/brito2025icml-crossregularization/)

BibTeX

@inproceedings{brito2025icml-crossregularization,
  title     = {{Cross-Regularization: Adaptive Model Complexity Through Validation Gradients}},
  author    = {Brito, Carlos Stein},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {5558-5577},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/brito2025icml-crossregularization/}
}