Learning to Optimize with Stochastic Dominance Constraints
Abstract
In real-world decision-making, uncertainty is important yet difficult to handle. Stochastic dominance provides a theoretically sound approach to comparing uncertain quantities, but optimization with stochastic dominance constraints is often computationally expensive, which limits practical applicability. In this paper, we develop a simple yet efficient approach for the problem, Light Stochastic Dominance Solver (light-SD), by leveraging properties of the Lagrangian. We recast the inner optimization in the Lagrangian as a learning problem for surrogate approximation, which bypasses the intractability and leads to tractable updates or even closed-form solutions for gradient calculations. We prove convergence of the algorithm and test it empirically. The proposed light-SD demonstrates superior performance on several representative problems ranging from finance to supply chain management.
Cite
Text
Dai et al. "Learning to Optimize with Stochastic Dominance Constraints." Artificial Intelligence and Statistics, 2023.Markdown
[Dai et al. "Learning to Optimize with Stochastic Dominance Constraints." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/dai2023aistats-learning/)BibTeX
@inproceedings{dai2023aistats-learning,
title = {{Learning to Optimize with Stochastic Dominance Constraints}},
author = {Dai, Hanjun and Xue, Yuan and He, Niao and Wang, Yixin and Li, Na and Schuurmans, Dale and Dai, Bo},
booktitle = {Artificial Intelligence and Statistics},
year = {2023},
pages = {8991-9009},
volume = {206},
url = {https://mlanthology.org/aistats/2023/dai2023aistats-learning/}
}