Portfolio Choices with Orthogonal Bandit Learning

Abstract

The investigation and development of new methods from diverse perspectives to shed light on portfolio choice problems has never stagnated in financial research. Recently, multi-armed bandits have drawn intensive attention in various machine learning applications in online settings. The tradeoff between exploration and exploitation to maximize rewards in bandit algorithms naturally establishes a connection to portfolio choice problems. In this paper, we present a bandit algorithm for conducting online portfolio choices by effectually exploiting correlations among multiple arms. Through constructing orthogonal portfolios from multiple assets and integrating with the upper confidence bound bandit framework, we derive the optimal portfolio strategy that represents the combination of passive and active investments according to a risk-adjusted reward function. Compared with oft-quoted trading strategies in finance and machine learning fields across representative real-world market datasets, the proposed algorithm demonstrates superiority in both risk-adjusted return and cumulative wealth.

Cite

Text

Shen et al. "Portfolio Choices with Orthogonal Bandit Learning." International Joint Conference on Artificial Intelligence, 2015.

Markdown

[Shen et al. "Portfolio Choices with Orthogonal Bandit Learning." International Joint Conference on Artificial Intelligence, 2015.](https://mlanthology.org/ijcai/2015/shen2015ijcai-portfolio/)

BibTeX

@inproceedings{shen2015ijcai-portfolio,
  title     = {{Portfolio Choices with Orthogonal Bandit Learning}},
  author    = {Shen, Weiwei and Wang, Jun and Jiang, Yu-Gang and Zha, Hongyuan},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2015},
  pages     = {974-},
  url       = {https://mlanthology.org/ijcai/2015/shen2015ijcai-portfolio/}
}