4.3 Article

Continuous-time mean-variance portfolio selection: A reinforcement learning framework

Journal

MATHEMATICAL FINANCE
Volume 30, Issue 4, Pages 1273-1308

Publisher

WILEY
DOI: 10.1111/mafi.12281

Keywords

empirical study; entropy regularization; Gaussian distribution; mean-variance portfolio selection; policy improvement; reinforcement learning; simulation; stochastic control; theorem; value function

Funding

  1. Nie Center for Intelligent Asset Management at Columbia
  2. Columbia University
  3. Nie Center for Intelligent Asset Management

Ask authors/readers for more resources

We approach the continuous-time mean-variance portfolio selection with reinforcement learning (RL). The problem is to achieve the best trade-off between exploration and exploitation, and is formulated as an entropy-regularized, relaxed stochastic control problem. We prove that the optimal feedback policy for this problem must be Gaussian, with time-decaying variance. We then prove a policy improvement theorem, based on which we devise an implementable RL algorithm. We find that our algorithm and its variant outperform both traditional and deep neural network based algorithms in our simulation and empirical studies.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available