4.7 Article

Cost-Sensitive Portfolio Selection via Deep Reinforcement Learning

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TKDE.2020.2979700

Keywords

Portfolio selection; reinforcement learning; deep learning; transaction cost

Funding

  1. National Natural Science Foundation of China (NSFC) [61876208, 71971164]
  2. NSFC [61836003]
  3. Program for Guangdong Introducing Innovative and Enterpreneurial Teams [2017ZT07X183]
  4. Guangdong Provincial Scientific and Technological Funds [2018B010107001, 2018B010108002]
  5. Pearl River S&T Nova Program of Guangzhou [201806010081]
  6. Tencent AI Lab Rhino-Bird Focused Research Program [JR201902]
  7. Fundamental Research Funds for the Central Universities [D2191240]

Ask authors/readers for more resources

Portfolio selection is a challenging task in finance, and this paper proposes a cost-sensitive method using deep reinforcement learning. The proposed method extracts price series patterns and asset correlations, and controls both transaction and risk costs effectively. Empirical results demonstrate its effectiveness and superiority in profitability, cost-sensitivity, and representation abilities.
Portfolio Selection is an important real-world financial task and has attracted extensive attention in artificial intelligence communities. This task, however, has two main difficulties: (i) the non-stationary price series and complex asset correlations make the learning of feature representation very hard; (ii) the practicality principle in financial markets requires controlling both transaction and risk costs. Most existing methods adopt handcraft features and/or consider no constraints for the costs, which may make them perform unsatisfactorily and fail to control both costs in practice. In this paper, we propose a cost-sensitive portfolio selection method with deep reinforcement learning. Specifically, a novel two-stream portfolio policy network is devised to extract both price series patterns and asset correlations, while a new cost-sensitive reward function is developed to maximize the accumulated return and constrain both costs via reinforcement learning. We theoretically analyze the near-optimality of the proposed reward, which shows that the growth rate of the policy regarding this reward function can approach the theoretical optimum. We also empirically evaluate the proposed method on real-world datasets. Promising results demonstrate the effectiveness and superiority of the proposed method in terms of profitability, cost-sensitivity and representation abilities.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available