4.7 Article

Self-attention based deep direct recurrent reinforcement learning with hybrid loss for trading signal generation

Journal

INFORMATION SCIENCES
Volume 623, Issue -, Pages 592-606

Publisher

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2022.12.042

Keywords

Recurrent reinforcement learning; Self-attention mechanism; Algorithmic trading; Decision support

Ask authors/readers for more resources

This paper proposes a novel algorithmic trading model based on recurrent reinforcement learning, which maximizes the expected rewards by extracting temporal features from complex observations. The model incorporates hybrid learning loss and self-attention mechanism to enhance the decision-making process. The proposed SA-DDR-HL model demonstrates superior performance compared to baseline benchmark models.
Algorithmic trading based on machine learning has the advantage of using intrinsic features and embedded causality in complex stock price time series. We propose a novel algorithmic trading model based on recurrent reinforcement learning, optimized for making consecutive trading signals. This paper elaborates on how temporal features from complex observation are optimally extracted to maximize the expected rewards of the reinforcement learning model. Our model incorporates the hybrid learning loss to allow sequences of hidden features for reinforcement learning to contain the original state's characteristics fully. The self-attention mechanism is also introduced to our model for learning the temporal importance of the hidden representation series, which helps the reinforcement learning model to be aware of temporal dependence for its decision-making. In this paper, we verify the effectiveness of proposed model using some major market indices and the representative stocks in each sector of S&P500. The augmented structure that we propose has a significant dominance on trading performance. Our proposed model, self-attention based deep direct recurrent reinforcement learning with hybrid loss (SA-DDR-HL), shows superior performance over well-known baseline benchmark models, including machine learning and time series models. (c) 2022 Elsevier Inc. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available