Cost-Efficient Reinforcement Learning for Optimal Trade Execution on Dynamic Market Environment

3Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Learning a high-performance trade execution model via reinforcement learning (RL) requires interaction with the real dynamic market. However, the massive interactions required by direct RL would result in a significant training overhead. In this paper, we propose a cost-efficient reinforcement learning (RL) approach called Deep Dyna-Double Q-learning (D3Q), which integrates deep reinforcement learning and planning to reduce the training overhead while improving the trading performance. Specifically, D3Q includes a learnable market environment model, which approximates the market impact using real market experience, to enhance policy learning via the learned environment. Meanwhile, we propose a novel state-balanced exploration scheme to solve the exploration bias caused by the non-increasing residual inventory during the trade execution to accelerate model learning. As demonstrated by our extensive experiments, the proposed D3Q framework significantly increases sample efficiency and outperforms state-of-the-art methods on average trading cost as well.

Cite

CITATION STYLE

APA

Chen, D., Zhu, Y., Liu, M., & Li, J. (2022). Cost-Efficient Reinforcement Learning for Optimal Trade Execution on Dynamic Market Environment. In Proceedings of the 3rd ACM International Conference on AI in Finance, ICAIF 2022 (pp. 386–393). Association for Computing Machinery, Inc. https://doi.org/10.1145/3533271.3561761

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free