An adaptive dual-level reinforcement learning approach for optimal trade execution
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim, Soohan | - |
dc.contributor.author | Kim, Jimyeong | - |
dc.contributor.author | Sul, Hong Kee | - |
dc.contributor.author | Hong, Youngjoon | - |
dc.date.accessioned | 2024-06-14T06:00:48Z | - |
dc.date.available | 2024-06-14T06:00:48Z | - |
dc.date.issued | 2024-10 | - |
dc.identifier.issn | 0957-4174 | - |
dc.identifier.issn | 1873-6793 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/74197 | - |
dc.description.abstract | The purpose of this research is to devise a tactic that can closely track the daily cumulative volume-weighted average price (VWAP) using reinforcement learning while minimizing the deviation from the VWAP. Previous studies often choose a relatively short trading horizon to implement their models, making it difficult to accurately track the daily cumulative VWAP since the stock price movement is often insignificant within the short trading horizon. On the other hand, training reinforcement learning models directly over a longer, daily horizon is burdensome due to extensive sequence length. Hence, there is a need for a method that can divide the long daily horizon into smaller, more manageable segments. We propose a method that leverages the U-shaped pattern of intraday stock trade volumes and uses Proximal Policy Optimization (PPO) as the learning algorithm. Our method follows a dual-level approach: a Transformer model that captures the overall (global) distribution of daily volumes in a U-shape, and a LSTM model that handles the distribution of orders within smaller (local) time intervals. The results from our experiments suggest that this dual-level architecture improves cumulative VWAP tracking accuracy compared to previous reinforcement learning approaches. The key finding is that explicitly accounting for the U-shaped intraday volume pattern leads to better performance in approximating the cumulative daily VWAP. This has implications for developing trading strategies that need to efficiently track VWAP over a full trading day. © 2024 Elsevier Ltd | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | Elsevier Ltd | - |
dc.title | An adaptive dual-level reinforcement learning approach for optimal trade execution | - |
dc.type | Article | - |
dc.identifier.doi | 10.1016/j.eswa.2024.124263 | - |
dc.identifier.bibliographicCitation | Expert Systems with Applications, v.252 | - |
dc.description.isOpenAccess | N | - |
dc.identifier.wosid | 001245090900001 | - |
dc.identifier.scopusid | 2-s2.0-85193754510 | - |
dc.citation.title | Expert Systems with Applications | - |
dc.citation.volume | 252 | - |
dc.type.docType | Article | - |
dc.publisher.location | 영국 | - |
dc.subject.keywordAuthor | Markov decision process | - |
dc.subject.keywordAuthor | Optimal trade execution | - |
dc.subject.keywordAuthor | Proximal policy optimization | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Volume-weighted average price | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Operations Research & Management Science | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Operations Research & Management Science | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
84, Heukseok-ro, Dongjak-gu, Seoul, Republic of Korea (06974)02-820-6194
COPYRIGHT 2019 Chung-Ang University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.