A Selective Portfolio Management Algorithm with Off-Policy Reinforcement Learning Using Dirichlet Distribution
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yang, Hyunjun | - |
dc.contributor.author | Park, Hyeonjun | - |
dc.contributor.author | Lee, Kyungjae | - |
dc.date.accessioned | 2023-03-08T04:48:40Z | - |
dc.date.available | 2023-03-08T04:48:40Z | - |
dc.date.issued | 2022-12 | - |
dc.identifier.issn | 2075-1680 | - |
dc.identifier.issn | 2075-1680 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/61125 | - |
dc.description.abstract | Existing methods in portfolio management deterministically produce an optimal portfolio. However, according to modern portfolio theory, there exists a trade-off between a portfolio's expected returns and risks. Therefore, the optimal portfolio does not exist definitively, but several exist, and using only one deterministic portfolio is disadvantageous for risk management. We proposed Dirichlet Distribution Trader (DDT), an algorithm that calculates multiple optimal portfolios by taking Dirichlet Distribution as a policy. The DDT algorithm makes several optimal portfolios according to risk levels. In addition, by obtaining the pi value from the distribution and applying importance sampling to off-policy learning, the sample is used efficiently. Furthermore, the architecture of our model is scalable because the feed-forward of information between portfolio stocks occurs independently. This means that even if untrained stocks are added to the portfolio, the optimal weight can be adjusted. We also conducted three experiments. In the scalability experiment, it was shown that the DDT extended model, which is trained with only three stocks, had little difference in performance from the DDT model that learned all the stocks in the portfolio. In an experiment comparing the off-policy algorithm and the on-policy algorithm, it was shown that the off-policy algorithm had good performance regardless of the stock price trend. In an experiment comparing investment results according to risk level, it was shown that a higher return or a better Sharpe ratio could be obtained through risk control. | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | MDPI | - |
dc.title | A Selective Portfolio Management Algorithm with Off-Policy Reinforcement Learning Using Dirichlet Distribution | - |
dc.type | Article | - |
dc.identifier.doi | 10.3390/axioms11120664 | - |
dc.identifier.bibliographicCitation | AXIOMS, v.11, no.12 | - |
dc.description.isOpenAccess | Y | - |
dc.identifier.wosid | 000902187200001 | - |
dc.identifier.scopusid | 2-s2.0-85144713898 | - |
dc.citation.number | 12 | - |
dc.citation.title | AXIOMS | - |
dc.citation.volume | 11 | - |
dc.type.docType | Article | - |
dc.publisher.location | 스위스 | - |
dc.subject.keywordAuthor | deep reinforcement learning | - |
dc.subject.keywordAuthor | exploration methods | - |
dc.subject.keywordAuthor | portfolio optimization | - |
dc.subject.keywordPlus | NETWORKS | - |
dc.relation.journalResearchArea | Mathematics | - |
dc.relation.journalWebOfScienceCategory | Mathematics, Applied | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
84, Heukseok-ro, Dongjak-gu, Seoul, Republic of Korea (06974)02-820-6194
COPYRIGHT 2019 Chung-Ang University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.