Reinforcement Learning for Optimizing Can-Order Policy with the Rolling Horizon Method
DC Field | Value | Language |
---|---|---|
dc.contributor.author | 노지성 | - |
dc.date.accessioned | 2023-09-04T05:32:07Z | - |
dc.date.available | 2023-09-04T05:32:07Z | - |
dc.date.issued | 2023-07 | - |
dc.identifier.issn | 2079-8954 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/114569 | - |
dc.description.abstract | This study presents a novel approach to a mixed-integer linear programming (MILP) model for periodic inventory management that combines reinforcement learning algorithms. The rolling horizon method (RHM) is a multi-period optimization approach that is applied to handle new information in updated markets. The RHM faces a limitation in easily determining a prediction horizon; to overcome this, a dynamic RHM is developed in which RL algorithms optimize the prediction horizon of the RHM. The state vector consisted of the order-up-to-level, real demand, total cost, holding cost, and backorder cost, whereas the action included the prediction horizon and forecasting demand for the next time step. The performance of the proposed model was validated through two experiments conducted in cases with stable and uncertain demand patterns. The results showed the effectiveness of the proposed approach in inventory management, particularly when the proximal policy optimization (PPO) algorithm was used for training compared with other reinforcement learning algorithms. This study signifies important advancements in both the theoretical and practical aspects of multi-item inventory management. | - |
dc.format.extent | 15 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | MDPI AG | - |
dc.title | Reinforcement Learning for Optimizing Can-Order Policy with the Rolling Horizon Method | - |
dc.type | Article | - |
dc.publisher.location | 스위스 | - |
dc.identifier.doi | 10.3390/systems11070350 | - |
dc.identifier.scopusid | 2-s2.0-85175107356 | - |
dc.identifier.wosid | 001072619000001 | - |
dc.identifier.bibliographicCitation | Systems, v.11, no.7, pp 350 - 364 | - |
dc.citation.title | Systems | - |
dc.citation.volume | 11 | - |
dc.citation.number | 7 | - |
dc.citation.startPage | 350 | - |
dc.citation.endPage | 364 | - |
dc.type.docType | Article | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | ssci | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Social Sciences - Other Topics | - |
dc.relation.journalWebOfScienceCategory | Social Sciences, Interdisciplinary | - |
dc.subject.keywordAuthor | can-order policy | - |
dc.subject.keywordAuthor | mixed-integer linear programming | - |
dc.subject.keywordAuthor | reinforcement learning | - |
dc.subject.keywordAuthor | rolling horizon method | - |
dc.subject.keywordAuthor | inventory management | - |
dc.identifier.url | https://www.mdpi.com/2079-8954/11/7/350 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.