Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Integrating Risk-Averse and Constrained Reinforcement Learning for Robust Decision-Making in High-Stakes Scenarios

Full metadata record
DC Field Value Language
dc.contributor.authorAhmad, Moiz-
dc.contributor.authorRamzan, Muhammad Babar-
dc.contributor.authorOmair, Muhammad-
dc.contributor.authorHabib, Muhammad Salman-
dc.date.accessioned2025-06-16T06:00:25Z-
dc.date.available2025-06-16T06:00:25Z-
dc.date.issued2024-07-
dc.identifier.issn2227-7390-
dc.identifier.issn2227-7390-
dc.identifier.urihttps://scholarworks.bwise.kr/erica/handle/2021.sw.erica/125637-
dc.description.abstractThis paper considers a risk-averse Markov decision process (MDP) with non-risk constraints as a dynamic optimization framework to ensure robustness against unfavorable outcomes in high-stakes sequential decision-making situations such as disaster response. In this regard, strong duality is proved while making no assumptions on the problem’s convexity. This is necessary for some real-world issues, e.g., in the case of deprivation costs in the context of disaster relief, where convexity cannot be ensured. Our theoretical results imply that the problem can be exactly solved in a dual domain where it becomes convex. Based on our duality results, an augmented Lagrangian-based constraint handling mechanism is also developed for risk-averse reinforcement learning algorithms. The mechanism is proved to be theoretically convergent. Finally, we have also empirically established the convergence of the mechanism using a multi-stage disaster response relief allocation problem while using a fixed negative reward scheme as a benchmark. © 2024 by the authors.-
dc.format.extent29-
dc.language영어-
dc.language.isoENG-
dc.publisherMultidisciplinary Digital Publishing Institute (MDPI)-
dc.titleIntegrating Risk-Averse and Constrained Reinforcement Learning for Robust Decision-Making in High-Stakes Scenarios-
dc.typeArticle-
dc.publisher.location스위스-
dc.identifier.doi10.3390/math12131954-
dc.identifier.scopusid2-s2.0-85198409366-
dc.identifier.wosid001269761800001-
dc.identifier.bibliographicCitationMathematics, v.12, no.13, pp 1 - 29-
dc.citation.titleMathematics-
dc.citation.volume12-
dc.citation.number13-
dc.citation.startPage1-
dc.citation.endPage29-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaMathematics-
dc.relation.journalWebOfScienceCategoryMathematics-
dc.subject.keywordPlusMANAGEMENT-
dc.subject.keywordPlusRELIEF-
dc.subject.keywordAuthoraugmented Lagrangian-
dc.subject.keywordAuthorconstrained reinforcement learning-
dc.subject.keywordAuthordynamic decision-making-
dc.subject.keywordAuthorMarkov risk-
dc.subject.keywordAuthornon-convexities-
dc.subject.keywordAuthorrobust decision-making-
dc.identifier.urlhttps://www.mdpi.com/2227-7390/12/13/1954-
Files in This Item
Go to Link
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher HABIB, MUHAMMAD SALMAN photo

HABIB, MUHAMMAD SALMAN
ERICA부총장 한양인재개발원 (ERICA 창의융합교육원)
Read more

Altmetrics

Total Views & Downloads

BROWSE