An enhanced deep reinforcement learning approach for efficient, effective, and equitable disaster relief distribution
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ahmad, Moiz | - |
dc.contributor.author | Tayyab, Muhammad | - |
dc.contributor.author | Habib, Muhammad Salman | - |
dc.date.accessioned | 2025-04-24T02:01:16Z | - |
dc.date.available | 2025-04-24T02:01:16Z | - |
dc.date.issued | 2025-03 | - |
dc.identifier.issn | 0952-1976 | - |
dc.identifier.issn | 1873-6769 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/125111 | - |
dc.description.abstract | Efficient disaster response, especially within the critical initial 72 h, is crucial for saving lives. However, allocating relief goods effectively to affected areas remains a complex challenge due to uncertainty, limited resources, and dynamic needs. This study addresses this challenge by proposing a multi-period integer nonlinear programming model for efficient, effective, and equitable distribution of relief goods during disaster response phase. To optimize relief allocation within entire 72-h, a novel decision-making approach is proposed that leverages the proximal policy optimization (PPO) algorithm. It uses deep residual neural networks for state-value and optimal action prediction with 5 value and 4 policy residual layers. Additionally, an algorithm-agnostic termination criterion based on episodic reward stall ensures effective convergence detection without requiring prior knowledge of optimal solution. The provided model and solution methods are validated through 30 hypothetical problem instances and a realistic earthquake response case study. The results demonstrate the superiority of proposed approach compared to traditional methods like dynamic programming, state-action-reward- state-action (SARSA), and Q-learning, in terms of both solution quality and sample efficiency. Notably, the deep residual networks and proposed termination criterion enable the PPO algorithm to achieve an average optimality gap of less than 10% for the majority of instances with consistent hyperparameters, while exhibiting significant sample efficiency gains, particularly for large-scale problems. This research empowers disaster managers with an efficient and timely relief delivery plan, ultimately contributing to saving lives in the face of disaster. Moreover, proposed termination criterion may improve the performance of reinforcement learning in other application areas. | - |
dc.format.extent | 28 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | PERGAMON-ELSEVIER SCIENCE LTD | - |
dc.title | An enhanced deep reinforcement learning approach for efficient, effective, and equitable disaster relief distribution | - |
dc.type | Article | - |
dc.publisher.location | 영국 | - |
dc.identifier.doi | 10.1016/j.engappai.2025.110002 | - |
dc.identifier.scopusid | 2-s2.0-85215370789 | - |
dc.identifier.wosid | 001405406700001 | - |
dc.identifier.bibliographicCitation | ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, v.143, pp 1 - 28 | - |
dc.citation.title | ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE | - |
dc.citation.volume | 143 | - |
dc.citation.startPage | 1 | - |
dc.citation.endPage | 28 | - |
dc.type.docType | Article | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Automation & Control Systems | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Automation & Control Systems | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.relation.journalWebOfScienceCategory | Engineering, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.subject.keywordPlus | RESOURCE-ALLOCATION | - |
dc.subject.keywordPlus | POSTDISASTER | - |
dc.subject.keywordPlus | ALGORITHM | - |
dc.subject.keywordAuthor | Disaster response | - |
dc.subject.keywordAuthor | Relief distribution | - |
dc.subject.keywordAuthor | Proximal policy optimization | - |
dc.subject.keywordAuthor | Q -learning | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Solution quality | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.