Intelligent Self-Optimization for Task Offloading in LEO-MEC-Assisted Energy-Harvesting-UAV Systems
- Authors
- Lakew, Demeke Shumeye; Tran, Anh-Tien; Dao, Nhu-Ngoc; Cho, Sungrae
- Issue Date
- Nov-2024
- Publisher
- IEEE Computer Society
- Keywords
- Autonomous aerial vehicles; computation offloading; deep reinforcement learning; energy harvesting; Internet of Things; LEO satellite; Low earth orbit satellites; resource allocation; Resource management; Satellites; Servers; Task analysis; UAV
- Citation
- IEEE Transactions on Network Science and Engineering, v.11, no.6, pp 1 - 14
- Pages
- 14
- Journal Title
- IEEE Transactions on Network Science and Engineering
- Volume
- 11
- Number
- 6
- Start Page
- 1
- End Page
- 14
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/72693
- DOI
- 10.1109/TNSE.2023.3349321
- ISSN
- 2327-4697
- Abstract
- Given the notable surge in Internet of Things (IoT) devices, low Earth orbit (LEO) satellites and unmanned aerial vehicles (UAVs) have emerged as promising networking components to supplement the network capacity and ensure seamless coverage in 6G, especially over remote areas. However, task offloading and resource management are challenging to realize because of the limited connectivity duration of LEO satellites attributable to their high mobility and UAVs limited resources. Thus, this paper proposes a network model in which mobile edge computing (MEC)-enabled multiple LEO satellites in-orbit provide computational services for a resource-constrained energy harvesting UAV (EH-UAV). The EH-UAV collects data from remote IoT/sensor devices and periodically generates a computational task. To optimize the system model, we formulate a joint LEO-MEC server selection, transmission power allocation, and partial task offloading decision-making problem to maximize the service satisfaction and alleviate energy dissipation under the constraints of connectivity duration, task deadline, and available energy. To circumvent the non-convexity and dynamicity of the problem, it is reformulated as a reinforcement learning problem and solved using a novel mixed discrete-continuous control deep reinforcement learning (<inline-formula><tex-math notation=LaTeX>$MDC^{2}-DRL$</tex-math></inline-formula>) based algorithm with an action shaping function. Simulation results demonstrate that <inline-formula><tex-math notation=LaTeX>$MDC^{2}-DRL$</tex-math></inline-formula> effectively converges and outperforms the existing methods. IEEE
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Software > School of Computer Science and Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.