Learning to Schedule Joint Radar-Communication with Deep Multi-Agent Reinforcement Learning
- Authors
- Lee, J.; Niyato, T.D.; Guan, Y.L.; Kim, D.I.
- Issue Date
- Jan-2022
- Publisher
- Institute of Electrical and Electronics Engineers Inc.
- Keywords
- Accidents; Automotive engineering; Cameras; communication; deep learning; Radar; Reinforcement learning; Reinforcement learning; Sensor systems; Sensors; task scheduling; vehicle safety
- Citation
- IEEE Transactions on Vehicular Technology, v.71, no.1, pp 406 - 422
- Pages
- 17
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE Transactions on Vehicular Technology
- Volume
- 71
- Number
- 1
- Start Page
- 406
- End Page
- 422
- URI
- https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/92784
- DOI
- 10.1109/TVT.2021.3124810
- ISSN
- 0018-9545
1939-9359
- Abstract
- Radar detection and communication are two of several sub-tasks essential for the operation of next-generation autonomous vehicles (AVs). The former is required for sensing and perception, more frequently so under various unfavorable environmental conditions such as heavy precipitation; the latter is needed to transmit time-critical data. Forthcoming proliferation of faster 5G networks utilizing mmWave is likely to lead to interference with automotive radar sensors, which has led to a body of research on the development of Joint Radar Communication (JRC) systems and solutions. This paper considers the problem of time-sharing for JRC, with the additional simultaneous objective of minimizing the average age of information (AoI) transmitted by a JRC-equipped AV. We first formulate the problem as a Markov Decision Process (MDP). We then propose a more general multi-agent system, with an appropriate medium access control protocol (MAC), which is formulated as a partially observed Markov game (POMG). To solve the POMG, we propose a multi-agent extension of the Proximal Policy Optimization (PPO) algorithm, along with algorithmic features to enhance learning from raw observations. Simulations are run with a range of environmental parameters to mimic variations in real-world operation. The results show that the chosen deep reinforcement learning methods allow the agents to obtain good results with minimal a priori knowledge about the environment. IEEE
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Information and Communication Engineering > School of Electronic and Electrical Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.