Performance Comparison of Moving Target Classification based on Deep Learning
- Authors
- Hur, Jun; Nam, Haewoon
- Issue Date
- Oct-2022
- Publisher
- IEEE
- Keywords
- Classification; CNN; Deep learning; Radar target detection; Res-UNet; U-Net
- Citation
- 2022 13th International Conference on Information and Communication Technology Convergence (ICTC), pp 1533 - 1535
- Pages
- 3
- Indexed
- OTHER
- Journal Title
- 2022 13th International Conference on Information and Communication Technology Convergence (ICTC)
- Start Page
- 1533
- End Page
- 1535
- URI
- https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/113749
- DOI
- 10.1109/ICTC55196.2022.9952676
- Abstract
- Radar target detection is a basic but important process of radar systems, and it is difficult to distinguish and measure targets in real-world environments. Therefore, distinguishing between humans and animals based on radar signals is a difficult task in the field of ground radar. The radar signal processing section uses the in-phase/quadraturephase (I/Q) matrix radar signal data and geolocation types as inputs and performs binary classification to classify animals and humans. In this radar signal processing, deep learning methods are adopted as feasible solutions. However, there is a limited lack
of training data in the real world and a problem with jamming signals, which are adversarial attacks. However, it is difficult to collect a lot of training data in a real-time environment. Reflecting this, we learn only some data from MAFAT Radar Challenge data to compare and analyze the classification performance of
conventional methods convolutional neural network (CNN) and
CNN-based U-Net and U-Net with residual blocks U-Net (ResUNet) algorithms
- Files in This Item
-
Go to Link
- Appears in
Collections - COLLEGE OF ENGINEERING SCIENCES > SCHOOL OF ELECTRICAL ENGINEERING > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/113749)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.