자율주행차량을 위한 CycleGAN 기반 Depth Completion 기법CycleGAN-Based Depth Completion for Autonomous Vehicles
- Other Titles
- CycleGAN-Based Depth Completion for Autonomous Vehicles
- Authors
- 응 웬민찌; 유명식
- Issue Date
- May-2022
- Publisher
- 한국통신학회
- Keywords
- depth completion; cycleGAN; semantic segmentation; autonomous vehicle; sensor fusion
- Citation
- 한국통신학회논문지, v.47, no.5, pp.781 - 788
- Journal Title
- 한국통신학회논문지
- Volume
- 47
- Number
- 5
- Start Page
- 781
- End Page
- 788
- URI
- http://scholarworks.bwise.kr/ssu/handle/2018.sw.ssu/43005
- DOI
- 10.7840/kics.2022.47.5.781
- ISSN
- 1226-4717
- Abstract
- Depth completion is a challenging task supporting the purpose of scene understanding and environment perception in an autonomous vehicle. The existing method considered multiple modals input such as RGB images and depth LIDAR images to utilize the complementary characteristics of those two sensors. However, traditional autoencoder approaches have shown limitations in representing the data in low dimensional space.
Moreover, depth discontinuity also happened when fusing the camera image and LIDAR image due to the light sensitivity in the RGB image. In our study, we are adapting CycleGAN focusing on learning the distribution of the data rather than the pixel density to reconstruct the depth into dense one. We also consider the semantic segmentation as additional input to mitigate the depth discontinuity problem. Our framework is trained and evaluated on the KITTI benchmark with synchronized data capturing various road scenery. The experimental results prove the proposed framework to be competitive performance and efficient in depth completion task.
- Files in This Item
-
Go to Link
- Appears in
Collections - ETC > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.