Camera and LiDAR-based point painted voxel region-based convolutional neural network for robust 3D object detection
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Xie, Han | - |
dc.contributor.author | Zheng, Wenqi | - |
dc.contributor.author | Chen, Yunfan | - |
dc.contributor.author | Shin, Hyunchul | - |
dc.date.accessioned | 2023-02-21T05:37:22Z | - |
dc.date.available | 2023-02-21T05:37:22Z | - |
dc.date.issued | 2022-09 | - |
dc.identifier.issn | 1017-9909 | - |
dc.identifier.issn | 1560-229X | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/111503 | - |
dc.description.abstract | Most of the three-dimensional (3D) object detection methods based on LiDAR point cloud data achieve relatively high performance in general cases. However, when the LiDAR points have noise or some corruptions, the detection performance can be severely affected. We propose a 3D object detection method that combines point cloud information with two-dimensional (2D) semantic segmentation information to enhance the feature representation for difficult cases, such as sparse, noisy, and partially absent data. Motivated by the Pointpainting techniques, we designed an early-stage fusion method based on a Voxel region-based convolutional neural network (R-CNN) architecture. The 2D semantic segmentation scores obtained by the Pointpainting techniques are appended to the raw point cloud data. The voxel-based features and 2D semantic information improve the performance in detecting instances when the point cloud is corrupted. In addition, we also designed a multiscale hierarchical region of interest pooling strategy that reduced the computational cost of Voxel R-CNN by at least 43%. Our method shows competitive results with the state-of-the-art methods on the standard KITTI dataset. In addition, three corrupted KITTI datasets, KITTI sparse (KITTI-S), KITTI jittering (KITTI-J), and KITTI dropout (KITTI-D), were used for robustness testing. With the noisy LiDAR points, our proposed point painted Voxel R-CNN achieved superior detection performance over that of the baseline Voxel R-CNN for the moderate case, with a notable improvement of 11.13% in average precision (AP) on the 3D object detection and 14.3% in AP on the bird's eye view object detection. (c) 2022 SPIE and IS&T | - |
dc.format.extent | 13 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | S P I E - International Society for Optical Engineering | - |
dc.title | Camera and LiDAR-based point painted voxel region-based convolutional neural network for robust 3D object detection | - |
dc.type | Article | - |
dc.publisher.location | 미국 | - |
dc.identifier.doi | 10.1117/1.JEI.31.5.053025 | - |
dc.identifier.wosid | 000877897300055 | - |
dc.identifier.bibliographicCitation | Journal of Electronic Imaging, v.31, no.5, pp 1 - 13 | - |
dc.citation.title | Journal of Electronic Imaging | - |
dc.citation.volume | 31 | - |
dc.citation.number | 5 | - |
dc.citation.startPage | 1 | - |
dc.citation.endPage | 13 | - |
dc.type.docType | Article | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Optics | - |
dc.relation.journalResearchArea | Imaging Science & Photographic Technology | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Optics | - |
dc.relation.journalWebOfScienceCategory | Imaging Science & Photographic Technology | - |
dc.subject.keywordPlus | R-CNN | - |
dc.subject.keywordAuthor | three-dimensional object detection | - |
dc.subject.keywordAuthor | LiDAR | - |
dc.subject.keywordAuthor | fusion | - |
dc.subject.keywordAuthor | computer vision | - |
dc.identifier.url | https://www.spiedigitallibrary.org/journals/journal-of-electronic-imaging/volume-31/issue-05/053025/Camera-and-LiDAR-based-point-painted-voxel-region-based-convolutional/10.1117/1.JEI.31.5.053025.full#_=_ | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.