Bi-PointFlowNet: Bidirectional Learning for Point Cloud Based Scene Flow Estimation
- Authors
- WENCAN, C.[WENCAN, CHENG]; Ko, J.H.[Ko, J.H.]
- Issue Date
- 2022
- Publisher
- Springer Science and Business Media Deutschland GmbH
- Keywords
- Bidirectional learning; Point cloud; Scene flow estimation
- Citation
- Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v.13688 LNCS, pp.108 - 124
- Indexed
- SCOPUS
- Journal Title
- Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
- Volume
- 13688 LNCS
- Start Page
- 108
- End Page
- 124
- URI
- https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/105739
- DOI
- 10.1007/978-3-031-19815-1_7
- ISSN
- 0302-9743
- Abstract
- Scene flow estimation, which extracts point-wise motion between scenes, is becoming a crucial task in many computer vision tasks. However, all of the existing estimation methods utilize only the unidirectional features, restricting the accuracy and generality. This paper presents a novel scene flow estimation architecture using bidirectional flow embedding layers. The proposed bidirectional layer learns features along both forward and backward directions, enhancing the estimation performance. In addition, hierarchical feature extraction and warping improve the performance and reduce computational overhead. Experimental results show that the proposed architecture achieved a new state-of-the-art record by outperforming other approaches with large margin in both FlyingThings3D and KITTI benchmarks. Codes are available at https://github.com/cwc1260/BiFlow. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Information and Communication Engineering > School of Electronic and Electrical Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.