Two-stream small-scale pedestrian detection network with feature aggregation for drone-view videos
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Xie, Han | - |
dc.contributor.author | Shin, Hyunchul | - |
dc.date.accessioned | 2022-12-20T05:49:09Z | - |
dc.date.available | 2022-12-20T05:49:09Z | - |
dc.date.issued | 2021-07 | - |
dc.identifier.issn | 0923-6082 | - |
dc.identifier.issn | 1573-0824 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/111271 | - |
dc.description.abstract | Detecting small-scale pedestrians in aerial images is a challenging task that can be difficult even for humans. Observing that the single image based method cannot achieve robust performance because of the poor visual cues of small instances. Considering that multiple frames may provide more information to detect such difficult case instead of only single frame, we design a novel video based pedestrian detection method with a two-stream network pipeline to fully utilize the temporal and contextual information of a video. An aggregated feature map is proposed to absorb the spatial and temporal information with the help of spatial and temporal sub-networks. To better capture motion information, a more refined flow net (SPyNet) is adopted instead of a simple flownet. In the spatial stream subnetwork, we modified the backbone network structure by increasing the feature map resolution with relatively larger receptive field to make it suitable for small-scale detection. Experimental results based on drone video datasets demonstrate that our approach improves detection accuracy in the case of small-scale instances and reduces false positive detections. By exploiting the temporal information and aggregating the feature maps, our two-stream method improves the detection performance by 8.48% in mean Average Precision (mAP) from that of the basic single stream R-FCN method, and it outperforms the state-of-the-art method by 3.09% on the Okutama Human-action dataset. | - |
dc.format.extent | 17 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | Kluwer Academic Publishers | - |
dc.title | Two-stream small-scale pedestrian detection network with feature aggregation for drone-view videos | - |
dc.type | Article | - |
dc.publisher.location | 네델란드 | - |
dc.identifier.doi | 10.1007/s11045-021-00764-1 | - |
dc.identifier.scopusid | 2-s2.0-85100703726 | - |
dc.identifier.wosid | 000616152400002 | - |
dc.identifier.bibliographicCitation | Multidimensional Systems and Signal Processing, v.32, no.3, pp 897 - 913 | - |
dc.citation.title | Multidimensional Systems and Signal Processing | - |
dc.citation.volume | 32 | - |
dc.citation.number | 3 | - |
dc.citation.startPage | 897 | - |
dc.citation.endPage | 913 | - |
dc.type.docType | Article | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Theory & Methods | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.subject.keywordAuthor | Pedestrian detection | - |
dc.subject.keywordAuthor | Feature aggregation | - |
dc.subject.keywordAuthor | Drone vision | - |
dc.subject.keywordAuthor | Neural network | - |
dc.subject.keywordAuthor | Deep learning | - |
dc.identifier.url | https://link.springer.com/article/10.1007/s11045-021-00764-1 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.