VISUALCENT: Visual Human Analysis using Dynamic Centroid Representation
- Authors
- 이영문
- Issue Date
- May-2025
- Publisher
- IEEE
- Citation
- IEEE International Conference on Automatic Face and Gesture Recognition, pp 1 - 5
- Pages
- 5
- Indexed
- FOREIGN
- Journal Title
- IEEE International Conference on Automatic Face and Gesture Recognition
- Start Page
- 1
- End Page
- 5
- URI
- https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/125173
- DOI
- 10.48550/arXiv.2504.19032 Focus to learn more
- Abstract
- We introduce VISUALCENT, a unified human pose and instance segmentation framework to address generalizability and scalability limitations to multi person visual human analysis. VISUALCENT leverages centroid based bottom up keypoint detection paradigm and uses Keypoint Heatmap incorporating Disk Representation and KeyCentroid to identify the optimal keypoint coordinates. For the unified segmentation task, an explicit keypoint is defined as a dynamic centroid called MaskCentroid to swiftly cluster pixels to specific human instance during rapid changes in human body movement or significantly occluded environment. Experimental results on COCO and OCHuman datasets demonstrate VISUALCENTs accuracy and real time performance advantages, outperforming existing methods in mAP scores and execution frame rate per second. The implementation is available on the project page.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - COLLEGE OF ENGINEERING SCIENCES > DEPARTMENT OF ROBOT ENGINEERING > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.