Semantic Recognition of Human-Object Interactions via Gaussian-based Elliptical Modelling and Pixel-Level Labeling
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Khalid, Nida | - |
dc.contributor.author | Ghadi, Yazeed Yasin | - |
dc.contributor.author | Gochoo, Munkhjargal | - |
dc.contributor.author | Jalal, Ahmad | - |
dc.contributor.author | Kim, Kibum | - |
dc.date.accessioned | 2023-08-16T07:33:52Z | - |
dc.date.available | 2023-08-16T07:33:52Z | - |
dc.date.issued | 2021-07 | - |
dc.identifier.issn | 2169-3536 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/113871 | - |
dc.description.abstract | Human-Object Interaction (HOI) recognition, due to its significance in many computer vision-based applications, requires in-depth and meaningful details from image sequences. Incorporating semantics in scene understanding has led to a deep understanding of human-centric actions. Therefore, in this research work, we propose a semantic HOI recognition system based on multi-vision sensors. In the proposed system, the de-noised RGB and depth images, via Bilateral Filtering (BLF), are segmented into multiple clusters using a Simple Linear Iterative Clustering (SLIC) algorithm. The skeleton is then extracted from segmented RGB and depth images via Euclidean Distance Transform (EDT). Human joints, extracted from the skeleton, provide the annotations for accurate pixel-level labeling. An elliptical human model is then generated via a Gaussian Mixture Model (GMM). A Conditional Random Field (CRF) model is trained to allocate a specific label to each pixel of different human body parts and an interaction object. Two semantic feature types that are extracted from each labeled body part of the human and labelled objects are: Fiducial points and 3D point cloud. Features descriptors are quantized using Fisher's Linear Discriminant Analysis (FLDA) and classified using K-ary Tree Hashing (KATH). In experimentation phase the recognition accuracy achieved with the Sports dataset is 92.88%, with the Sun Yat-Sen University (SYSU) 3D HOI dataset is 93.5% and with the Nanyang Technological University (NTU) RGB+D dataset it is 94.16%. The proposed system is validated via extensive experimentation and should be applicable to many computer-vision based applications such as healthcare monitoring, security systems and assisted living etc. © 2013 IEEE. | - |
dc.format.extent | 18 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | Institute of Electrical and Electronics Engineers Inc. | - |
dc.title | Semantic Recognition of Human-Object Interactions via Gaussian-based Elliptical Modelling and Pixel-Level Labeling | - |
dc.type | Article | - |
dc.publisher.location | 미국 | - |
dc.identifier.doi | 10.1109/ACCESS.2021.3101716 | - |
dc.identifier.scopusid | 2-s2.0-85111592359 | - |
dc.identifier.wosid | 000684675000001 | - |
dc.identifier.bibliographicCitation | IEEE Access, v.9, pp 111249 - 111266 | - |
dc.citation.title | IEEE Access | - |
dc.citation.volume | 9 | - |
dc.citation.startPage | 111249 | - |
dc.citation.endPage | 111266 | - |
dc.type.docType | 정기학술지(Article(Perspective Article포함)) | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordPlus | BODY PARTS | - |
dc.subject.keywordPlus | NETWORKS | - |
dc.subject.keywordPlus | SEGMENTATION | - |
dc.subject.keywordAuthor | 3D point cloud | - |
dc.subject.keywordAuthor | fiducial points | - |
dc.subject.keywordAuthor | human-object interaction | - |
dc.subject.keywordAuthor | K-ary tree hashing | - |
dc.subject.keywordAuthor | pixel labeling | - |
dc.subject.keywordAuthor | semantic segmentation | - |
dc.subject.keywordAuthor | super-pixels | - |
dc.identifier.url | https://ieeexplore.ieee.org/document/9502603 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.