A resource conscious human action recognition framework using 26-layered deep convolutional neural network
- Authors
- Khan, M.A.; Zhang, Y.-D.; Khan, S.A.; Attique, M.; Rehman, A.; Seo, S.
- Issue Date
- Nov-2021
- Publisher
- Springer
- Keywords
- Action recognition; CNN architecture; ELM; Features fusion; Features selection
- Citation
- Multimedia Tools and Applications, v.80, no.28-29, pp 35827 - 35849
- Pages
- 23
- Journal Title
- Multimedia Tools and Applications
- Volume
- 80
- Number
- 28-29
- Start Page
- 35827
- End Page
- 35849
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/62113
- DOI
- 10.1007/s11042-020-09408-1
- ISSN
- 1380-7501
1432-1882
- Abstract
- Vision-based human action recognition (HAR) is a hot topic of research from the decade due to a few popular applications such as visual surveillance and robotics. For correct action recognition, various local and global points are requires known as features. These features modified during the variation in human movement. But due to a bit change in several human actions, the features of these actions are mixed that degrade the recognition performance. In this article, we design a new 26-layered Convolutional Neural Network (CNN) architecture for accurate complex action recognition. The features are extracted from the global average pooling layer and fully connected (FC) layer, and fused by a proposed high entropy-based approach. Further, we propose a feature selection method name Poisson distribution along with Univariate Measures (PDaUM). Few of fused CNN features are irrelevant, and few of them are redundant that makes the incorrect prediction among complex human actions. Therefore, the proposed PDaUM based approach selects only the strongest features that later passed to the Extreme Learning Machine (ELM) and Softmax for final recognition. Four datasets are using for experimental analysis - HMDB51 (51 classes), UCF Sports (10 classes), KTH (6 classes), and Weizmann (10 classes). On these datasets, the ELM classifier gives an improved performance as compared to a Softmax classifier. The achieved accuracy on each dataset is 81.4%, 99.2%, 98.3%, and 98.7%, respectively. Comparison with existing techniques, it is shown that the proposed architecture gives better performance in terms of accuracy and testing time. © 2020, Springer Science+Business Media, LLC, part of Springer Nature.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - ETC > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/62113)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.