Action recognition using Kinematics Posture Feature on 3D skeleton joint locations
Article
Ahad, M. A. R., Ahmed, M., Antar, A. D., Makihara, Y. and Yagi. Y. 2021. Action recognition using Kinematics Posture Feature on 3D skeleton joint locations. Pattern Recognition Letters. 145, pp. 216-224. https://doi.org/10.1016/j.patrec.2021.02.013
Authors | Ahad, M. A. R., Ahmed, M., Antar, A. D., Makihara, Y. and Yagi. Y. |
---|---|
Abstract | Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance. |
Keywords | AI; Activity recognition; Skeleton; Vision; Deep learning |
Journal | Pattern Recognition Letters |
Journal citation | 145, pp. 216-224 |
ISSN | 0167-8655 |
Year | 2021 |
Publisher | Elsevier |
Publisher's version | License File Access Level Anyone |
Digital Object Identifier (DOI) | https://doi.org/10.1016/j.patrec.2021.02.013 |
Publication dates | |
Online | 03 Mar 2021 |
May 2021 | |
Publication process dates | |
Accepted | 26 Feb 2021 |
Deposited | 04 Dec 2023 |
Copyright holder | © 2021, The Authors |
https://repository.uel.ac.uk/item/8wz27
Download files
82
total views32
total downloads2
views this month0
downloads this month