Skeleton Based View Invariant Human Action Recognition using Convolutional Neural Networks
K. Vijaya Prasad1, P. V. V. Kishore2, O. Srinivasa Rao3
1K. Vijaya Prasad, Research Scholar, Department of ECE, JNT University Kakinada, Kakinada, India.
2P. V. V. Kishore, Professor, Department of ECE, K L University, KLEF, Guntur, India.
3O. Srinivasa Rao, Professor, Department of CSE, UCEK, JNT University Kakinada, Kakinada, India.
Manuscript received on 18 March 2019 | Revised Manuscript received on 25 March 2019 | Manuscript published on 30 July 2019 | PP: 4860-4867 | Volume-8 Issue-2, July 2019 | Retrieval Number: B3547078219/19©BEIESP | DOI: 10.35940/ijrte.B3547.078219
Open Access | Ethics and Policies | Cite | Mendeley | Indexing and Abstracting
© The Authors. Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP). This is an open access article under the CC-BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)
Abstract: The skeletal based human action recognition has its significant applications in the field of human computer interaction and human recognition from surveillance videos. However, the tasks suffers from the major challenges like view variance and noise in the data. These problems are limiting the performance of human action recognition. This paper focuses to solve these problems by adopting sequence based view invariant transform to effectively represent the spatio-temporal information of the skeletal data. The task of human action recognition in this paper is performed in three stages. Firstly, the raw 3D skeletal joint data obtained from the Microsoft Kinect sensor is transformed to eliminate the problem of view variations on a spatio-temporal data by implementing sequence-based view invariant transform. In the second stage, the transformed joint locations of the skeletal data will be converted to RGB images by a color coding technique and forms a transformed joint location maps (TJLMs) . As a third stage, the discriminating features were extracted by the novel CNN architecture to performs the human action recognition task by means of class scores. Noticeable amount of recognition scores are achieved. Extensive experiments in four difficult 3D action datasets constantly show our method’s superiority. The performance of the proposed method is compared with the other state-of-the-art methods.
Keywords: Human Action Recognition, Sequence Based View Invariant Transform, Convolutional Neural Networks.
Scope of the Article: Pattern Recognition