YogaNet: 3-D Yoga Asana Recognition Using Joint Angular Displacement Maps With ConvNets (original) (raw)
2019, IEEE Transactions on Multimedia
Representing 3D motion-capture sensor data with 2D color-coded joint distance maps (JDMs) as input to a deep neural network has been shown to be effective for 3D skeletal-based human action recognition tasks. However, the joint distances are limited by their ability to represent rotational joint movements, which account for a considerable amount of information in human action classification tasks. Moreover, for the subject, view and time invariance in the recognition process, the deep classifier needs training on JDMs along different coordinate axes from multiple streams. To overcome the above shortcomings of JDMs, we propose integrating joint angular movements along with the joint distances in a spatiotemporal color-coded image called a joint angular displacement map (JADM). In the literature, multistream deep convolutional neural networks (CNNs) have been employed to achieve invariance across subjects and views for 3D human action data, which is achieved by sacrificing training time for accuracy. To improve the recognition accuracy with reduced training times, we propose to test our JADMs with a singlestream deep CNN model. To test and analyze the proposed method, we chose video sequences of yoga. The 3D motion-capture data represent a complex set of actions with lateral and rotational spatiotemporal variations. We validated the proposed method using 3D traditional human action data from the publicly available datasets HDM05 and CMU. The proposed model can accurately recognize 3D yoga actions, which may help in building a 3D model-based yoga assistant tool.
Sign up for access to the world's latest research.
checkGet notified about relevant papers
checkSave papers to use in your research
checkJoin the discussion with peers
checkTrack your impact