Meta‐action descriptor for action recognition in RGBD video

Action recognition is one of the hottest research topics in computer vision. Recent methods represent actions based on global or local video features. These approaches, however, lack semantic structure and may not provide a deep insight into the essence of an action. In this work, the authors argue...

Full description

Bibliographic Details
Main Authors: Min Huang, Song‐Zhi Su, Guo‐Rong Cai, Hong‐Bo Zhang, Donglin Cao, Shao‐Zi Li
Format: Article
Language:English
Published: Wiley 2017-06-01
Series:IET Computer Vision
Subjects:
Online Access:https://doi.org/10.1049/iet-cvi.2016.0252
Description
Summary:Action recognition is one of the hottest research topics in computer vision. Recent methods represent actions based on global or local video features. These approaches, however, lack semantic structure and may not provide a deep insight into the essence of an action. In this work, the authors argue that semantic clues, such as joint positions and part‐level motion clustering, help verify actions. To this end, a meta‐action descriptor for action recognition in RGBD video is proposed in this study. Specifically, two discrimination‐based strategies – dynamic and discriminative part clustering – are introduced to improve accuracy. Experiments conducted on the MSR Action 3D dataset show that the proposed method significantly outperforms the methods without joint position semantic.
ISSN:1751-9632
1751-9640