Flexible human action recognition in depth video sequences using masked joint trajectories

Published in EURASIP Journal on Image and Video Processing, 2016

Recommended citation: Tejero-de-Pablos, A., Nakashima, Y., Yokoya, N., Díaz-Pernas, F. J., & Martínez-Zarzuela, M. (2016). Flexible human action recognition in depth video sequences using masked joint trajectories. EURASIP Journal on Image and Video Processing, 2016(1), 1-12

Human action recognition applications are greatly benefited from the use of commodity depth sensors that are capable of skeleton tracking. Some of these applications (e.g. customizable gesture interfaces) require learning of new actions at runtime and may not count with many training instances. This paper presents a human action recognition method designed for flexibility, which allows taking users’ feedback to improve recognition performance and to add a new action instance without computationally expensive optimization for training classifiers. Our nearest neighbor-based action classifier adopts dynamic time warping to handle variability in execution rate. In addition, it uses the confidence values associated to each tracked joint position to mask erroneous trajectories for robustness against noise. We evaluate the proposed method with various datasets with different frame rates, actors, and noise. The experimental results demonstrate its adequacy for learning of actions from depth sequences at runtime. We achieve an accuracy comparable to the state-of-the-art on the challenging MSR-Action3D dataset.

Download here

Bibtex:

@article{tejero2016flexible,
  title={Flexible human action recognition in depth video sequences using masked joint trajectories},
  author={Tejero-de-Pablos, Antonio and Nakashima, Yuta and Yokoya, Naokazu and D{\'\i}az-Pernas, Francisco-Javier and Mart{\'\i}nez-Zarzuela, Mario},
  journal={EURASIP Journal on Image and Video Processing},
  volume={2016},
  number={1},
  pages={1--12},
  year={2016},
  publisher={Springer}
}