Skip to main content
Article
Expandable data-driven graphical modeling of human actions based on salient postures
IEEE Transactions on Circuits and Systems for Video Technology
  • Wanqing Li, University of Wollongong
  • Zhengyou Zhang, Microsoft Research
  • Zicheng Liu, Microsoft Research
Document Type
Journal Article
Publication Date
1-1-2008
Abstract

This paper presents a graphical model for learning and recognizing human actions. Specifically, we propose to encode actions in a weighted directed graph, referred to as action graph, where nodes of the graph represent salient postures that are used to characterize the actions and are shared by all actions. The weight between two nodes measures the transitional probability between the two postures represented by the two nodes. An action is encoded as one or multiple paths in the action graph. The salient postures are modeled using Gaussian mixture models (GMMs). Both the salient postures and action graph are automatically learned from training samples through unsupervised clustering and expectation and maximization (EM) algorithm. The proposed action graph not only performs effective and robust recognition of actions, but it can also be expanded efficiently with new actions. An algorithm is also proposed for adding a new action to a trained action graph without compromising the existing action graph. Extensive experiments on widely used and challenging data sets have verified the performance of the proposed methods, its tolerance to noise and viewpoints, its robustness across different subjects and data sets, as well as the effectiveness of the algorithm for learning new actions.

RIS ID
25640
Citation Information
Wanqing Li, Zhengyou Zhang and Zicheng Liu. "Expandable data-driven graphical modeling of human actions based on salient postures" IEEE Transactions on Circuits and Systems for Video Technology Vol. 18 Iss. 11 (2008) p. 1499 - 1510
Available at: http://works.bepress.com/wli/33/