Poselet Key-framing: A Model for Human Activity Recognition

 

In this paper, we develop a new model for recognizing human actions.

June 23, 2013
IEEE Conference on Computer Vision Pattern Recognition (CVPR) 2013

 

Authors

Michalis Raptis (Disney Research)

Leonid Sigal (Disney Research)

Poselet Key-framing: A Model for Human Activity Recognition

Abstract

An action is modeled as a very sparse sequence of temporally local discriminative key frames – collections of partial key-poses of the actor(s), depicting key states in the action sequence. We cast the learning of key frames in a max-margin discriminative framework, where we treat key frames as latent variables. This allows us to (jointly) learn a set of most discriminative key frames while also learning the local temporal context between them. Key frames are encoded using a spatially-localizable pose let-like representation with HoG and BoW components learned from weak annotations, we rely on structured SVM formulation to align our components and mine for hard negatives to boost localization performance. This results in a model that supports spatio-temporal localization and is insensitive to dropped frames or partial observations. We show classification performance that is competitive with the state of the art on the benchmark UT-Interaction dataset and illustrate that our model outperforms prior methods in an on-line streaming setting.

Copyright Notice