UNIVERSITY OF CALIFORNIA,
SANTA CRUZ

 
 
 
 
Generic Human Action Recognition from a Single Example
 
 

Hae Jong Seo and Peyman Milanfar

 
 
Abstract
 
 

We present a novel human action recognition method based on space-time locally adaptive regression kernels and the matrix cosine similarity measure. The proposed method operates using a single example (e.g., short video clip) of an action of interest to find similar matches. It does not require prior knowledge (learning) about actions being sought; and does not require foreground/background segmentation, or any motion estimation or tracking. Our method is based on the computation of the so-called local steering kernels as space-time descriptors from a query video, which measure the likeness of a voxel to its surroundings. Salient features are extracted from said descriptors and compared against analogous features from the target video. This comparison is done using a matrix generalization of the cosine similarity measure. The algorithm yields a scalar resemblance volume with each voxel here, indicating the likelihood of similarity between the query video and all cubes in the target video. By employing nonparametric significance tests and non-maxima suppression, we detect the presence and location of actions similar to the given query video. High performance is demonstrated on the challenging set of action data indicating successful detection of actions in the presence of fast motion, different contexts and even when multiple complex actions occur simultaneously within the field of view of the camera. Further experiments on the Weizmann dataset and the KTH dataset for action categorization task demonstrate that the proposed method achieves improvement over other (state-of-the-art) algorithms.

  • System Overview
  • Experimental Results
  • Matlab implementation
  • See more details and examples in the following papers .

  • Go to object detection page

  • Go to saliency detection page

  •    
     

    System Overview

     
     
     
     

     
     

    Results on the Shechtman's Action Dataset

     
     
    (a) Query video
    (b) Target Video
    (c) Resemblance Volume
    (d) Detected Actions



     
     

    (a) Query video
    (b) Target Video
    (c) Resemblance Volume
    (d) Detected Actions



     

    (a) Multiple Query videos
    (b) Target Video
    (c) Resemblance Volumes



     
     
     
     
     
     
     

    Matlab Package

     
     
    The package contains the software that can be used to detect action using a single query, as explained in the PAMI paper above. The included demonstration files (demo*.m) provide the easiest way to learn how to use the code.

    Disclaimer: This is experimental software. It is provided for non-commercial research purposes only. Use at your own risk. No warranty is implied by this distribution. Copyright © 2010 by University of California.

    File updated: June 2011




     
    Acknowledgement
     
     
    This work was supported in part by the US Air Force Grant F49550-07-1-0365.
     
     
     
     
    last update on June 10th, 2009