• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Generalized Hough transform based time invariant action recognition with 3D pose information
 
  • Details
  • Full
Options
2014
Conference Paper
Title

Generalized Hough transform based time invariant action recognition with 3D pose information

Abstract
Human action recognition has emerged as an important field in the computer vision community due to its large number of applications such as automatic video surveillance, content based video-search and human robot interaction. In order to cope with the challenges that this large variety of applications present, recent research has focused more on developing classifiers able to detect several actions in more natural and unconstrained video sequences. The invariance discrimination tradeoff in action recognition has been addressed by utilizing a Generalized Hough Transform. As a basis for action representation we transform 3D poses into a robust feature space, referred to as pose descriptors. For each action class a one-dimensional temporal voting space is constructed. Votes are generated from associating pose descriptors with their position in time relative to the end of an action sequence. Training data consists of manually segmented action sequences. In the detection phase valid human 3D poses are assumed as input, e.g. originating from 3D sensors or monocular pose reconstruction methods. The human 3D poses are normalized to gain view-independence and transformed into (i) relative limb-angle space to ensure independence of non-adjacent joints or (ii) geometric features. In (i) an action descriptor consists of the relative angles between limbs and their temporal derivatives. In (ii) the action descriptor consists of different geometric features. In order to circumvent the problem of time-warping we propose to use a codebook of prototypical 3D poses which is generated from sample sequences of 3D motion capture data. This idea is in accordance with the concept of equivalence classes in action space. Results of the codebook method are presented using the Kinect sensor and the CMU Motion Capture Database.
Author(s)
Münch, David  
Hübner, Wolfgang  
Arens, Michael  
Mainwork
Optics and Photonics for Counterterrorism, Crime Fighting, and Defence X and Optical Materials and Biomaterials in Security and Defence Systems Technology XI  
Conference
Conference "Optics and Photonics for Counterterrorism, Crime Fighting, and Defence" 2014  
Conference "Optical Materials and Biomaterials in Security and Defence Systems Technology" 2014  
Open Access
File(s)
Download (1.06 MB)
Rights
Use according to copyright law
DOI
10.1117/12.2065805
10.24406/publica-r-385361
Language
English
Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB  
Keyword(s)
  • action recognition

  • Generalized Hough transform

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024