Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

Tracking in object action space

Kruger, Volker LU orcid and Herzog, Dennis (2013) In Computer Vision and Image Understanding 117(7). p.764-789
Abstract

In this paper we focus on the joint problem of tracking humans and recognizing human action in scenarios such as a kitchen scenario or a scenario where a robot cooperates with a human, e.g., for a manufacturing task. In these scenarios, the human directly interacts with objects physically by using/manipulating them or by, e.g., pointing at them such as in "Give me that...", 10 recognize these types of human actions is difficult because (a) they ought to be recognized independent of scene parameters such as viewing direction and (b) the actions are parametric, where the parameters are either object-dependent or as, e.g., in the case of a pointing direction convey important information. One common way to achieve recognition is by using 3D... (More)

In this paper we focus on the joint problem of tracking humans and recognizing human action in scenarios such as a kitchen scenario or a scenario where a robot cooperates with a human, e.g., for a manufacturing task. In these scenarios, the human directly interacts with objects physically by using/manipulating them or by, e.g., pointing at them such as in "Give me that...", 10 recognize these types of human actions is difficult because (a) they ought to be recognized independent of scene parameters such as viewing direction and (b) the actions are parametric, where the parameters are either object-dependent or as, e.g., in the case of a pointing direction convey important information. One common way to achieve recognition is by using 3D human body tracking followed by action recognition based on the captured tracking data. For the kind of scenarios considered here we would like to argue that 3D body tracking and action recognition should be seen as an intertwined problem that is primed by the objects on which the actions are applied. In this paper, we are looking at human body tracking and action recognition from a object-driven perspective. Instead of the space of human body poses we consider the space of the object affordances, i.e., the space of possible actions that are applied on a given object. This way, 3D body tracking reduces to action tracking in the object (and context) primed parameter space of the object affordances. This reduces the high- dimensional joint-space to a low-dimensional action space. In our approach, we use parametric hidden Markov models to represent parametric movements; particle filtering is used to track in the space of action parameters. We demonstrate its effectiveness on synthetic and on real image sequences using human-upper body single arm actions that involve objects.

(Less)
Please use this url to cite or link to this publication:
author
and
publishing date
type
Contribution to journal
publication status
published
keywords
Action recognition, Parametric gestures, Pose estimation, Tracking
in
Computer Vision and Image Understanding
volume
117
issue
7
pages
26 pages
publisher
Elsevier
external identifiers
  • scopus:85028093142
ISSN
1077-3142
DOI
10.1016/j.cviu.2013.02.002
language
English
LU publication?
no
id
c14ce51b-6a9e-4503-bff0-0c785c3b91eb
date added to LUP
2019-06-28 09:19:18
date last changed
2022-02-15 21:57:37
@article{c14ce51b-6a9e-4503-bff0-0c785c3b91eb,
  abstract     = {{<p>In this paper we focus on the joint problem of tracking humans and recognizing human action in scenarios such as a kitchen scenario or a scenario where a robot cooperates with a human, e.g., for a manufacturing task. In these scenarios, the human directly interacts with objects physically by using/manipulating them or by, e.g., pointing at them such as in "Give me that...", 10 recognize these types of human actions is difficult because (a) they ought to be recognized independent of scene parameters such as viewing direction and (b) the actions are parametric, where the parameters are either object-dependent or as, e.g., in the case of a pointing direction convey important information. One common way to achieve recognition is by using 3D human body tracking followed by action recognition based on the captured tracking data. For the kind of scenarios considered here we would like to argue that 3D body tracking and action recognition should be seen as an intertwined problem that is primed by the objects on which the actions are applied. In this paper, we are looking at human body tracking and action recognition from a object-driven perspective. Instead of the space of human body poses we consider the space of the object affordances, i.e., the space of possible actions that are applied on a given object. This way, 3D body tracking reduces to action tracking in the object (and context) primed parameter space of the object affordances. This reduces the high- dimensional joint-space to a low-dimensional action space. In our approach, we use parametric hidden Markov models to represent parametric movements; particle filtering is used to track in the space of action parameters. We demonstrate its effectiveness on synthetic and on real image sequences using human-upper body single arm actions that involve objects.</p>}},
  author       = {{Kruger, Volker and Herzog, Dennis}},
  issn         = {{1077-3142}},
  keywords     = {{Action recognition; Parametric gestures; Pose estimation; Tracking}},
  language     = {{eng}},
  month        = {{01}},
  number       = {{7}},
  pages        = {{764--789}},
  publisher    = {{Elsevier}},
  series       = {{Computer Vision and Image Understanding}},
  title        = {{Tracking in object action space}},
  url          = {{http://dx.doi.org/10.1016/j.cviu.2013.02.002}},
  doi          = {{10.1016/j.cviu.2013.02.002}},
  volume       = {{117}},
  year         = {{2013}},
}