Anticipating human activities for reactive robotic response
2013
https://doi.org/10.1109/IROS.2013.6696634Abstract
An important aspect of human perception is anticipation, which we use extensively in our day-today activities when interacting with other humans as well as with our surroundings. Anticipating which activities will a human do next (and how to do them) in useful for many applications, for example, anticipation enables an assistive robot to plan ahead for reactive responses in the human environments. In this work, we represent each possible future using an anticipatory temporal conditional random field (ATCRF) that models the rich spatial-temporal relations through object affordances. We then consider each ATCRF as a particle and represent the distribution over the potential futures using a set of particles. In extensive evaluation on CAD-120 human activity RGB-D dataset, for new subjects (not seen in the training set), we obtain an activity anticipation accuracy (defined as whether one of top three predictions actually happened) of 75.4%, 69.2% and 58.1% for an anticipation time of 1, 3 and 10 seconds respectively. Finally, we also use our algorithm on a robot for performing a few reactive responses. 1
References (9)
- H. S. Koppula and A. Saxena, "Anticipating human activities using object affordances for reactive robotic response," in RSS, 2013.
- K. Tang, L. Fei-Fei, and D. Koller, "Learning latent temporal structure for complex event detection," in CVPR, 2012.
- M. Rohrbach, S. Amin, M. Andriluka, and B. Schiele, "A database for fine grained activity detection of cooking activities," in CVPR, 2012.
- H. Pirsiavash and D. Ramanan, "Detecting activities of daily living in first-person camera views," in CVPR, 2012.
- J.-K. Min and S.-B. Cho, "Activity recognition based on wearable sensors using selection/fusion hybrid ensemble," in SMC, 2011.
- H. S. Koppula, R. Gupta, and A. Saxena, "Learning human activities and object affordances from rgb-d videos," IJRR, 2013.
- J. Sung, C. Ponce, B. Selman, and A. Saxena, "Unstructured human activity detection from rgbd images," in ICRA, 2012.
- B. Ni, G. Wang, and P. Moulin, "Rgbd-hudaact: A color-depth video database for human daily activity recognition," in ICCV Workshop on CDC4CV, 2011.
- M. Montemerlo, S. Thrun, and W. Whittaker, "Conditional particle filters for simultaneous mobile robot localization and people-tracking," in ICRA, 2002.