Models of Gaze Control for Manipulation Tasks
https://doi.org/10.1145/0000000.0000000Abstract
Human studies have shown that gaze shifts are mostly driven by the current task demands. In manipulation tasks, gaze leads action to the next manipulation target. One explanation is that fixations gather information about task relevant properties, where task relevance is signalled by reward. This work presents new computational models of gaze shifting, where the agent imagines ahead in time the informational effects of possible gaze fixations. Building on our previous work, the contributions of this paper are: a) the presentation of two new gaze control models; b) comparison of their performance to our previous model; c) results showing the fit of all these models to previously published human data; and d) integration of a visual search process. The first new model selects the gaze that most reduces positional uncertainty of landmarks (Unc), and the second maximises expected rewards by reducing positional uncertainty (RU). Our previous approach maximises the expected gain in cumulative reward by reducing positional uncertainty (RUG). In experiment b) the models are tested on a simulated humanoid robot performing a manipulation task, and each model's performance is characterised by varying three environmental variables. This experiment provides evidence that the RUG model has the best overall performance. In experiment c) we compare the hand-eye coordination timings of the models in a robot simulation to those obtained from human data. This provides evidence that only the models that incorporate both uncertainty and reward (RU and RUG) match human data.
References (28)
- A:16 • J. Nunez-Varela and J. L. Wyatt A:20 • REFERENCES BAJCSY, R. 1988. Active perception. Proceedings of the IEEE 76, 8, 966-1005.
- BALLARD, D., HAYHOE, M., LI, F., AND WHITEHEAD, S. 1992. Hand-eye coordination during sequential tasks [and discussion].
- Phil. Trans. Royal Society of London. Series B: Biological Sciences 337, 1281, 331-339.
- BRADTKE, S. AND DUFF, M. 1995. Reinforcement learning methods for continuous-time Markov decision problems. Adv. in Neural Inf. Proc. Sys. 8, 393-400.
- CASSANDRA, A. R. 1998. Exact and approximate algorithms for partially observable markov decision processes. Ph.D. thesis, Brown Univ., Rhode Island.
- EREZ, T., TRAMPER, J., SMART, W., AND GIELEN, S. 2011. A POMDP model of eye-hand coordination. In Proc. 25th Conf. on AI. California, 952-957.
- FINDLAY, J. AND GILCHRIST, I. 2003. Active Vision. Oxford Univ. Press, Oxford.
- JOHANSSON, R., WESTLING, G., BACKSTROM, A., AND FLANAGAN, J. R. 2001. Eye-hand coordination in object manipulation. Journal of Neuroscience 21, 17, 6917-6932.
- JOHANSSON, R. S. AND FLANAGAN, J. R. 2009. Sensorimotor control of manipulation. Encyclopedia of Neuroscience 8, 593-604.
- LAND, M. 2009. Vision, eye movements, and natural behavior. Visual neuroscience 26, 1, 51-62.
- LEVENSHTEIN, V. 1966. Binary codes capable of correcting deletions insertions and reversals. Soviet Phys. Doklady 10, 707-10.
- METTA, G., SANDINI, G., VERNON, D., NATALE, L., AND NORI, F. 2008. The iCub humanoid robot: An open platform for research in embodied cognition. In Proc. ACM Performance Metrics for Int. Sys. MD, USA, 50-56.
- NAJEMNIK, J. AND GEISLER, W. 2005. Optimal eye movement strategies in visual search. Nature 434, 7031, 387-391.
- NAVALPAKKAM, V., KOCH, C., RANGEL, A., AND PERONA, P. 2010. Optimal reward harvesting in complex perceptual environ- ments. Proceedings of the National Academy of Sciences 107, 11, 5232-5237.
- NUNEZ-VARELA, J., WYATT, J., AND RAVINDRAN, B. 2012a. Gaze allocation analysis for a visually guided manipulation task. In Proceedings of SAB 2012. Denmark, 44-53.
- NUNEZ-VARELA, J., WYATT, J., AND RAVINDRAN, B. 2012b. Where do I look now? Gaze allocation during visually guided manipulation. In Proceedings of ICRA 2012. USA, 4444-4449.
- PATTACINI, U., NORI, F., NATALE, L., METTA, G., AND SANDINI, G. 2010. An experimental evaluation of a novel minimum-jerk cartesian controller for humanoid robots. In IEEE/RSJ Int. Conf. on Int. Robots, Sys. Taipei, Taiwan, 1668-1674.
- PUTERMAN, M. L. 1994. Markov Decision Processes: Discrete Stochastic Dynamic Programming. Wiley-Intersc., New York.
- RENNINGER, L., VERGHESE, P., AND COUGHLAN, J. 2010. Where to look next? Eye movements reduce local uncertainty. Journal of Vision 7, 3, 1-17.
- ROTHKOPF, C., BALLARD, D., AND HAYHOE, M. 2007. Task and context determine where you look. J. of Vision 7, 14, 1-20.
- SCHUTZ, A. AND GEGENFURTNER, K. 2010. Dynamic integration of saliency and reward information for saccadic eye move- ments. Journal of Vision 10, 7, 551-551.
- SHIBATA, T., VIJAYAKUMAR, S., CONRADT, J., AND SCHAAL, S. 2001. Biomimetic oculomotor control. Adaptive Behavior 9, 3-4, 189-207.
- SPRAGUE, N., BALLARD, D., AND ROBINSON, A. 2007. Modeling embodied visual behaviors. ACM TAP. 4, 2, 1-23.
- STEINMAN, R. 2003. Gaze control under natural conditions. In The Visual Neurosciences, L. Chalupa and J. Werner, Eds. MIT Press, 1339-1356.
- SULLIVAN, B., JOHNSON, L., ROTHKOPF, C., BALLARD, D., AND HAYHOE, M. 2012. The effect of uncertainty and reward on fixation behavior in a driving task. Journal of Vision 12, 9, 1259-1259.
- SUTTON, R. S. AND BARTO, A. G. 1998. Introduction to Reinforcement Learning. MIT Press Cambridge, Cambridge, MA.
- SUTTON, R. S., PRECUP, D., AND SINGH, S. 1999. Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. AI Journal 112, 1, 181-211.
- THRUN, S., BURGARD, W., AND FOX, D. 2008. Probabilistic Robotics. MIT Press Cambridge, Cambridge, MA. YARBUS, A. 1967. Eye movements and vision. Plenum Press, New York.