Learning unknown event models
Abstract
Agents with incomplete environment models are likely to be surprised, and this represents an opportunity to learn. We investigate approaches for situated agents to detect surprises, discriminate among different forms of surprise, and hypothesize new models for the unknown events that surprised them. We instantiate these approaches in a new goal reasoning agent (named FOOLMETWICE), investigate its performance in simulation studies, and report that it produces plans with significantly reduced execution cost in comparison to not learning models for surprising events.
References (21)
- Bridewell, W., Langley, P., Todorovski, L., & Džeroski, S. (2008). Inductive process modeling. Machine learning, 71(1), 1- 32. Ghallab, M., Nau, D., & Traverso, P. (2004). Automated planning: Theory & practice. San Mateo, CA: Morgan Kaufmann.
- Goldman, W. (1973). The princess bride. San Diego, CA: Harcourt Brace.
- Hiatt, L.M., Khemlani, S.S., & Trafton, J.G. (2012). An explanatory reasoning framework for embodied agents. Biologically Inspired Cognitive Architectures, 1, 23-31.
- Klenk, M., Molineaux, M., & Aha, D.W. (2013). Goal-driven autonomy for responding to unexpected events in strategy simulations. Computational Intelligence, 29(2), 187-206.
- Leake, D. B. (1991), Goal-based explanation evaluation. Cognitive Science, 15, 509-545.
- Molineaux, M., Aha, D.W., & Kuter, U. (2011). Learning event models that explain anomalies. In T. Roth-Berghofer, N. Tintarev, & D.B. Leake (Eds.) Explanation-Aware Computing: Papers from the IJCAI Workshop. Barcelona, Spain.
- Molineaux, M., Klenk, M., & Aha, D.W. (2010a). Goal-driven autonomy in a Navy strategy simulation. In Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence. Atlanta, GA: AAAI Press.
- Molineaux, M., Klenk, M., & Aha, D.W. (2010b). Planning in dynamic environments: Extending HTNs with nonlinear continuous effects. In Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence. Atlanta, GA: AAAI Press.
- Molineaux, M., Kuter, U., & Klenk, M. (2012). DiscoverHistory: Understanding the past in planning and execution. Proceedings of the Eleventh International Conference on Autonomous Agents and Multiagent Systems (Volume 2) (pp. 989-996). Valencia, Spain: International Foundation for Autonomous Agents and Multiagent Systems.
- Mourao, K., Zettlemoyer, L. S., Petrick, R., & Steedman, M. (2012). Learning strips operators from noisy and incomplete observations. arXiv preprint arXiv:1210.4889.
- Nau, D.S. (2007). Current trends in automated planning. AI Magazine, 28(4), 43-58.
- Nau, D., Au, T.-C., Ilghami, O, Kuter, U, Murdock, J.W., Wu, D., & Yaman, F. (2003). SHOP2: An HTN planning system. Journal of Artificial Intelligence Research, 20, 379-404.
- Nguyen, T.H.D., & Leong, T.Y. (2009). A surprise triggered adaptive and reactive (STAR) framework for online adaptation in non-stationary environments. In Proceedings of the Fifth Artificial Intelligence and Interactive Digital Entertainment Conference. Stanford, CA: AAAI Press.
- Pang, W., & Coghill, G.M. (2010). Learning qualitative differential equation models: A survey of algorithms and applications. Knowledge Engineering Review, 25(1), 69-107.
- Pasula, H. M., Zettlemoyer, L. S., & Kaelbling, L. P. (2007). Learning Symbolic Models of Stochastic Domains. J. Artif. Intell. Res.(JAIR), 29, 309-352.
- Quinlan, J. R. (1990). Learning logical definitions from relations. Machine learning, 5(3), 239-266.
- Ranasinghe, N., & Shen, W.-M. (2008). Surprised-based learning for developmental robotics. Proceedings of the ECSIS Symposium on Learning and Adaptive Behaviors for Robotic Systems (pp. 65- 70). Edinburgh, Scotland: IEEE Press.
- Sutton, R.S., & Barto, A.G. (1998). Reinforcement learning: An introduction. Cambridge, MA: MIT Press.
- Weber, B., Mateas, M., & Jhala, A. (2012). Learning from demonstration for goal-driven autonomy. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence. Toronto, Canada: AAAI Press.
- Zhuo, H.H., Hu, D.H., Hogg, C., Yang, Q., & Muñoz-Avila, H. (2009). Learning HTN method preconditions and action models from partial observations. Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence (pp. 1804-1810). Pasadena, CA: AAAI Press.
- Zhuo, H. H., Yang, Q., Hu, D. H., & Li, L. (2010). Learning complex action models with quantifiers and logical implications. Artificial Intelligence, 174(18), 1540-1569.