Academia.eduAcademia.edu

Outline

All Together Now

2013, Lecture Notes in Computer Science

https://doi.org/10.1007/978-3-642-40415-3_33

Abstract

While virtual humans are proven tools for training, education and research, they are far from realizing their full potential. Advances are needed in individual capabilities, such as character animation and speech synthesis, but perhaps more importantly, fundamental questions remain as to how best to integrate these capabilities into a single framework that allows us to efficiently create characters that can engage users in meaningful and realistic social interactions. This integration requires in-depth, inter-disciplinary understanding few individuals, or even teams of individuals, possess. We help address this challenge by introducing the ICT Virtual Human Toolkit 1 , which offers a flexible framework for exploring a variety of different types of virtual human systems, from virtual listeners and question-answering characters to virtual role-players. We show that due to its modularity, the Toolkit allows researchers to mix and match provided capabilities with their own, lowering the barrier of entry to this multidisciplinary research challenge.

References (45)

  1. Rossen B., Lok B., A crowdsourcing method to develop virtual human conversational agents, International Journal of HCS, pp. 301-319 (2012)
  2. Bickmore, T.W., Bukhari L., Pfeiffer L., Paasche-Orlow M., Shanahan C., Hospital Bud- dy: A Persistent Emotional Support Companion Agent for Hospital Patients, Springer Ber- lin/Heidelberg, pp. 492-495 (2012)
  3. D'Mello, S. K., & Graesser, A. C. AutoTutor and affective AutoTutor: Learning by talking with cognitively and emotionally intelligent computers that talk back. ACM Transactions on Interactive Intelligent Systems, Volume 2, Issue 4, Article 23 (2012)
  4. Lane H.C., Noren,D., Auerbach D., Birch M., Swartout W., Intelligent Tutoring Goes to the Museum in the Big City: A Pedagogical Agent for ISE, Artificial Intelligence in Edu- cation, pp. 155-162 (2011)
  5. Johnson, W.L., Valente, A., Tactical Language and Culture Training Systems: Using AI to Teach Foreign Languages and Cultures. pp. 72-83 (2009)
  6. Campbell J., Core, M., Artstein R., Armstrong L., Hartholt A., Wilson C., Georgila K., Morbini F., Haynes E., Gomboc D., Birch M., Bobrow J., Lane H., Gerten J., Leuski A., Traum D., Trimmer M., DiNinni R., Bosack M., Jones T., Clark R., Yates K. Developing INOTS to support interpersonal skills practice. In Proceedings of the Thirty-second Annu- al IEEE Aerospace Conference, pp. 1-14, (2011)
  7. Pocketsphinx: A Free, Real-Time Continuous Speech Recognition System for hand-Held Devices, vol. 1, pp. 185-188 (2006)
  8. Littlewort G., Whitehill J., Wu T., Fasel I., Frank M., Movellan J., and Bartlett M., The Computer Expression Recognition Toolbox (CERT). Proc. IEEE International Conference on Automatic Face and Gesture Recognition (2011)
  9. Bickmore, T.W., Schulman D., Shaw, G., DTask & LiteBody: Open Source, Standards- based Tools for Building Web-deployed Embodied Conversational Agents, Intelligent Vir- tual Agents, PP. 425-431 (2009)
  10. Stone M., Specifying Generation of Referring Expressions by Example. AAAI Spring Symposium on NLG in Spoken and Written Dialogue, pp. 133-140 (2003)
  11. I.Poggi, C.Pelachaud, F. de Rosis, V. Carofiglio, B. De Carolis, GRETA. A Believable Embodied Conversational Agent, Multimodal Intelligent Information Presentation, (2005)
  12. Van Welbergen, H., Reidsma, D., Ruttkay, Z.M., Zwiers, J.: Elckerlyc: A BML realizer for continuous, multimodal interaction with a virtual human. Multimodal UI (2010)
  13. Shapiro, A.: Building a Character Animation System. In: The Fourth International Confer- ence on Motion in Games, Edinburgh, UK, November, 2011
  14. Heloir, A. and Kipp, M.: A Realtime Engine for Interactive Embodied Agents, Intelligent Virtual Agents, pp. 393-404, 2009
  15. Taylor, P., Black, A., Caley, R., The architecture of the Festival speech synthesis system. Third ESCA Workshop in Speech Synthesis, pp. 147-151. (1998)
  16. Heylen, D.K.J., Kopp, S., Marsella, S.C., Pelachaud, C., Vilhjálmsson, H.H. The Next Step towards a Function Markup Language. IVA (2008)
  17. Kopp, S., Krenn, B., Marsella, S., Marshall, A., Pelachaud, C., Pirker, H., Th´orisson, K., Vilhj´almsson, H., Towards a Common Framework for Multimodal Generation: The Be- havior Markup Language (LNAI), vol. 4133, pp. 205-217. (2006)
  18. H. van Welbergen, Y. Xu, M. Thiebaux, W.W. Feng, J. Fu, D. Reidsma and A. Shapiro Demonstrating and Testing the BML Compliance of BML Realizers. IVA (2011)
  19. Schröder, M. (2010). The SEMAINE API: Towards a standards-based framework for building emotion-oriented systems. Advances in Human-Machine Interaction (2010)
  20. Quigley, M., Conley, K., Gerkey, B.P., Faust, J., Foote, T., Leibs, J., Wheeler, R., and Ng, A.Y., ROS: an open-source Robot Operating System, ICRA Open Source Software, (2009)
  21. Leuski, A and Traum, D. NPCEditor: Creating virtual human dialogue using information retrieval techniques. AI Magazine, 32(2):42-56, (2011).
  22. Leuski, A., Traum D.. A statistical approach for text processing in virtual humans. In Pro- ceedings of the 26th Army Science Conference, Orlando, Florida, USA, December 2008.
  23. Lee, J. and S. Marsella. Nonverbal Behavior Generator for Embodied Conversational Agents, IVA (2006)
  24. Feng, A.W., Xu, Y, Shapiro, A. : An Example-Based Motion Synthesis Technique for Lo- comotion and Object Manipulation, SIGGRAPH (2012)
  25. Thiebaux, M., Lance, B., Marsella, S.: Real-time Expressive Gaze for Virtual Humans, AAMAS, vol. 1, pp. 321-328 (2008)
  26. Kallmann, M., Marsella, S.: Hierarchical Motion Controllers for Real-time Autonomous Virtual Humans, Intelligent Virtual Agents, pp. 253-265, (2005)
  27. Feng, A.W., Huang, Y, Shapiro, A., An Analysis of Motion Blending Techniques, The Fifth International Conference on Motion in Games, Rennes, France, November 2012
  28. Feng, A.W., Huang, Y, Xu, Y, Shapiro, A., Automating the Transfer of a Generic Set of Behaviors Onto a Virtual Character, Conference on Motion in Games, (2012)
  29. J. Wagner, F. Lingenfelser, N. Bee, and E. Andre. Social signal interpretation (ssi). KI - Kuenstliche Intelligenz, 25:251-256, (2011)
  30. T. Baltrusaitis, P. Robinson, and L.-P. Morency. 3D constrained local model for rigid and non-rigid facial tracking. IEEE Computer Vision and Pattern Recognition, June (2012)
  31. L.-P. Morency, J. Whitehill, and J. Movellan. Generalized adaptive view-based appearance model: Integrated framework for monocular head pose estimation. Automatic Face and Gesture Recognition, pp. 1-8, (2008)
  32. E. Suma, B. Lange, A. Rizzo, D. Krum, and M. Bolas, "FAAST: The Flexible Action and Articulated Skeleton Toolkit," Proceedings of IEEE Virtual Reality, pp. 247-248, (2011).
  33. Scherer, S., Marsella, S., Stratou G., Xu X., Morbini F., Egan A., Rizzo A., Morency, L.P., Perception Markup Language: Towards a Standardized Representation of Perceived Non- verbal Behaviors, IVA pp. 455-463 (2012)
  34. Leuski A, Kennedy B, Patel R, Traum DR. Asking questions to limited domain virtual characters: how good does speech recognition have to be? ASC (2006)
  35. Artstein R, Gandhe S, Leuski A, Traum DR. Field Testing of an interactive question- answering character. ELRA, LREC (2008)
  36. Swartout W, Traum DR, Artstein R, Noren D, Debevec P, Bronnenkant K, et al. Ada and Grace: Toward Realistic and Engaging Virtual Museum Guides. IVA pp. 286-300 (2010)
  37. Jonathan Gratch, Anna Okhmatovskaia, Francois Lamothe, Stacy Marsella, Mathieu Mo- rales, R. J. van der Werf, Louis-Philippe Morency, Virtual Rapport, IVA (2006)
  38. Rizzo A., Forbell E., Lange B., Buckwalter J.G., Williams J., Sagae K., Traum D., SimCoach: An Online Intelligent Virtual Agent System for Breaking Down Barriers to Care for Service Members and Veterans, Chapter in Healing War Trauma (2012) 45.
  39. Morbini F., DeVault D., Sagae K., Gerten J., Nazarian A., Traum D., FLoReS: A Forward Looking, Reward Seeking, Dialogue Manager, Spoken Dialog Systems (2012)
  40. Hartholt A., Gratch J., Weiss L., Leuski A., Morency L.P., Marsella S, At the Virtual Fron- tier: Introducing Gunslinger, a Multi-Character, Mixed-Reality, Story-Driven Experience IVA pp. 500-501 (2009)
  41. Kenny, P.G., Parsons, T.D., Gratch, J. Rizzo, A., Evaluation of Justina: A Virtual Patient with PTSD, Lecture Notes in Computer Science, pp. 394-408 (2008)
  42. Traum DR, Marsella S, Gratch J, Lee J, Hartholt A. Multi-party, Multi-issue, Multi- strategy Negotiation for Multi-modal Virtual Agents. IVA pp. 117-130(2008)
  43. Gratch J., Hartholt A., Dehghani M., Marsella S., Virtual Humans: A New Toolkit for Cognitive Science Research, CogSci (2013)
  44. Khooshabeh P., McCall C., Gandhe S., Gratch J., Blascovich J., Does it matter if a com- puter jokes, Extended abstracts on Human Factors in Computer Systems, pp. 77-86 (2011)
  45. Batrinca, L., Stratou, G., Morency, L.P., Scherer, S. Cicero -Towards a Multimodal Virtu- al Audience Platform for Public Speaking Training. IVA (2013)