Gesture recognition
2001
Abstract
1. Introduction A primary goal of virtual environments (VE) is to provide natural, efficient, powerful, and flexible interaction. Gesture as an input modality can help meet these requirements. Human gestures are certainly natural and flexible, and may often be efficient and powerful, especially as compared with alternative interaction modes. This chapter will cover automatic gesture recognition, particularly computer vision based techniques that do not require the user to wear extra sensors, clothing or equipment.
References (52)
- References Anonymous. (1987). Visual Signals. U.S. Army Field Manual FM-2160. Available: http://155.217.58.58/atdls.html.
- Assan, M. and Grobel, K. (1997). Video-based sign language recognition using hidden Markov models. In I. Wachsmuth and M. Fröhlich (Eds.), Gesture and Sign Language in Human-Computer Interaction. Proc. International Gesture Workshop, Bielefeld, Germany.
- Baudel, T. and Beaudouin-Lafon, M. (1993). CHARADE: remote control of objects using free-hand gestures. Communications of the ACM (pp. 28-35), Vol. 36, No. 7.
- Black, M. and Yacoob, Y. (1995). Tracking and recognizing rigid and non-rigid facial motions using local parametric models of image motion. Proc. ICCV (pp. 374-381). Cambridge, MA.
- Bobick, A. (1997). Movement, activity, and action: the role of knowledge in the perception of motion. Royal Society Workshop on Knowledge-based Vision in Man and Machine. London, England.
- Böhm, K., Broll, W., and Solokewicz, M. (1994). Dynamic gesture recognition using neural networks; a fundament for advanced interaction construction. In S. Fisher, J. Merrit, and M. Bolan (Eds.), Stereoscopic Displays and Virtual Reality Systems . SPIE Conference on Electronic Imaging Science and Technology, Vol. 2177, San Jose, CA.
- Bolt, R. A. (1980). Put-That-There: Voice and gesture at the graphics interface. Computer Graphics, 14.3 (pp. 262-270).
- Boyd, J. and Little, J. (1998). Shape of Motion and the Perception of Human Gaits. IEEE Workshop on Empirical Evaluation Methods in Computer Vision. CVPR 98, Santa Barbara, CA.
- Cadoz, C. (1994). Les réalités virtuelles. Dominos, Flammarion, 1994.
- Cassell, J., Steedman, M., Badler, N., Pelachaud, C., Stone, M., Douville, B., Prevost, S., and Achorn, B. (1994). Modeling the interaction between speech and gesture. Proceedings of the Sixteenth Conference of the Cognitive Science Society.
- Cohen, P. R., Johnston, M., McGee, D., Oviatt, S., Pittman, J., Smith, I., Chen, L., and Clow, J. (1997). QuickSet: Multimodal interaction for distributed applications, Proceedings of the Fifth Annual International Multimodal Conference (pp 31-40), Seattle, WA, ACM Press.
- Cutler, R. and Turk, M. (1998). View-based interpretation of real-time optical flow for gesture recognition. Proc. Third International Conference on Automatic Face and Gesture Recognition. Nara, Japan.
- Davis, J. and Bobick, A. (1997). The representation and recognition of human movement using temporal trajectories. Proc. IEEE Conference on Computer Vision and Pattern Recognition. Puerto Rico.
- Ekman, P. and Friesen, W. V. (1978). Facial action coding system: A technique for the measurement of facial movement. Palo Alto, Calif.: Consulting Psychologists Press.
- Essa, I. and Pentland, A. (1997). Coding, Analysis, Interpretation and Recognition of Facial Expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence,Vol. 19 (7), IEEE Computer Society Press.
- Fels, S. and Hinton, G. (1995). Glove-TalkII: An Adaptive Gesture-to-Formant Interface. CHI'95. Denver, CO.
- Freeman, W., Tanaka, K., Ohta, J., and Kyuma, K. (1996). Computer vision for computer games. Proc. Second International Conference on Automatic Face and Gesture Recognition. Killington, VT.
- Haritaoglu, I., Harwood, D., and Davis, L. (1998). W4: Who? When? Where? What? A real time system for detecting and tracking people. Proc. Third International Conference on Automatic Face and Gesture Recognition. Nara, Japan.
- Huang, T. and Pavlovic, V. (1995). Hand gesture modeling, analysis, and synthesis. Proc. International Workshop on Automatic Face-and Gesture-Recognition. Zurich.
- Hummels, C. and Stappers, P (1998). Meaningful gestures for human computer interaction: beyond hand gestures. Proc. Third International Conference on Automatic Face and Gesture Recognition. Nara, Japan.
- Johnson, T. (1963). Sketchpad III: Three Dimensional Graphical Communication with a Digital Computer. AFIPS Spring Joint Computer Conference, 23. (pp. 347-353).
- Kadous, W. (1996). Computer recognition of Auslan signs with PowerGloves. Proc. Workshop on the Integration of Gesture in Language and Speech, Wilmington, DE.
- Kendon, A. (1972). Some relationships between body motion and speech. In A. W. Siegman and B. Pope (Eds.), Studies in Dyadic Communication, New York, Pergamon Press.
- La Cascia, M., Isidoro, J., and Sclaroff, S. (1998). Head tracking via robust registration in texture map images. Proc. IEEE Conference on Computer Vision and Pattern Recognition. Santa Barbara, CA.
- Landay, J. A. and Myers, B. A. (1995). Interactive sketching for the early stages of user interface design. Proceedings of CHI'95 (pp. 43-50).
- Latoschik, M. and Wachsmuth, I. (1997). Exploiting distant pointing gestures for object selection in a virtual environment. In I. Wachsmuth, and M. Fröhlich (Eds.), Gesture and Sign Language in Human-Computer Interaction. International Gesture Workshop, Bielefeld, Germany.
- Lee, C. and Xu, Y. (1996). Online, Interactive Learning of Gestures for Human/Robot Interfaces. 1996 IEEE International Conference on Robotics and Automation, Vol. 4 (pp. 2982-2987). Minneapolis, MN.
- Long, A., Landay, J., and Rowe, L. (1998). PDA and Gesture Uses in Practice: Insights for Designers of Pen- Based User Interfaces. Report #CSD-97-976, CS Division, EECS Department. UC Berkeley. Berkeley, CA.
- Lucente, M., Zwart, G., and George, A. (1998). Visualization Space: a testbed for deviceless multimodal user interface. Intelligent Environments Symposium. AAAI Spring Symposium Series. Stanford, CA.
- Maggioni, C. and Kämmerer, B. (1998). GestureComputer -history, design and applications. In R. Cipolla and A. Pentland (Eds.), Computer Vision for Human-Machine Interaction. Cambridge University Press.
- Marrin, T. and Picard, R. (1998). The Conductor's Jacket: a Testbed for Research on Gestural and Affective Expression. XII Colloquium for Musical Informatics, Gorizia, Italy.
- McNeill. D. (1992). Hand and Mind: What Gestures Reveal about Thought. Chicago: University of Chicago Press.
- Moses, Y., Reynard, D., and Blake, A. (1995). Determining facial expressions in real time. Proc. Fifth International Conference on Computer Vision. Cambridge, MA.
- Mulder, A. (1996). Hand gestures for HCI. Technical Report 96-1, School of Kinesiology. Simon Fraser University.
- Oliver, N., Pentland, A., and Bérard, F. (1997). LAFTER: Lips and face real time tracker. Proc. IEEE Conference on Computer Vision and Pattern Recognition, Puerto Rico.
- Otsuka, T. and Ohya, J. (1998). Recognizing abruptly changing facial expressions from time-sequential face images. Proc. IEEE Conference on Computer Vision and Pattern Recognition. Santa Barbara, CA.
- Oviatt, S. L. (1996). Multimodal interfaces for dynamic interactive maps. Proceedings of CHI'96 Human Factors in Computing Systems (pp. 95-102). ACM Press, NY.
- Paradiso, J. and Sparacino, F. (1997). Optical Tracking for Music and Dance Performance. Fourth Conference on Optical 3-D Measurement Techniques. Zurich, Switzerland.
- Pavlovic.
- V, Sharma, R., and Huang, T. (1996). Gestural interface to a visual computing environment for molecular biologists. Proc. Second International Conference on Automatic Face and Gesture Recognition. Killington, VT.
- Stark, M. and Kohler, M. (1995). Video based gesture recognition for human computer interaction. In W. D. Fellner (Ed.), Modeling -Virtual Worlds -Distributed Graphics.
- Starner, T. and Pentland, A. (1995). Visual recognition of American Sign Language using hidden Markov models. Proc. International Workshop on Automatic Face-and Gesture-Recognition. Zurich.
- Sturman, J. (1992). Whole-hand Input. Ph.D. Thesis, MIT Media Laboratory. Cambridge, MA.
- Takahashi, T. and Kishino, F. (1991). Gesture coding based in experiments with a hand gesture interface device. SIGCHI Bulletin, 23(2), (pp. 67-73).
- Triesch, J. and von der Malsburg, C. (1996). Robust classification of hand postures against complex backgrounds. Proc. Second International Conference on Automatic Face and Gesture Recognition. Killington, VT.
- Väänänen, K. and Böhm, K. (1992). Gesture driven interaction as a human factor in virtual environments -an approach with neural networks. Proc. Virtual Reality Systems . British Computer Society, Academic Press.
- Weng, J. and Cui, Y. (1998). Recognition of hand signs from complex backgrounds. In R. Cipolla and A. Pentland (Eds.), Computer Vision for Human-Machine Interaction. Cambridge University Press.
- Wexelblat, A. (1994). A Feature-Based Approach to Continuous-Gesture Analysis. M.S. Thesis, MIT Media Laboratory, Cambridge, MA.
- Wren, C., Azarbayejani, A., Darrell, T., and Pentland, A. (1996). Pfinder: real-time tracking of the human body. Proc. Second International Conference on Automatic Face and Gesture Recognition. Killington, VT.
- Yamamoto, J., Kondo, T., Yamagiwa, T., and Yamanaka, K. (1998). Skill recognition. Proc. Third International Conference on Automatic Face and Gesture Recognition. Nara, Japan.
- Zeleznik, R.C., Herndon, K.P., and Hughes J.F. (1996). Sketch: An interface for sketching 3D scenes. Computer Graphics (Proceedings of SIGGRAPH '96).
- Zelinsky, A. and Heinzmann, J. (1996). Real-time visual recognition of facial gestures for human-computer interaction. Proc. Second International Conference on Automatic Face and Gesture Recognition. Killington, VT.