Academia.eduAcademia.edu

Outline

Multimodal sensing-based camera applications

2011, Proceedings of SPIE- The …

https://doi.org/10.1117/12.871934

Abstract

The increased sensing and computing capabilities of mobile devices can provide for enhanced mobile user experience. Integrating the data from different sensors offers a way to improve application performance in camera-based applications. A key advantage of using cameras as an input modality is that it enables recognizing the context. Therefore, computer vision has been traditionally utilized in user interfaces to observe and automatically detect the user actions. The imaging applications can also make use of various sensors for improving the interactivity and the robustness of the system. In this context, two applications fusing the sensor data with the results obtained from video analysis have been implemented on a Nokia Nseries mobile device. The first solution is a real-time user interface that can be used for browsing large images. The solution enables the display to be controlled by the motion of the user's hand using the built-in sensors as complementary information. The second application is a real-time panorama builder that uses the device's accelerometers to improve the overall quality, providing also instructions during the capture. The experiments show that fusing the sensor data improves camera-based applications especially when the conditions are not optimal for approaches using camera data alone.

References (20)

  1. J. Rekimoto, "Tilting operations for small screen interfaces," in 9th annual ACM symposium on User interface software and technology, pp. 167-168, ACM Press, 1996.
  2. K. Hinckley, J. S. Pierce, M. Sinclair, and E. Horvitz, "Sensing techniques for mobile interaction," in 13th annual ACM symposium on User Interface Software and Technology, pp. 91-100, 2000.
  3. T. Capin, K. Pulli, and T. Akenine-Möller, "The state of the art in mobile graphics research," IEEE Computer Graphics and Applications 1, pp. 74-84, 2008.
  4. M. Möhring, C. Lessig, and O. Bimber, "Optical tracking and video see-through ar on consumer cell phones," in Workshop on Virtual and Augmented Reality of the GI-Fachgruppe AR/VR, pp. 193-204, 2004.
  5. M. Rohs, "Real-world interaction with camera-phones," in 2nd International Symposium on Ubiquitous Computing Systems, pp. 39-48, 2004.
  6. J. Hwang, J. Jung, and G. J. Kim, "Hand-held virtual reality: A feasibility study," in ACM Virtual Reality Software and Technology, pp. 356-363, 2006.
  7. S. DiVerdi and T. Höllerer, "Groundcam: A tracking modality for mobile mixed reality," in IEEE Virtual Reality, pp. 75-82, 2007.
  8. X. Liu, D. Doermann, and H. Li, "Fast camera motion estimation for hand-held devices and applications," in 4th International Conference on Mobile and Ubiquitous Multimedia, pp. 103-108, 2005.
  9. S. Kratz and R. Ballagas, "Gesture recognition using motion estimation on mobile phones," in 3rd Intl. Workshop on Pervasive Mobile Interaction Devices at Pervasive 2007, 2007.
  10. A. Adams, N. Gelfand, and K. Pulli, "Viewfinder alignment," in Eurographics 2008, pp. 597-606, 2008.
  11. S. J. Ha, S. H. Lee, N. I. Cho, S. K. Kim, and B. Son, "Embedded panoramic mosaic system using auto-shot interface," IEEE Transactions on Consumer Electronics , pp. 16-24, 2008.
  12. R. E. Kalman, "A new approach to linear filtering and prediction problems," Transactions of the ASME- Journal of Basic Engineering (82), pp. 35-45, 1960.
  13. G. S. W. Klein and T. Drummond, "Tightly integrated sensor fusion for robust visual tracking," in Proc. British Machine Vision Conference, pp. 787-796, 2002.
  14. B. Jiang, U. Neumann, and S. You, "A robust hybrid tracking system for outdoor augmented reality," Virtual Reality Conference, IEEE , p. 3, 2004.
  15. J. Hannuksela, P. Sangi, and J. Heikkilä, "Vision-based motion estimation for interaction with mobile devices," Computer Vision and Image Understanding: Special Issue on Vision for Human-Computer Inter- action (1-2), pp. 188-195, 2007.
  16. P. M. Kuhn, Algorithms, Complexity Analysis and VLSI Architectures for MPEG-4 Motion Estimation, Kluwer Academic Publishers, Norwell, MA, USA, 1999.
  17. M. Bordallo, J. Hannuksela, O. Silven, and M. Vehviläinen, "Graphics hardware accelerated panorama builder for mobile phones.," Proc. SPIE Multimedia on Mobile Devices 2009 (72560D), 2009.
  18. G. W. Fitzmaurice, "Situated information spaces and spatially aware palmtop computers," Communications of the ACM (7), pp. 38-49, 1993.
  19. P. Eslambolchilar and R. Murray-Smith, "Tilt-based automatic zooming and scaling in mobile devices -a state-space implementation," in Mobile Human-Computer Interaction, pp. 120-131, 2004.
  20. M. Bordallo López, J. Boutellier, and O. Silvén, "Implementing mosaic stitching on mobile phones," in Finnish Signal Processing Symposium, 2007.