Skip to main content

Video as Input: Spiral Search with the Sparse Angular Sampling

  • Conference paper
Computer and Information Sciences – ISCIS 2006 (ISCIS 2006)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 4263))

Included in the following conference series:

Abstract

This paper presents an improved cross-correlation algorithm for template-based object tracking: the reduced spiral search with a sparse angular sampling. The basic parameters of the algorithm for the real-time face tracking were evaluated regarding their impact on the algorithm performance. They are the minimum number of pixels and the size of the template, the correlation threshold and drifting, and the parameters of the search – radius, shift, direction, and rotation of the template. We demonstrated that the information provided by the grid-like template might be reduced to 16 pixels with a grid step of 15 pixels. A spiral search in 8 directions with a minimum shift of 1 pixel decreases the number of computations by 20 times. Being activated sequentially the template rotation does not increase the performance, but doing the tracking adaptive and robust.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Bérard, F.: The perceptual window: Head motion as a new input stream. In: IFIP Conference on Human-Computer Interaction, pp. 238–244 (1999)

    Google Scholar 

  2. Betke, M., Gips, J., Fleming, P.: The Camera Mouse: Visual Tracking of Body Features to Provide Computer Access For People with Severe Disabilities. IEEE Transactions on Neural Systems and Rehabilitation Engineering 10(1), 1–10 (2002)

    Article  Google Scholar 

  3. Blalock, T.N., et al.: Method and Device for Tracking Relative Movement by Correlating Signals From an Array of Photoelements, US Patent 5729008 (1998)

    Google Scholar 

  4. Brunelli, R., Poggio, T.: Face recognition: features versus templates. IEEE Trans. Pattern Analysis and Machine Intelligence 15(10), 1042–1052 (1993)

    Article  Google Scholar 

  5. Comaniciu, D., Ramesh, V., Meer, P.: Real-Time Tracking of Non-Rigid Objects using Mean Shift. In: IEEE Conf. Computer Vision and Pattern Recognition (CVPR 2000), vol. 2, pp. 142–149 (2002)

    Google Scholar 

  6. Crowley, J.L., Berard, F., Coutaz, J.: Finger Tracking as an Input Device for Augmented Reality. In: Int. Workshop on Face and Gesture Recognition, pp. 195–200 (1995)

    Google Scholar 

  7. FaceMOUSE. Product information. Web site (2005), http://www.aidalabs.com/

  8. Intel Image Processing library Open CV, http://www.intel.com/technology/~/opencv/

  9. Jilin Tu, T., Huang, T., Tao, H.: Face As Mouse Through Visual Face Tracking. In: Proc. of the 2nd Canadian Conf Computer and Robot Vision, pp. 339–346 (2005)

    Google Scholar 

  10. Lewis, J.P.: Fast Template Matching. Vision Interface, 120–123 (1995)

    Google Scholar 

  11. Product information on EyeTwig.com. Website (2005), http://www.eyetwig.com

  12. Zhang, S.-C., Liu, Z.-Q.: A robust, real-time ellipse detector. J. Pattern Recognition 38, 273–287 (2005)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Evreinova, T.V., Evreinov, G., Raisamo, R. (2006). Video as Input: Spiral Search with the Sparse Angular Sampling. In: Levi, A., Savaş, E., Yenigün, H., Balcısoy, S., Saygın, Y. (eds) Computer and Information Sciences – ISCIS 2006. ISCIS 2006. Lecture Notes in Computer Science, vol 4263. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11902140_58

Download citation

Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Publish with us

Policies and ethics