Skip to main content

Multi-camera Egocentric Activity Detection for Personal Assistant

  • Conference paper
Book cover Advances in Multimedia Modeling

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 7733))

Abstract

We demonstrate an egocentric human activity assistant system that has been developed to aid people in doing explicitly encoded motion behavior, such as operating a home infusion pump in sequence. This system is based on a robust multi-camera egocentric human behavior detection approach. This approach detects individual actions in interesting hot regions by spatio-temporal mid-level features, which are built by spatial bag-of-words method in time sliding window. Using a specific infusion pump as a test case, our goal is to detect individual human actions in the operations of a home medical device to see whether the patient is correctly performing the required actions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Fathi, A., Ren, X., Rehg, J.: Learning to Recognize Objects in Egocentric Activities. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR (2011)

    Google Scholar 

  2. Gao, Z., Detyniecki, M., Chen, M.-Y., Hauptmann, A.G., Wactlar, H.D., Cai, A.: The Application of Spatio-temporal Feature and Multi-Sensor in Home Medical Devices. International Journal of Digital Content Technology and its Applications(IJDCTA) 4(6), 69–78 (2010)

    Google Scholar 

  3. Laptev, I.: On space-time interest points. International Journal of Computer Vision 64(2-3), 107–123 (2005)

    Article  Google Scholar 

  4. Fan, J., Wu, Y., Dai, S.: Discriminative Spatial Attention for Robust Tracking. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part I. LNCS, vol. 6311, pp. 480–493. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  5. Chen, M.Y., Hauptmann, A.: MoSIFT: Recognizing human actions in surveillance videos. CMU-CS-09-161, Carnegie Mellon University (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zhang, L., Gao, Y., Tong, W., Ding, G., Hauptmann, A. (2013). Multi-camera Egocentric Activity Detection for Personal Assistant. In: Li, S., et al. Advances in Multimedia Modeling. Lecture Notes in Computer Science, vol 7733. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35728-2_50

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-35728-2_50

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-35727-5

  • Online ISBN: 978-3-642-35728-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics