Abstract
This chapter proposes an object movement detection method in household environments. The proposed method detects “object placement” and “object removal” via images captured by environment-embedded cameras. When object movement detection is performed in household environments, there are several difficulties: the method needs to detect object movements robustly even if sizes of objects are small, the method must discriminate objects and non-objects such as humans. In this work, we propose an object movement detection method by detecting “stable changes”, which are changing from the recorded state but which change are settled. To categorize objects and non-objects via the stable changes even though non-objects make long-term changes (e.g. a person is sitting down), we employ motion history of changed regions. In addition, to classify object placement and object removal, we use multiple-layered background model, called the layered background model and edge subtraction technique. The experiment shows the system can detect objects robustly and in sufficient frame-rates.
An Erratum for this chapter can be found at http://dx.doi.org/10.1007/978-3-642-17554-1_11
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Gupta, A., Kembhavi, A., Davis, L.: Observing human-object interactions: using spatial and functional compatibility for recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 31(10), 1775–1789 (2009)
Navneet, D., Triggs, B.: Histograms of oriented gradients for human detection. In: CVPR (2005)
Sabzmeydani, P., Mori, G.: Detecting pedestrians by learning shapelet features. In: CVPR (2007)
Itti, L., Baldi, P.: A principled approach to detecting surprising events in video. In: CVPR (2005)
Gould, S., Arfvidsson, J., Kaehler, A., Sapp, B., Messner, M., Bradski, G., Baumstarck, P., Chung, S., Ng, A.: Peripheral-foveal vision for real-time object recognition and tracking in video. In: IJCAI (2007)
Maki, K., Katayama, N., Shimada, N., Shirai, Y.: Image-based automatic detection of indoor scene events and interactive inquiry. In: ICPR (2008)
Kim, K., Chalidabhongse, T., Harwood, D., Davis, L.: Real-time foreground–background segmentation using codebook model. Real-Time Imaging 11(3), 172–185 (2005)
Tian, Y., Lu, M., Hampapur, A.: Robust and efficient foreground analysis for real-time video surveillance. In: CVPR (2005)
Harville, M.: A framework for high-level feedback to adaptive, per-pixel, mixture-of-gaussian background models. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002. LNCS, vol. 2352, pp. 543–560. Springer, Heidelberg (2002)
Grabner, H., Roth, P., Grabner, M., Bischof, H.: Autonomous learning of a robust background model for change detection. In: Performance Evaluation of Tracking and Surveillance (PETS) Workshop at CVPR (2006)
Shimosaka, M., Murasaki, K., Mori, T., Sato, T.: Human shape reconstruction via graph cuts for voxel-based markerless motion capture in intelligent environment. In: IUCS (2009)
Sun, J., Zhang, W., Tang, X., Shum, H.-Y.: Background cut. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3952, pp. 628–641. Springer, Heidelberg (2006)
Boykov, Y., Kolmogorov, V.: An experimental comparison of min-cut/max-flow algorithms for energy minimization in vision. IEEE Transactions on Pattern Analysis and Machine Intelligence 26(9), 1124–1137 (2004)
Horprasert, T., Harwood, D., Davis, L.: A robust background subtraction and shadow detection. In: ACCV (2000)
Rosten, E., Drummond, T.: Machine learning for high-speed corner detection. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3951, pp. 430–443. Springer, Heidelberg (2006)
Gevers, T., Smeulders, A.: PicToSeek: combining color and shape invariant features for image retrieval. IEEE Transactions on Image Processing 9(1), 102–119 (2000)
Nguyen, H., Smeulders, A.: Fast occluded object tracking by a robust appearance filter. IEEE Transactions on Pattern Analysis and Machine Intelligence 26(8), 1099–1104 (2004)
Pérez, P., Hue, C., Vermaak, J., Gangnet, M.: Color-based probabilistic tracking. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002. LNCS, vol. 2350, pp. 661–675. Springer, Heidelberg (2002)
Connell, J., Senior, A., Hampapur, A., Tian, Y.L., Brown, L., Pankanti, S.: Detection and tracking in the IBM peoplevision system. In: ICME (2004)
Odashima, S., Mori, T., Shimosaka, M., Noguchi, H., Sato, T.: Object movement event detection for household environments via layered-background model and keypoint-based tracking. In: International Workshop on Video Event Categorization, Tagging and Retrieval (2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Odashima, S., Mori, T., Simosaka, M., Noguchi, H., Sato, T. (2011). Event Understanding of Human-Object Interaction: Object Movement Detection via Stable Changes. In: Zhang, J., Shao, L., Zhang, L., Jones, G.A. (eds) Intelligent Video Event Analysis and Understanding. Studies in Computational Intelligence, vol 332. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-17554-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-642-17554-1_9
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-17553-4
Online ISBN: 978-3-642-17554-1
eBook Packages: EngineeringEngineering (R0)