Abstract
In this talk I will present ongoing work at Leeds on building models of video activity. I will present techniques, both supervised and unsupervised, for learning the spatio-temporal structure of tasks and events from video or other sensor data. In both cases, the representation will exploit qualititive spatio-temporal relations. A novel method for robustly transforming video data to qualitative relations will be presented. For supervised learning I will show how the supervisory burden can be reduced using what we term “deictic supervision”, whilst in the unsupervised case I will present a method for learning the most likely interpretation of the training data. I will also show how objects can be “functionally categorised” according to their spatio-temporal behaviour and how the use of type information can help in the learning process, especially in the presence of noise. I will present results from several domains including a kitchen scenario and an aircraft apron.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Cohn, A.G. (2010). Mining Video Data: Learning about Activities. In: Bi, Y., Williams, MA. (eds) Knowledge Science, Engineering and Management. KSEM 2010. Lecture Notes in Computer Science(), vol 6291. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-15280-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-642-15280-1_1
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-15279-5
Online ISBN: 978-3-642-15280-1
eBook Packages: Computer ScienceComputer Science (R0)