Abstract
In this work, we analyze the effectiveness of perceptual features to automatically annotate video clips in domain-specific video digital libraries. Typically, automatic annotation is provided by computing clip similarity with respect to given examples, which constitute the knowledgebase, in accordance with a given ontology or a classification scheme. Since the amount of training clips is normally very large, we propose to automatically extract some prototypes, or visual concepts, for each class instead of using the whole knowledge base. The prototypes are generated after a Complete Link clustering based on perceptual features with an automatic selection of the number of clusters. Context based information are used in an intra-class clustering framework to provide selection of more discriminative clips. Reducing the number of samples makes the matching process faster and lessens the storage requirements. Clips are annotated following the MPEG-7 directives to provide easier portability. Results are provided on videos taken from sports and news digital libraries.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Jaimes, A., Christel, M., Gilles, S., Sarukkai, R., Ma, W.: Multimedia information retrieval: what is it, and why isn’t anyone using it? In: MIR 2005. Proc. of the 7th ACM SIGMM Int’l Workshop on Multimedia Information Retrieval, pp. 3–8. ACM Press, New York (2005)
Bertini, M., Cucchiara, R., Del Bimbo, A., Torniai, C.: Video Annotation with Pictorially Enriched Ontologies. In: Proceedings of IEEE International Conference on Multimedia and Expo, Amsterdam, The Netherlands, pp. 1428–1431. IEEE Computer Society Press, Los Alamitos (2005)
Chang, S., Chen, W., Meng, H.J., Sundaram, H., Zhong, D.: A Fully Automated Content-Based Video Search Engine Supporting Spatiotemporal Queries. IEEE Transactions on Circuits and System for Video Technology 8(5), 602–615 (1998)
Grana, C., Tardini, G., Cucchiara, R.: MPEG-7 Compliant Shot Detection in Sport Videos. In: ISM 2005. Proc. of the IEEE Int’l Symposium on Multimedia, pp. 395–402 (2005)
Grana, C., Vezzani, R., Bulgarelli, D., Barbieri, F., Cucchiara, R., Bertini, M., Torniai, C., Del Bimbo, A.: PEANO: Pictorial Enriched ANnotation of VideO. In: Proceedings of the 14th ACM international Conference on Multimedia, pp. 793–794 (2006)
ISO/IEC Std. 15 938-3: Information technology - Multimedia content description interface - Part 3: Visual (2003)
Jain, A.K., Dubes, R.C.: Algorithms for clustering data. Prentice-Hall, Englewood Cliffs, NJ (1988)
Dunn, J.C.: A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. Journal of Cybernetics 3(3), 32–57 (1973)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Grana, C., Vezzani, R., Cucchiara, R. (2007). Prototypes Selection with Context Based Intra-class Clustering for Video Annotation with Mpeg7 Features. In: Thanos, C., Borri, F., Candela, L. (eds) Digital Libraries: Research and Development. DELOS 2007. Lecture Notes in Computer Science, vol 4877. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-77088-6_26
Download citation
DOI: https://doi.org/10.1007/978-3-540-77088-6_26
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-77087-9
Online ISBN: 978-3-540-77088-6
eBook Packages: Computer ScienceComputer Science (R0)