skip to main content
10.1145/2671188.2749338acmconferencesArticle/Chapter ViewAbstractPublication PagesicmrConference Proceedingsconference-collections
short-paper

Exploiting Multiple Web Resources towards Collecting Positive Training Samples for Visual Concept Learning

Published:22 June 2015Publication History

ABSTRACT

The number of images uploaded to the web is enormous and is rapidly increasing. The purpose of our work is to use these for acquiring positive training data for visual concept learning. Manually creating training data for visual concept classifiers is an expensive and time consuming task. We propose an approach which automatically collects positive training samples from the Web by constructing a multitude of text queries and retaining for each query only very few top-ranked images returned by each one of the different web image search engines (Google, Flickr and Bing). In this way, we sift the burden of false positive rejection to the Web search engines and directly assemble a rich set of high-quality positive training samples. Experiments on forty concepts, evaluated on the ImageNet dataset, show the merit of the proposed approach.

References

  1. P. Over, G. Awad, et al., "Trecvid 2014 - an overview of the goals, tasks, data, evaluation mechanisms and metrics," in Proc. TRECVID 2014. NIST, USA, 2014.Google ScholarGoogle Scholar
  2. X. Li, C. G. Snoek, et al., "Harvesting social images for bi-concept search," IEEE Trans. on Multimedia, vol. 14, no. 4, pp. 1091--1104, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. X. Li, C. G. Snoek, and M. Worring, "Unsupervised multi-feature tag relevance learning for social image retrieval," in Proc. Int. Conf. on Image and Video Retrieval. ACM, 2010, pp. 10--17. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. J. Deng, W. Dong, et al., "Imagenet: A large-scale hierarchical image database," in Proc. Int. Conf. on Computer Vision and Pattern Recognition. IEEE, 2009, pp. 248--255.Google ScholarGoogle Scholar
  5. S. Zhu, C.-W. Ngo, and Y.-G. Jiang, "Sampling and ontologically pooling web images for visual concept learning," IEEE Trans. on Multimedia, vol. 14, no. 4, pp. 1068--1078, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. F. Schroff, A. Criminisi, and A. Zisserman, "Harvesting image databases from the web," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 33, no. 4, pp. 754--766, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. L.-J. Li and L. Fei-Fei, "Optimol: automatic online picture collection via incremental model learning," Int. Journal of Computer Vision, vol. 88, no. 2, pp. 147--168, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. X. Li and C. G. Snoek, "Visual categorization with negative examples for free," in Proc. 17th ACM Int. Conf. on Multimedia, 2009, pp. 661--664. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. C. Fellbaum, WordNet: An Electronic Lexical Database, Bradford Books, 1998.Google ScholarGoogle ScholarCross RefCross Ref
  10. L. Han, A. Kashyap, et al., "Umbc ebiquity-core: Semantic textual similarity systems," in Proc. 2nd Joint Conf. on Lexical and Computational Semantics, 2013, vol. 1, pp. 44--52.Google ScholarGoogle Scholar
  11. D. Carvalho, C. Callı, et al., "Easyesa: A low-effort infrastructure for explicit semantic analysis," in Proc. 13th Int. Semantic Web Conference (ISWC), 2014.Google ScholarGoogle Scholar
  12. J. Ramos, "Using tf-idf to determine word relevance in document queries," in Proc. 1st Instructional Conf. on Machine Learning, 2003.Google ScholarGoogle Scholar
  13. F. Markatopoulou, N. Pittaras, et al., "A study on the use of a binary local descriptor and color extensions of local descriptors for video concept detection," in Proc. MultiMedia Modeling, Springer, 2015, pp. 282--293.Google ScholarGoogle Scholar
  14. A. F. Smeaton, P. Over, and W. Kraaij, "Evaluation campaigns and trecvid," in Proc. 8th ACM Int. Workshop on Multimedia Information Retrieval, 2006, pp. 321--330. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Exploiting Multiple Web Resources towards Collecting Positive Training Samples for Visual Concept Learning

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ICMR '15: Proceedings of the 5th ACM on International Conference on Multimedia Retrieval
      June 2015
      700 pages
      ISBN:9781450332743
      DOI:10.1145/2671188

      Copyright © 2015 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 22 June 2015

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      ICMR '15 Paper Acceptance Rate48of127submissions,38%Overall Acceptance Rate254of830submissions,31%

      Upcoming Conference

      ICMR '24
      International Conference on Multimedia Retrieval
      June 10 - 14, 2024
      Phuket , Thailand

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader