Loading web-font TeX/Main/Regular
Cross-Modal Learning to Rank via Latent Joint Representation | IEEE Journals & Magazine | IEEE Xplore

Cross-Modal Learning to Rank via Latent Joint Representation

Publisher: IEEE

Abstract:

Cross-modal ranking is a research topic that is imperative to many applications involving multimodal data. Discovering a joint representation for multimodal data and lear...View more

Abstract:

Cross-modal ranking is a research topic that is imperative to many applications involving multimodal data. Discovering a joint representation for multimodal data and learning a ranking function are essential in order to boost the cross-media retrieval (i.e., image-query-text or text-query-image). In this paper, we propose an approach to discover the latent joint representation of pairs of multimodal data (e.g., pairs of an image query and a text document) via a conditional random field and structural learning in a listwise ranking manner. We call this approach cross-modal learning to rank via latent joint representation (CML ^{2}\text{R} ). In CML ^{2}\text{R} , the correlations between multimodal data are captured in terms of their sharing hidden variables (e.g., topics), and a hidden-topic-driven discriminative ranking function is learned in a listwise ranking manner. The experiments show that the proposed approach achieves a good performance in cross-media retrieval and meanwhile has the capability to learn the discriminative representation of multimodal data.
Published in: IEEE Transactions on Image Processing ( Volume: 24, Issue: 5, May 2015)
Page(s): 1497 - 1509
Date of Publication: 12 February 2015

ISSN Information:

PubMed ID: 25700450
Publisher: IEEE

Funding Agency:


References

References is not available for this document.