Abstract
Content is increasingly available in multiple modalities (such as images, text, and video), each of which provides a different representation of some entity. The cross-modal retrieval problem is: given the representation of an entity in one modality, find its best representation in all other modalities. We propose a novel approach to this problem based on pairwise classification. The approach seamlessly applies to both the settings where ground-truth annotations for the entities are absent and present. In the former case, the approach considers both positive and unlabelled links that arise in standard cross-modal retrieval datasets. Empirical comparisons show improvements over state-of-the-art methods for cross-modal retrieval.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2015 SIAM International Conference on Data Mining |
Editors | Jieping Ye, Suresh Venkatasubramanian |
Publisher | Society for Industrial and Applied Mathematics Publications |
Pages | 199-207 |
Number of pages | 9 |
ISBN (Electronic) | 9781510811522 |
DOIs | |
Publication status | Published - 2015 |
Event | SIAM International Conference on Data Mining 2015 - Vancouver, Canada Duration: 30 Apr 2015 → 2 May 2015 https://epubs.siam.org/doi/book/10.1137/1.9781611974010 |
Conference
Conference | SIAM International Conference on Data Mining 2015 |
---|---|
Abbreviated title | SDM 2015 |
Country/Territory | Canada |
City | Vancouver |
Period | 30/04/15 → 2/05/15 |
Internet address |