Abstract
Content is increasingly available in multiple modalities (such as images, text, and video), each of which provides a different representation of some entity. The cross-modal retrieval problem is: given the representation of an entity in one modality, find its best representation in all other modalities. We propose a novel approach to this problem based on pairwise classification. The approach seamlessly applies to both the settings where ground-truth annotations for the entities are absent and present. In the former case, the approach considers both positive and unlabelled links that arise in standard cross-modal retrieval datasets. Empirical comparisons show improvements over state-of-the-art methods for cross-modal retrieval.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 2015 SIAM International Conference on Data Mining |
| Editors | Jieping Ye, Suresh Venkatasubramanian |
| Publisher | Society for Industrial and Applied Mathematics Publications |
| Pages | 199-207 |
| Number of pages | 9 |
| ISBN (Electronic) | 9781510811522 |
| DOIs | |
| Publication status | Published - 2015 |
| Event | SIAM International Conference on Data Mining 2015 - Vancouver, Canada Duration: 30 Apr 2015 → 2 May 2015 https://epubs.siam.org/doi/book/10.1137/1.9781611974010 |
Conference
| Conference | SIAM International Conference on Data Mining 2015 |
|---|---|
| Abbreviated title | SDM 2015 |
| Country/Territory | Canada |
| City | Vancouver |
| Period | 30/04/15 → 2/05/15 |
| Internet address |
Fingerprint
Dive into the research topics of 'Cross-Modal Retrieval: A Pairwise Classification Approach'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver