Cross-Modal Retrieval: A Pairwise Classification Approach

Aditya Krishna Menon, Didi Sudan, Sanjay Chawla

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    9 Citations (Scopus)

    Abstract

    Content is increasingly available in multiple modalities (such as images, text, and video), each of which provides a different representation of some entity. The cross-modal retrieval problem is: given the representation of an entity in one modality, find its best representation in all other modalities. We propose a novel approach to this problem based on pairwise classification. The approach seamlessly applies to both the settings where ground-truth annotations for the entities are absent and present. In the former case, the approach considers both positive and unlabelled links that arise in standard cross-modal retrieval datasets. Empirical comparisons show improvements over state-of-the-art methods for cross-modal retrieval.

    Original languageEnglish
    Title of host publicationProceedings of the 2015 SIAM International Conference on Data Mining
    EditorsJieping Ye, Suresh Venkatasubramanian
    PublisherSociety for Industrial and Applied Mathematics Publications
    Pages199-207
    Number of pages9
    ISBN (Electronic)9781510811522
    DOIs
    Publication statusPublished - 2015
    EventSIAM International Conference on Data Mining 2015 - Vancouver, Canada
    Duration: 30 Apr 20152 May 2015
    https://epubs.siam.org/doi/book/10.1137/1.9781611974010

    Conference

    ConferenceSIAM International Conference on Data Mining 2015
    Abbreviated titleSDM 2015
    Country/TerritoryCanada
    CityVancouver
    Period30/04/152/05/15
    Internet address

    Fingerprint

    Dive into the research topics of 'Cross-Modal Retrieval: A Pairwise Classification Approach'. Together they form a unique fingerprint.

    Cite this