Learning Saliency from Single Noisy Labelling: A Robust Model Fitting Perspective

Jing Zhang, Yuchao Dai*, Tong Zhang, Mehrtash Harandi, Nick Barnes, Richard Hartley

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

15 Citations (Scopus)

Abstract

The advances made in predicting visual saliency using deep neural networks come at the expense of collecting large-scale annotated data. However, pixel-wise annotation is labor-intensive and overwhelming. In this paper, we propose to learn saliency prediction from a single noisy labelling, which is easy to obtain (e.g., from imperfect human annotation or from unsupervised saliency prediction methods). With this goal, we address a natural question: Can we learn saliency prediction while identifying clean labels in a unified framework? To answer this question, we call on the theory of robust model fitting and formulate deep saliency prediction from a single noisy labelling as robust network learning and exploit model consistency across iterations to identify inliers and outliers (i.e., noisy labels). Extensive experiments on different benchmark datasets demonstrate the superiority of our proposed framework, which can learn comparable saliency prediction with state-of-the-art fully supervised saliency methods. Furthermore, we show that simply by treating ground truth annotations as noisy labelling, our framework achieves tangible improvements over state-of-the-art methods.

Original languageEnglish
Article number9303417
Pages (from-to)2866-2873
Number of pages8
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume43
Issue number8
DOIs
Publication statusPublished - 1 Aug 2021

Fingerprint

Dive into the research topics of 'Learning Saliency from Single Noisy Labelling: A Robust Model Fitting Perspective'. Together they form a unique fingerprint.

Cite this