SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks
Citations Over TimeTop 1% of 2015 papers
Abstract
Saliency in Context (SALICON) is an ongoing effort that aims at understanding and predicting visual attention. Conventional saliency models typically rely on low-level image statistics to predict human fixations. While these models perform significantly better than chance, there is still a large gap between model prediction and human behavior. This gap is largely due to the limited capability of models in predicting eye fixations with strong semantic content, the so-called semantic gap. This paper presents a focused study to narrow the semantic gap with an architecture based on Deep Neural Network (DNN). It leverages the representational power of high-level semantics encoded in DNNs pretrained for object recognition. Two key components are fine-tuning the DNNs fully convolutionally with an objective function based on the saliency evaluation metrics, and integrating information at different image scales. We compare our method with 14 saliency models on 6 public eye tracking benchmark datasets. Results demonstrate that our DNNs can automatically learn features particularly for saliency prediction that surpass by a big margin the state-of-the-art. In addition, our model ranks top to date under all seven metrics on the MIT300 challenge set.
Related Papers
- → Exploring the Semantic Gap for Movie Recommendations(2017)43 cited
- → Mining user hidden semantics from image content for image retrieval(2007)6 cited
- → Relevance feedback for semantics based image retrieval(2002)22 cited
- Method mapping image low-level features to high-level semantics(2004)
- → Image retrieval based on semantics of intra-region color properties(2008)2 cited