Encoding based saliency detection for videos and images
Citations Over TimeTop 10% of 2015 papers
Abstract
We present a novel video saliency detection method to support human activity recognition and weakly supervised training of activity detection algorithms. Recent research has emphasized the need for analyzing salient information in videos to minimize dataset bias or to supervise weakly labeled training of activity detectors. In contrast to previous methods we do not rely on training information given by either eye-gaze or annotation data, but propose a fully unsupervised algorithm to find salient regions within videos. In general, we enforce the Gestalt principle of figure-ground segregation for both appearance and motion cues. We introduce an encoding approach that allows for efficient computation of saliency by approximating joint feature distributions. We evaluate our approach on several datasets, including challenging scenarios with cluttered background and camera motion, as well as salient object detection in images. Overall, we demonstrate favorable performance compared to state-of-the-art methods in estimating both ground-truth eye-gaze and activity annotations.
Related Papers
- → Intelligent Detection of Missing and Unattended Objects in Complex Scene of Surveillance Videos(2012)10 cited
- → Surface-based general 3D object detection and pose estimation(2014)5 cited
- → Towards Object Detection from Motion(2019)1 cited
- → Object extraction from stereo vision using continuity of disparity map(2012)
- → 2A1-S02 Unknown Object Detection and Picking Based on Impacting and Keypoint Tracking(2015)