Deep GoogLeNet Features for Visual Object Tracking
Citations Over TimeTop 15% of 2018 papers
Abstract
Convolutional Neural Network (CNN) has recently become very popular in visual object tracking due to their strong feature representation capabilities. Almost all of the CNN based trackers currently use the features extracted from shallow convolutional layers of VGGNet architecture. This paper presents an investigation of the impact of deep convolutional layer features in an object tracking framework. In this study, we demonstrate for the first time, the viability of features extracted from deep layers of GoogLeNet CNN architecture for the purpose of object tracking. We integrated GoogLeNet features in a discriminative correlation filter based tracking framework. Our experimental results show that the GoogLeNet features provides significant computational advantages over the conventionally used VGGNet features, without much compromise on the tracking performance. It was observed that features obtained from inception modules of GoogLeNet have high depths. Further, Principal Component Analysis (PCA) was employed to reduce the dimensionality of the extracted features. This greatly reduces the computational cost and thus improve the speed of the tracking process. Extensive evaluation have been performed on three benchmark datasets: OTB, ALOV300++ and VOT2016 datasets and its performances are measured in terms of metrics like F-score, One Pass Evaluation, robustness and accuracy.
Related Papers
- → Adaptive Feature Selection Siamese Networks for Visual Tracking(2020)9 cited
- → Densely Connected Discriminative Correlation Filters for Visual Tracking(2018)5 cited
- → Attention fusion and target-uncertain detection for discriminative tracking(2023)3 cited
- → Cascaded Generative and Discriminative Learning for Visual Tracking(2013)2 cited
- → Current progress in discriminative object tracking(2017)