What Value Do Explicit High Level Concepts Have in Vision to Language Problems?
Citations Over TimeTop 1% of 2016 papers
Abstract
Much recent progress in Vision-to-Language (V2L) problems has been achieved through a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). This approach does not explicitly represent high-level semantic concepts, but rather seeks to progress directly from image features to text. In this paper we investigate whether this direct approach succeeds due to, or despite, the fact that it avoids the explicit representation of high-level information. We propose a method of incorporating high-level concepts into the successful CNN-RNN approach, and show that it achieves a significant improvement on the state-of-the-art in both image captioning and visual question answering. We also show that the same mechanism can be used to introduce external semantic information and that doing so further improves performance. We achieve the best reported results on both image captioning and VQA on several benchmark datasets, and provide an analysis of the value of explicit high-level concepts in V2L problems.
Related Papers
- → OSCAR and ActivityNet: an Image Captioning model can effectively learn a Video Captioning dataset(2021)1 cited
- → Video Captioning via Hierarchical Reinforcement Learning(2017)22 cited
- → Boosted Attention: Leveraging Human Attention for Image Captioning(2019)1 cited
- → Image Captioning Methodologies Using Deep Learning: A Review(2020)
- → Image Captioning using Neural Networks(2022)