CLIPScore: A Reference-free Evaluation Metric for Image Captioning
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing2021pp. 7514–7528
Citations Over TimeTop 1% of 2021 papers
Abstract
Image captioning has conventionally relied on reference-based automatic evaluations, where machine captions are compared against captions written by humans. This is in contrast to the reference-free manner in which humans assess caption quality.
Related Papers
- → Video-Captioning Evaluation Metric for Segments (VEMS): A Metric for Segment-level Evaluation of Video Captions with Weighted Frames(2023)2 cited
- → OSCAR and ActivityNet: an Image Captioning model can effectively learn a Video Captioning dataset(2021)1 cited
- → Comprehensive Comparative Study on Several Image Captioning Techniques Based on Deep Learning Algorithm(2021)1 cited
- → CIC Chinese Image Captioning Based on Image Label Information(2021)
- → Image Captioning using Neural Networks(2022)