Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LCEval: Learned Composite Metric for Caption Evaluation (2012.13136v1)

Published 24 Dec 2020 in cs.AI

Abstract: Automatic evaluation metrics hold a fundamental importance in the development and fine-grained analysis of captioning systems. While current evaluation metrics tend to achieve an acceptable correlation with human judgements at the system level, they fail to do so at the caption level. In this work, we propose a neural network-based learned metric to improve the caption-level caption evaluation. To get a deeper insight into the parameters which impact a learned metrics performance, this paper investigates the relationship between different linguistic features and the caption-level correlation of the learned metrics. We also compare metrics trained with different training examples to measure the variations in their evaluation. Moreover, we perform a robustness analysis, which highlights the sensitivity of learned and handcrafted metrics to various sentence perturbations. Our empirical analysis shows that our proposed metric not only outperforms the existing metrics in terms of caption-level correlation but it also shows a strong system-level correlation against human assessments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Naeha Sharif (4 papers)
  2. Lyndon White (6 papers)
  3. Mohammed Bennamoun (124 papers)
  4. Wei Liu (1135 papers)
  5. Syed Afaq Ali Shah (16 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.