Abstract

We propose BERTScore, an automatic evaluation metric for text generation. Analogously to common metrics, BERTScore computes a similarity score for each token in the candidate sentence with each token in the reference sentence. However, instead of exact matches, we compute token similarity using contextual embeddings. We evaluate using the outputs of 363 machine translation and image captioning systems. BERTScore correlates better with human judgments and provides stronger model selection performance than existing metrics. Finally, we use an adversarial paraphrase detection task to show that BERTScore is more robust to challenging examples when compared to existing metrics.

Keywords

Security tokenParaphraseComputer scienceMachine translationSimilarity (geometry)Artificial intelligenceSentenceMetric (unit)Selection (genetic algorithm)Natural language processingTask (project management)Closed captioningMachine learningSpeech recognitionImage (mathematics)

Related Publications

Publication Info

Year
2019
Type
preprint
Citations
2001
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2001
OpenAlex

Cite This

Tianyi Zhang, Varsha Kishore, Felix Wu et al. (2019). BERTScore: Evaluating Text Generation with BERT. arXiv (Cornell University) . https://doi.org/10.48550/arxiv.1904.09675

Identifiers

DOI
10.48550/arxiv.1904.09675