Abstract
While recent progress on abstractive summarization has led to remarkably fluent summaries, factual errors in generated summaries still severely limit their use in practice. In this paper, we evaluate summaries produced by state-of-the-art models via crowdsourcing and show that such errors occur frequently, in particular with more abstractive models. We study whether textual entailment predictions can be used to detect such errors and if they can be reduced by reranking alternative predicted summaries. That leads to an interesting downstream application for entailment models. In our experiments, we find that out-of-the-box entailment models trained on NLI datasets do not yet offer the desired performance for the downstream task and we therefore release our annotations as additional test data for future extrinsic evaluations of NLI.
Original language | English |
---|---|
Title of host publication | ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 2214-2220 |
Number of pages | 7 |
ISBN (Electronic) | 9781950737482 |
State | Published - 2020 |
Event | 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019 - Florence, Italy Duration: 28 Jul 2019 → 2 Aug 2019 |
Publication series
Name | ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference |
---|
Conference
Conference | 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019 |
---|---|
Country/Territory | Italy |
City | Florence |
Period | 28/07/19 → 2/08/19 |
Bibliographical note
Publisher Copyright:© 2019 Association for Computational Linguistics.
Funding
This work has been supported by the German Research Foundation through the research training group “Adaptive Preparation of Information from Heterogeneous Sources” (AIPHES, GRK 1994/1) and the German-Israeli Project Cooperation (DIP, grant DA 1600/1-1 and grant GU 798/17-1).
Funders | Funder number |
---|---|
DIP | GU 798/17-1, DA 1600/1-1 |
German-Israeli Project Cooperation | |
Deutsche Forschungsgemeinschaft | GRK 1994/1 |