Abstract
Neural abstractive summarization models have drastically improved in the recent years. However, the summaries generated by these models generally suffer from issues such as: not capturing the critical facts in source documents, and containing facts that are inconsistent with the source documents. In this work, we present a general framework to train abstractive summarization models to alleviate such issues. We first train a sequence-to-sequence model to summarize documents, and then further train this model in a Reinforcement Learning setting with question-answering based rewards. We evaluate the summaries generated by the this framework using multiple automatic measures and human judgements. The experimental results show that the question-answering rewards can be used as a general framework to improve neural abstractive summarization. Particularly, the results from human evaluations show that the summaries generated by our approach are preferred over 30% of the time over the summaries generated by general abstractive summarization models.
| Original language | English |
|---|---|
| Title of host publication | Findings of the Association for Computational Linguistics, Findings of ACL |
| Subtitle of host publication | EMNLP 2021 |
| Editors | Marie-Francine Moens, Xuanjing Huang, Lucia Specia, Scott Wen-Tau Yih |
| Publisher | Association for Computational Linguistics (ACL) |
| Pages | 518-526 |
| Number of pages | 9 |
| ISBN (Electronic) | 9781955917100 |
| DOIs | |
| State | Published - 2021 |
| Externally published | Yes |
| Event | 2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 - Punta Cana, Dominican Republic Duration: 7 Nov 2021 → 11 Nov 2021 |
Publication series
| Name | Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 |
|---|
Conference
| Conference | 2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 |
|---|---|
| Country/Territory | Dominican Republic |
| City | Punta Cana |
| Period | 7/11/21 → 11/11/21 |
Bibliographical note
Publisher Copyright:© 2021 Association for Computational Linguistics.