Abstract
Recent advancements in self-attention neural network architectures have raised the bar for open-ended text generation. Yet, while current methods are capable of producing a coherent text which is several hundred words long, attaining control over the content that is being generated—as well as evaluating it—are still open questions. We propose a controlled generation task which is based on expanding a sequence of facts, expressed in natural language, into a longer narrative. We introduce human-based evaluation metrics for this task, as well as a method for deriving a large training dataset. We evaluate three methods on this task, based on fine-tuning pre-trained models. We show that while auto-regressive, unidirectional Language Models such as GPT2 produce better fluency, they struggle to adhere to the requested facts. We propose a plan-and-cloze model (using fine-tuned XLNet) which produces competitive fluency while adhering to the requested content.
Original language | English |
---|---|
Title of host publication | COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference |
Editors | Donia Scott, Nuria Bel, Chengqing Zong |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 2329-2345 |
Number of pages | 17 |
ISBN (Electronic) | 9781952148279 |
DOIs | |
State | Published - 2020 |
Event | 28th International Conference on Computational Linguistics, COLING 2020 - Virtual, Online, Spain Duration: 8 Dec 2020 → 13 Dec 2020 |
Publication series
Name | COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference |
---|
Conference
Conference | 28th International Conference on Computational Linguistics, COLING 2020 |
---|---|
Country/Territory | Spain |
City | Virtual, Online |
Period | 8/12/20 → 13/12/20 |
Bibliographical note
Publisher Copyright:© 2020 COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference. All rights reserved.