Abstract
Long-context question answering (QA) tasks require reasoning over a long document or multiple documents. Addressing these tasks often benefits from identifying a set of evidence spans (e.g., sentences), which provide supporting evidence for answering the question. In this work, we propose a novel method for equipping long-context QA models with an additional sequence-level objective for better identification of the supporting evidence. We achieve this via an additional contrastive supervision signal in finetuning, where the model is encouraged to explicitly discriminate supporting evidence sentences from negative ones by maximizing question-evidence similarity. The proposed additional loss exhibits consistent improvements on three different strong long-context transformer models, across two challenging question answering benchmarks - HotpotQA and QAsper.
| Original language | English |
|---|---|
| Title of host publication | NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics |
| Subtitle of host publication | Human Language Technologies, Proceedings of the Conference |
| Publisher | Association for Computational Linguistics (ACL) |
| Pages | 2872-2879 |
| Number of pages | 8 |
| ISBN (Electronic) | 9781955917711 |
| DOIs | |
| State | Published - 2022 |
| Event | 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022 - Hybrid, Seattle, United States Duration: 10 Jul 2022 → 15 Jul 2022 |
Publication series
| Name | NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference |
|---|
Conference
| Conference | 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022 |
|---|---|
| Country/Territory | United States |
| City | Hybrid, Seattle |
| Period | 10/07/22 → 15/07/22 |
Bibliographical note
Publisher Copyright:© 2022 Association for Computational Linguistics.
Funding
We thank the BIU-NLP lab and the Semantic Scholar research team at AI2 for fruitful discussions and helpful feedback. The work described herein was supported by the PBC fellowship for outstanding PhD candidates in data science, in part by by the Israel Science Foundation grant 2827/21, and by a grant from the Israel Ministry of Science and Technology.
| Funders | Funder number |
|---|---|
| Israel Science Foundation | 2827/21 |
| Ministry of science and technology, Israel | |
| Planning and Budgeting Committee of the Council for Higher Education of Israel |
Fingerprint
Dive into the research topics of 'Long Context Question Answering via Supervised Contrastive Learning'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver