Abstract
One of the goals of natural language understanding is to develop models that map sentences into meaning representations. However, training such models requires expensive annotation of complex structures, which hinders their adoption. Learning to actively-learn (LTAL) is a recent paradigm for reducing the amount of labeled data by learning a policy that selects which samples should be labeled. In this work, we examine LTAL for learning semantic representations, such as QA-SRL. We show that even an oracle policy that is allowed to pick examples that maximize performance on the test set (and constitutes an upper bound on the potential of LTAL), does not substantially improve performance compared to a random policy. We investigate factors that could explain this finding and show that a distinguishing characteristic of successful applications of LTAL is the interaction between optimization and the oracle policy selection process. In successful applications of LTAL, the examples selected by the oracle policy do not substantially depend on the optimization procedure, while in our setup the stochastic nature of optimization strongly affects the examples selected by the oracle. We conclude that the current applicability of LTAL for improving data efficiency in learning semantic meaning representations is limited.
Original language | English |
---|---|
Title of host publication | CoNLL 2019 - 23rd Conference on Computational Natural Language Learning, Proceedings of the Conference |
Publisher | Association for Computational Linguistics |
Pages | 452-462 |
Number of pages | 11 |
ISBN (Electronic) | 9781950737727 |
State | Published - 2019 |
Externally published | Yes |
Event | 23rd Conference on Computational Natural Language Learning, CoNLL 2019 - Hong Kong, China Duration: 3 Nov 2019 → 4 Nov 2019 |
Publication series
Name | CoNLL 2019 - 23rd Conference on Computational Natural Language Learning, Proceedings of the Conference |
---|
Conference
Conference | 23rd Conference on Computational Natural Language Learning, CoNLL 2019 |
---|---|
Country/Territory | China |
City | Hong Kong |
Period | 3/11/19 → 4/11/19 |
Bibliographical note
Publisher Copyright:© 2019 Association for Computational Linguistics.
Funding
We thank Julian Michael and Oz Anani for their useful comments and feedback. This research was supported by The U.S-Israel Binational Science Foundation grant 2016257, its associated NSF grant 1737230 and The Yandex Initiative for Machine Learning.
Funders | Funder number |
---|---|
Yandex Initiative for Machine Learning | |
National Science Foundation | 1737230 |
United States-Israel Binational Science Foundation | 2016257 |