context2vec: Learning generic context embedding with bidirectional LSTM

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

415 Scopus citations

Abstract

Context representations are central to various NLP tasks, such as word sense disambiguation, named entity recognition, co-reference resolution, and many more. In this work we present a neural model for efficiently learning a generic context embedding function from large corpora, using bidirectional LSTM. With a very simple application of our context representations, we manage to surpass or nearly reach state-of-the-art results on sentence completion, lexical substitution and word sense disambiguation tasks, while substantially outperforming the popular context representation of averaged word embeddings. We release our code and pre-trained models, suggesting they could be useful in a wide variety of NLP tasks.

Original languageEnglish
Title of host publicationCoNLL 2016 - 20th SIGNLL Conference on Computational Natural Language Learning, Proceedings
PublisherAssociation for Computational Linguistics (ACL)
Pages51-61
Number of pages11
ISBN (Electronic)9781945626197
DOIs
StatePublished - 2016
Event20th SIGNLL Conference on Computational Natural Language Learning, CoNLL 2016 - Berlin, Germany
Duration: 11 Aug 201612 Aug 2016

Publication series

NameCoNLL 2016 - 20th SIGNLL Conference on Computational Natural Language Learning, Proceedings

Conference

Conference20th SIGNLL Conference on Computational Natural Language Learning, CoNLL 2016
Country/TerritoryGermany
CityBerlin
Period11/08/1612/08/16

Bibliographical note

Publisher Copyright:
© 2016 Association for Computational Linguistics.

Fingerprint

Dive into the research topics of 'context2vec: Learning generic context embedding with bidirectional LSTM'. Together they form a unique fingerprint.

Cite this