Abstract
Recent work has shown that neural-embedded word representations capture many relational similarities, which can be recovered by means of vector arithmetic in the embedded space. We show that Mikolov et al.’s method of first adding and subtracting word vectors, and then searching for a word similar to the result, is equivalent to searching for a word that maximizes a linear combination of three pairwise word similarities. Based on this observation, we suggest an improved method of recovering relational similarities, improving the state-of-the-art results on two recent word-analogy datasets. Moreover, we demonstrate that analogy recovery is not restricted to neural word embeddings, and that a similar amount of relational similarities can be recovered from traditional distributional word representations.
Original language | English |
---|---|
Title of host publication | CoNLL 2014 - 18th Conference on Computational Natural Language Learning, Proceedings |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 171-180 |
Number of pages | 10 |
ISBN (Electronic) | 9781941643020 |
DOIs | |
State | Published - 2014 |
Event | 18th Conference on Computational Natural Language Learning, CoNLL 2014 - Baltimore, United States Duration: 26 Jun 2014 → 27 Jun 2014 |
Publication series
Name | CoNLL 2014 - 18th Conference on Computational Natural Language Learning, Proceedings |
---|
Conference
Conference | 18th Conference on Computational Natural Language Learning, CoNLL 2014 |
---|---|
Country/Territory | United States |
City | Baltimore |
Period | 26/06/14 → 27/06/14 |
Bibliographical note
Publisher Copyright:© 2014 Association for Computational Linguistics.
Funding
∗Supported by the European Community’s Seventh Framework Programme (FP7/2007-2013) under grant agreement no. 287923 (EXCITEMENT).
Funders | Funder number |
---|---|
FP7/2007 | 287923 |
Seventh Framework Programme |