Abstract
The tremendous success of word embeddings in improving the ability of computers to perform natural language tasks has shifted the research on language representation from word representation to focus on sentence representation. This shift introduced a plethora of methods for learning vector representations of sentences, many of them based on compositional methods over word embeddings. These vectors are used as features for subsequent machine learning tasks or for pretraining in the context of deep learning. However, not much is known about the properties that are encoded in these sentence representations and about the language information they encapsulate. Recent studies analyze the encoded representations and the kind of information they capture. In this paper, we analyze results from a previous study on the ability of models to encode basic properties such as content, order, and length. Our analysis led to new insights, such as the effect of word frequency or word distance on the ability to encode content and order.
Original language | English |
---|---|
Article number | 8030297 |
Journal | IBM Journal of Research and Development |
Volume | 61 |
Issue number | 4 |
State | Published - 1 Jul 2017 |
Bibliographical note
Publisher Copyright:© 2017 IBM.
Funding
We thank G. Lev, B. Carmeli, E. Mezuman, O. Sar-Shalom, M. Ninio, and E. Shaked from the Machine Learning Technologies group at the IBM Research - Haifa lab for the fruitful discussion and contributions. This work was supported in part by the CONSENSUS project, which is funded by the European Commission within its FP7 Programme (contract 611688 funded under call FP7-ICT-10-5.4).
Funders | Funder number |
---|---|
European Commission | FP7-ICT-10-5.4, 611688 |