Vector space word representations are learned from distributional information
of words in large corpora. Although such statistics are semantically
informative, they disregard the valuable information that is contained in
semantic lexicons such as WordNet, FrameNet, and the Paraphrase Database. This
paper proposes a method for refining vector space representations using
relational information from semantic lexicons by encouraging linked words to
have similar vector representations, and it makes no assumptions about how the
input vectors were constructed. Evaluated on a battery of standard lexical
semantic evaluation tasks in several languages, we obtain substantial
improvements starting with a variety of word vector models. Our refinement
method outperforms prior techniques for incorporating semantic lexicons into
the word vector training algorithms.
Description
[1411.4166] Retrofitting Word Vectors to Semantic Lexicons
%0 Generic
%1 faruqui2014retrofitting
%A Faruqui, Manaal
%A Dodge, Jesse
%A Jauhar, Sujay K.
%A Dyer, Chris
%A Hovy, Eduard
%A Smith, Noah A.
%D 2014
%K mlnlp nlp retrofitting wordembeddings
%T Retrofitting Word Vectors to Semantic Lexicons
%U http://arxiv.org/abs/1411.4166
%X Vector space word representations are learned from distributional information
of words in large corpora. Although such statistics are semantically
informative, they disregard the valuable information that is contained in
semantic lexicons such as WordNet, FrameNet, and the Paraphrase Database. This
paper proposes a method for refining vector space representations using
relational information from semantic lexicons by encouraging linked words to
have similar vector representations, and it makes no assumptions about how the
input vectors were constructed. Evaluated on a battery of standard lexical
semantic evaluation tasks in several languages, we obtain substantial
improvements starting with a variety of word vector models. Our refinement
method outperforms prior techniques for incorporating semantic lexicons into
the word vector training algorithms.
@misc{faruqui2014retrofitting,
abstract = {Vector space word representations are learned from distributional information
of words in large corpora. Although such statistics are semantically
informative, they disregard the valuable information that is contained in
semantic lexicons such as WordNet, FrameNet, and the Paraphrase Database. This
paper proposes a method for refining vector space representations using
relational information from semantic lexicons by encouraging linked words to
have similar vector representations, and it makes no assumptions about how the
input vectors were constructed. Evaluated on a battery of standard lexical
semantic evaluation tasks in several languages, we obtain substantial
improvements starting with a variety of word vector models. Our refinement
method outperforms prior techniques for incorporating semantic lexicons into
the word vector training algorithms.},
added-at = {2018-05-14T10:16:35.000+0200},
author = {Faruqui, Manaal and Dodge, Jesse and Jauhar, Sujay K. and Dyer, Chris and Hovy, Eduard and Smith, Noah A.},
biburl = {https://www.bibsonomy.org/bibtex/22bc03bfd3e60b3d0e7f5f074cf237414/albinzehe},
description = {[1411.4166] Retrofitting Word Vectors to Semantic Lexicons},
interhash = {328fdf26ef5a8ce387e454ddc1461710},
intrahash = {2bc03bfd3e60b3d0e7f5f074cf237414},
keywords = {mlnlp nlp retrofitting wordembeddings},
note = {cite arxiv:1411.4166Comment: Proceedings of NAACL 2015},
timestamp = {2018-05-14T10:16:35.000+0200},
title = {Retrofitting Word Vectors to Semantic Lexicons},
url = {http://arxiv.org/abs/1411.4166},
year = 2014
}