@dallmann

Colorless Green Recurrent Networks Dream Hierarchically.

, , , , and . NAACL-HLT, page 1195-1205. Association for Computational Linguistics, (2018)

Abstract

Recurrent neural networks (RNNs) have achieved impressive results in a variety of lin- guistic processing tasks, suggesting that they can induce non-trivial properties of language. We investigate here to what extent RNNs learn to track abstract hierarchical syntactic struc- ture. We test whether RNNs trained with a generic language modeling objective in four languages (Italian, English, Hebrew, Russian) can predict long-distance number agreement in various constructions. We include in our evaluation nonsensical sentences where RNNs cannot rely on semantic or lexical cues (“The colorless green ideas ideas I ate with the chair sleep sleep furiously”), and, for Italian, we com- pare model performance to human intuitions. Our language-model-trained RNNs make re- liable predictions about long-distance agree- ment, and do not lag much behind human performance. We thus bring support to the hypothesis that RNNs are not just shallow- pattern extractors, but they also acquire deeper grammatical competence.

Links and resources

Tags

community

  • @dallmann
  • @dblp
@dallmann's tags highlighted