In this paper, we propose a novel neural network model called RNN
Encoder-Decoder that consists of two recurrent neural networks (RNN). One RNN
encodes a sequence of symbols into a fixed-length vector representation, and
the other decodes the representation into another sequence of symbols. The
encoder and decoder of the proposed model are jointly trained to maximize the
conditional probability of a target sequence given a source sequence. The
performance of a statistical machine translation system is empirically found to
improve by using the conditional probabilities of phrase pairs computed by the
RNN Encoder-Decoder as an additional feature in the existing log-linear model.
Qualitatively, we show that the proposed model learns a semantically and
syntactically meaningful representation of linguistic phrases.
Beschreibung
Learning Phrase Representations using RNN Encoder-Decoder for
Statistical Machine Translation
%0 Generic
%1 cho2014learning
%A Cho, Kyunghyun
%A van Merrienboer, Bart
%A Gulcehre, Caglar
%A Bahdanau, Dzmitry
%A Bougares, Fethi
%A Schwenk, Holger
%A Bengio, Yoshua
%D 2014
%K reading
%T Learning Phrase Representations using RNN Encoder-Decoder for
Statistical Machine Translation
%U http://arxiv.org/abs/1406.1078
%X In this paper, we propose a novel neural network model called RNN
Encoder-Decoder that consists of two recurrent neural networks (RNN). One RNN
encodes a sequence of symbols into a fixed-length vector representation, and
the other decodes the representation into another sequence of symbols. The
encoder and decoder of the proposed model are jointly trained to maximize the
conditional probability of a target sequence given a source sequence. The
performance of a statistical machine translation system is empirically found to
improve by using the conditional probabilities of phrase pairs computed by the
RNN Encoder-Decoder as an additional feature in the existing log-linear model.
Qualitatively, we show that the proposed model learns a semantically and
syntactically meaningful representation of linguistic phrases.
@misc{cho2014learning,
abstract = {In this paper, we propose a novel neural network model called RNN
Encoder-Decoder that consists of two recurrent neural networks (RNN). One RNN
encodes a sequence of symbols into a fixed-length vector representation, and
the other decodes the representation into another sequence of symbols. The
encoder and decoder of the proposed model are jointly trained to maximize the
conditional probability of a target sequence given a source sequence. The
performance of a statistical machine translation system is empirically found to
improve by using the conditional probabilities of phrase pairs computed by the
RNN Encoder-Decoder as an additional feature in the existing log-linear model.
Qualitatively, we show that the proposed model learns a semantically and
syntactically meaningful representation of linguistic phrases.},
added-at = {2016-02-03T09:14:51.000+0100},
author = {Cho, Kyunghyun and van Merrienboer, Bart and Gulcehre, Caglar and Bahdanau, Dzmitry and Bougares, Fethi and Schwenk, Holger and Bengio, Yoshua},
biburl = {https://www.bibsonomy.org/bibtex/21b65748e52783133211bce750d8e0361/lrieger},
description = {Learning Phrase Representations using RNN Encoder-Decoder for
Statistical Machine Translation},
interhash = {a4bf56db9d1f80d8681c1b47de0569b3},
intrahash = {1b65748e52783133211bce750d8e0361},
keywords = {reading},
note = {cite arxiv:1406.1078Comment: EMNLP 2014},
timestamp = {2016-02-03T09:14:51.000+0100},
title = {Learning Phrase Representations using RNN Encoder-Decoder for
Statistical Machine Translation},
url = {http://arxiv.org/abs/1406.1078},
year = 2014
}