Abstract

We introduce a class of probabilistic continuous translation models called Recurrent Continuous Translation Models that are purely based on continuous representations for words, phrases and sentences and do not rely on alignments or phrasal translation units. The models have a generation and a conditioning aspect. The generation of the translation is modelled with a target Recurrent Language Model, whereas the conditioning on the source sentence is modelled with a Convolutional Sentence Model. Through various experiments, we show first that our models obtain a perplexity with respect to gold translations that is > 43% lower than that of stateof-the-art alignment-based translation models. Secondly, we show that they are remarkably sensitive to the word order, syntax, and meaning of the source sentence despite lacking alignments. Finally we show that they match a state-of-the-art system when rescoring n-best lists of translations.

Keywords

PerplexityComputer scienceSentenceTranslation (biology)Natural language processingArtificial intelligenceSyntaxTransfer-based machine translationLanguage modelProbabilistic logicWord (group theory)Machine translationExample-based machine translationLinguistics

Affiliated Institutions

Related Publications

Publication Info

Year
2013
Type
article
Pages
1700-1709
Citations
1330
Access
Closed

External Links

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

1330
OpenAlex

Cite This

Nal Kalchbrenner, Phil Blunsom (2013). Recurrent Continuous Translation Models. , 1700-1709. https://doi.org/10.18653/v1/d13-1176

Identifiers

DOI
10.18653/v1/d13-1176