Unsupervised Approaches
for Neural Machine Translation
Facebook: Howard Lo (羅右鈞)
GitHub: howardyclo
Outline
Motivation
Core Techniques
Unsupervised Machine Translation using Monolingual Corpora Only
Model
Objective 1: Denoising Autoencoder
Objective 2: Back-Translation (Cross Domain Training)
Objective 3: Adversarial Training
Final Objective
Unsupervised Training Algorithm
Model Selection
Since no parallel data for validation, they use the surrogate criterion:
Datasets
Baselines
Experiment Results
Experiment Results
Horizontal lines: Performance of unsupervised NMT that leverages 15 million monolingual sentences.
Close to the performance of supervised NMT trained on 100,000 parallel sentences.
Ablation Study
Unsupervised Neural Machine Translation
Similarities between the Papers
Differences between the Papers
Differences between the Papers
Differences between the Papers
Differences between the Papers
Quantitative Analysis in Paper2
Preserve keywords but lack of fluency and adequacy.
Lack of adequacy.
Beyond word-by-word translation.
Can model structural differences between languages.
Code
References
References