Skip to content

Course project for NLP class, Fall 2021 at ETH Zurich: Assessing the robustness of the Seq2Seq neural machine translation.

Notifications You must be signed in to change notification settings

jyanqa/GRU-seq2seq-Model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 

Repository files navigation

GRU-seq2seq-Model

A course project for Natural Language Processing class, Fall 2021 at ETH Zurich.

Do noises break GRU Seq2Seq Translation Model?

  • The project aims to assess the robustness of the Seq2Seq neural machine translation by testing trained model on data which had been added different types of noises. The research is thus focused on presenting these features in the setting of French-English translation.

  • Two big general categories of noises including character-level and word-level are added to the clean texts.

  • Built model is tested on the 'dirty' data. BLEU-4 is the metric used to evaluate the performance NMT model.

  • The experiment reveals that word-level noises have less ability to break seq2seq model meanwhile randomizing character noises including Fully Random and Middle Random are two character-level noises whose the most negative affect to the model.

Data | Code | Report

About

Course project for NLP class, Fall 2021 at ETH Zurich: Assessing the robustness of the Seq2Seq neural machine translation.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages