-
Notifications
You must be signed in to change notification settings - Fork 0
/
NOTES
46 lines (36 loc) · 1.66 KB
/
NOTES
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
Web links
========================================
http://www.statmt.org/moses/?n=Moses.Tutorial
http://www.statmt.org/moses/?n=Moses.Baseline
http://www.statmt.org/moses/?n=FactoredTraining.AlignWords
http://www.statmt.org/moses/?n=FactoredTraining.TrainingParameters
http://www.statmt.org/moses/?n=FactoredTraining.EMS
http://www.statmt.org/moses/uploads/Moses/config
http://www.statmt.org/moses/?n=FactoredTraining.RunGIZA
Relevant files
========================================
working/experiments/model/phrase-table.1.gz
working/experiments/lm/project-syndicate.lm.6
Experiment results
========================================
Sentences, cleaned and stuff:
working/experiments/corpus/project-syndicate.clean.1.en
working/experiments/corpus/project-syndicate.clean.1.fr
Given corpus files "news-commentary-v8.fr-en.{fr|en}" truncated to 5000
sentences, the standard Moses phrase-based translation netted a Bleu
score of:
alignment file: working/experiments/model/aligned.1.grow-diag-final-and
newstest2010: 15.45 (1.080) BLEU
Without tuning but with normal alignment:
alignment file: working/experiments/model/aligned.1.grow-diag-final-and
newstest2010: 14.66 (1.076) BLEU
Without tuning and with phrase-based alignment matrix:
alignment file: new-align.txt.grow-diag-final-and
newstest2010: 14.42 (1.074) BLEU
Without tuning and with simple GIZA++ intersection alignment:
alignment file: working/experiments/model/aligned.22.intersection
newstest2010: 14.31 (0.999) BLEU
Without tuning and with phrase-based alignment matrix built from simple
intersection:
alignment file: new-align.txt.intersection
newstest2010: 14.42 (1.074) BLEU