Difference between revisions of "User:Aboelhamd/progress"
Line 14: | Line 14: | ||
=== Week 1 (April 19 : April 25) === |
=== Week 1 (April 19 : April 25) === |
||
''' Day 1 (Friday April 19) ''' |
|||
Latest evaluation scores were lower than traditional apertium's LRLM resolution, by far, unfortunately.<br /> |
Latest evaluation scores were lower than traditional apertium's LRLM resolution, by far, unfortunately.<br /> |
||
Debugged the code to see what's the cause of such low evaluation score.<br /> |
Debugged the code to see what's the cause of such low evaluation score.<br /> |
||
Line 21: | Line 21: | ||
''' Day 2 (Saturday April 20) ''' |
|||
The evaluation results were better, but still not better than traditional apertium's LRLM resolution.<br /> |
The evaluation results were better, but still not better than traditional apertium's LRLM resolution.<br /> |
||
Debugged the code to see why the score still worse than apertium's.<br /> |
Debugged the code to see why the score still worse than apertium's.<br /> |
||
Line 38: | Line 38: | ||
''' Day 3 (Sunday April 21) ''' |
|||
The bug of some incomplete sentences was a problem of pointers. Before that, I was depending on pointers and the program was working good but a memory leaks problem emerged when I had done some changes, so instead of solving the leaks problems I turned to use stack instead of the heap, and with not enough testing after that, we got this bug.<br /> |
The bug of some incomplete sentences was a problem of pointers. Before that, I was depending on pointers and the program was working good but a memory leaks problem emerged when I had done some changes, so instead of solving the leaks problems I turned to use stack instead of the heap, and with not enough testing after that, we got this bug.<br /> |
||
Now, I will remove that stack solution again and turn to pointers and try to solve the leaks problem, so the program will work as well as previous version.<br /> |
Now, I will remove that stack solution again and turn to pointers and try to solve the leaks problem, so the program will work as well as previous version.<br /> |
||
Line 44: | Line 44: | ||
''' Day 4 (Monday April 22) ''' |
|||
Continue spa-eng training. <br/> |
Continue spa-eng training. <br/> |
||
Read the paper Neural Machine Translation with Extended Context. [https://arxiv.org/abs/1708.05943]<br /> |
Read the paper Neural Machine Translation with Extended Context. [https://arxiv.org/abs/1708.05943]<br /> |
||
''' Day 5 (Tuesday April 23) ''' |
|||
Continue spa-eng training. <br/> |
Continue spa-eng training. <br/> |
||
Debugged kir-tur pair for possible bug, as the chunker output has only default and unknown chunks. <br /> |
Debugged kir-tur pair for possible bug, as the chunker output has only default and unknown chunks. But still needs more debugging <br /> |
||
''' Day 6 (Wednesday April 24) ''' |
|||
Training is finished.<br /> |
|||
Modifying some of the code regarding evaluation, to update them with the last bug fixed.<br /> |
|||
Evaluation is finished, and the result scores still lower than apertium's.<br /> |
|||
{| class="wikitable" |
|||
⚫ | |||
|+ Using 100% of training data, 6-gram LM, max entropy models, sampling ambiguous combinations |
|||
! |
|||
! Apertium LRLM |
|||
! Apertium ambiguous |
|||
|- |
|||
! WER (Word Error Rate) |
|||
| 78.41 |
|||
| 76.93 |
|||
|- |
|||
! PER (Position-independent word Error Rate) |
|||
| 61.86 |
|||
| 57.62 |
|||
|- |
|||
! BLEU (Bi-Lingual Evaluation Understudy) |
|||
| 14.13 |
|||
| 13.72 |
|||
|} |
|||
<br /> |
|||
⚫ | |||
Working on training with 10%, 25%, 50%, 75% of the data, to evaluate the results with respect the data size.<br /> |
|||
=== Week 2 (April 26 : May 2) === |
=== Week 2 (April 26 : May 2) === |
Revision as of 08:19, 26 April 2019
GSOC 2019 progress : Extend weighted transfer rules
The code is uploaded regularly in this repo [1].
My working days will be everyday except for Thursday, 5 hours per day, at least for the first phase only.
Contents
- 1 Phase 1 (April 19 : May 16)
- 2 Phase 2 (June 28 : July 25)
- 3 Phase 3 (July 26 : August 19)
Phase 1 (April 19 : May 16)
From April 19 to May 16 and from June 21 to June 28.
It's shifted because of my exams.
Week 1 (April 19 : April 25)
Day 1 (Friday April 19)
Latest evaluation scores were lower than traditional apertium's LRLM resolution, by far, unfortunately.
Debugged the code to see what's the cause of such low evaluation score.
Figured out that there was a bug in normalizing the LM scores of the target ambiguous sentences. The LM score is a log base 10 the probability of the sentence, and as the magnitude gets higher, the sentence normalizing probability should get lower, and I was doing the inverse of that.
The easiest solution was to modify the score-sentences script, and instead of getting the score, will get its reciprocal.
Day 2 (Saturday April 20)
The evaluation results were better, but still not better than traditional apertium's LRLM resolution.
Debugged the code to see why the score still worse than apertium's.
Found a bug in generating the ambiguous combinations and still working on solving it.
The bug resulted some incomplete sentences like:
Sentence : Reciprocal of negative LM score
Resumption of the period of sessions : 0.0552026445652
Resumption of session period : 0.0739337147853
Resumption of the session period : 0.0753641871191
Resumption of the period of : 0.0757469108192
Resumption of period of : 0.0684245152522
Resumption of the period : 0.0809504851312
And as shown, the best score went to incomplete sentence, which is one of the reasons why we got bad evaluation score.
Day 3 (Sunday April 21)
The bug of some incomplete sentences was a problem of pointers. Before that, I was depending on pointers and the program was working good but a memory leaks problem emerged when I had done some changes, so instead of solving the leaks problems I turned to use stack instead of the heap, and with not enough testing after that, we got this bug.
Now, I will remove that stack solution again and turn to pointers and try to solve the leaks problem, so the program will work as well as previous version.
The bug was solved, and started training spa-eng pair again.
Day 4 (Monday April 22)
Continue spa-eng training.
Read the paper Neural Machine Translation with Extended Context. [2]
Day 5 (Tuesday April 23)
Continue spa-eng training.
Debugged kir-tur pair for possible bug, as the chunker output has only default and unknown chunks. But still needs more debugging
Day 6 (Wednesday April 24)
Training is finished.
Modifying some of the code regarding evaluation, to update them with the last bug fixed.
Evaluation is finished, and the result scores still lower than apertium's.
Apertium LRLM | Apertium ambiguous | |
---|---|---|
WER (Word Error Rate) | 78.41 | 76.93 |
PER (Position-independent word Error Rate) | 61.86 | 57.62 |
BLEU (Bi-Lingual Evaluation Understudy) | 14.13 | 13.72 |
Day 7 (Thursday April 25)
Working on training with 10%, 25%, 50%, 75% of the data, to evaluate the results with respect the data size.
Week 2 (April 26 : May 2)
Day 1 (Friday April 26)
Day 2 (Saturday April 27)
Day 3 (Sunday April 28)
Day 4 (Monday April 29)
Day 5 (Tuesday April 30)
Day 6 (Wednesday May 1)
Day 7 (Thursday May 2)
Week 3 (May 3 : May 9)
Day 1 (Friday May 3)
Day 2 (Saturday May 4)
Day 3 (Sunday May 5)
Day 4 (Monday May 6)
Day 5 (Tuesday May 7)
Day 6 (Wednesday May 8)
Day 7 (Thursday May 9)
Week 4 (May 10 : May 16)
Day 1 (Friday May 10)
Day 2 (Saturday May 11)
Day 3 (Sunday May 12)
Day 4 (Monday May 13)
Day 5 (Tuesday May 14)
Day 6 (Wednesday May 15)
Day 7 (Thursday May 16)
Week 5 (June 21 : June 27)
After my exams.
Day 1 (Friday June 21)
Day 2 (Saturday June 22)
Day 3 (Sunday June 23)
Day 4 (Monday June 24)
Day 5 (Tuesday June 25)
Day 6 (Wednesday June 26)
Day 7 (Thursday June 27)
Phase 2 (June 28 : July 25)
Week 1 (June 28 : July 4)
Day 1 (Friday June 28)
Day 2 (Saturday June 29)
Day 3 (Sunday June 30)
Day 4 (Monday July 1)
Day 5 (Tuesday July 2)
Day 6 (Wednesday July 3)
Day 7 (Thursday July 4)
Week 2 (July 5 : July 11)
Day 1 (Friday July 5)
Day 2 (Saturday July 6)
Day 3 (Sunday July 7)
Day 4 (Monday July 8)
Day 5 (Tuesday July 9)
Day 6 (Wednesday July 10)
Day 7 (Thursday July 11)
Week 3 (July 12 : July 18)
Day 1 (Friday July 12)
Day 2 (Saturday July 13)
Day 3 (Sunday July 14)
Day 4 (Monday July 15)
Day 5 (Tuesday July 16)
Day 6 (Wednesday July 17)
Day 7 (Thursday July 18)
Week 4 (July 19 : July 25)
Day 1 (Friday July 19)
Day 2 (Saturday July 20)
Day 3 (Sunday July 21)
Day 4 (Monday July 22)
Day 5 (Tuesday July 23)
Day 6 (Wednesday July 24)
Day 7 (Thursday July 25)