User:Aboelhamd

From Apertium
Revision as of 22:03, 4 April 2019 by Aboelhamd (talk | contribs)
Jump to navigation Jump to search

GSOC 2019 : Extend weighted transfer rules[1]

Personal Details

General Summary

I am Aboelhamd Aly, a 24 years old Egyptian computer engineer. My first tongue is Arabic and not hieroglyphic :) . I am currently living in Alexandria, Egypt, and I intend to study masters abroad after finishing my undergraduate study. I love languages, AI and hence NLP. I have some research and industry experience in NLP, machine learning, parallel programming and optimization. I have been working alongside Sevilay Bayatli (piraye) on introducing new module (weighted transfer rule) to apertium , and that encouraged me to choose the idea "Extend weighted transfer rules" to continue our work, extend, integrate and deploy the full module.


Contacts

Email : aboelhamd.abotreka@gmail.com
Facebook : https://www.facebook.com/aboelhamd
LinkedIn : https://www.linkedin.com/in/aboelhamd-aly-76384a102/
IRC : aboelhamd
Github : https://github.com/aboelhamd
Time zone : GMT+2


Education

I am a senior bachelor student at Alexandria University in Egypt. Recently I have been granted a scholarship to study masters in data science at Innopolis University in Russia. My undergraduate major is computer engineering, which dealt with everything in computers from the lowest level of zeros and ones to the highest level of HCI (human and computer interaction, mainly deals with user interface).
The subjects I loved the most were artificial intelligence, machine leaning, data mining and deep learning, and that's because I see very great potential in the AI field that could solve many of the problems humans face today.


Languages Love

I love languages very much, especially Arabic, because it's a very very beautiful language and of course because it's the language of our holy scripture (Quraan), which I memorize more than half of it. Also I love Arabic literature and I have wrote several Arabic poems and short stories. All of that gave me a very good knowledge of classical and modern Arabic morphology, syntax and derivation. After Arabic comes English which I also love very much but surely not proficient at it like Arabic.
And so my love to languages and AI lead me to work in natural language processing field to combine my passion and knowledge in it.


Last Year GSoC

Last year I tried to contribute in apertium to introduce a new pair (Arabic-Syriac) but I failed miserably, because I wasn't familiar at all with Syriac nor with apertium and also I began late in time which made me hasty, that I needed a less overwhelming project. I then applied to classical language tool-kit project to enhance some Classical Arabic functionalities there and that was my proposal[2]. Unfortunately I wasn't accepted in the program, though my mentor told me then, that Google gave them less spots than what they asked for :( , and that the other 3 applicants was postgraduate students that have more experience in the field and in open-source projects :( .
And after that I decided to contribute in an open-source project to gain both experiences and to try again next year, and here I am now :) .


Experience

Apertium

Sevilay and Me have been working into introducing the weighted transfer rules for months now. And we re-implemented a new module to handle ambiguous transfer rules , which parses, matches, applies transfer rules to the source and target sentence, then train maximum entropy models to be able to choose the best ambiguous rule for any given pattern.


Online courses

I had taken many online courses with wide spectrum of the computer engineering field. One that I am very proud of, is udacity's machine-learning nano-degree[2] which is a six-months program, consists of many courses and practical projects regarding machine learning.


Industry

Last summer I was hired as a software engineer intern in Brightskies tech. company. After the internship I was hired as a part-time software engineer.
Our team is working on parallel programming, optimization and machine learning projects. The 2 biggest companies we are working with are Intel and Aramco.
My role is working on understanding, implementing, optimizing some seismic algorithms and kernels, besides doing some research on some machine learning algorithms and topics.


Why interested in apertium ?

- I am very interested in NLP in general.
- Apertium has very noble goal, which is bringing languages with scarce data to life by linking them with machine translation of other languages.
- I have previous contribution in apertium and willing to build on it.



Project Idea

Weighted transfer rules

When we have more than one transfer rule that can be applied to a given pattern, we call this an ambiguous situation. Apertium resolve this ambiguous situation by choosing the left-to-right longest match (LRLM) rule/s to apply. And that's of course is not adequate with all the word/s that follow that pattern/s. To solve this problem we introduced a way to make this ambiguous rules weighted for certain word/s that follow the ambiguous pattern/s. And this is done by training very huge corpus to capture better expressive weights.
1- First we train an n-gram -we use n=5- source language model.
2- We apply all ambiguous transfer rules for each ambiguous pattern in the given sentence, separately from each other -we apply LRLM rules to all other ambiguous patterns-, and get score from the n-gram model for each of the ambiguous sentences for that given pattern.
3- These scores are then written in some files, each file contains the scores of an ambiguous pattern. These files are considered the datasets for the tool (we use yasmet) which train target language max entropy models.
4- After having the models, our module is now ready for use. By using beam search algorithm we choose the best possible target.
For more detailed explanation you could refer to this documentation[3].


Weighted transfer rules extension

Now that weighted transfer module we worked on was built to apply only chunker transfer rules. And the idea want to extend that to interchunk and postchunk transfer rules too. Both of them are similar to the chunker, but with some differences. For example interchunk def-cats will refer to the tags of the chunk itself and not the lexical forms it contains like chunker, and for postchunk they refer to name of the chunk and has nothing to do with tags now. Also chunk element has different use, because it deals now with chunks not words. Also there are some differences in clip element attributes between the three transfer files. All these differences may be considered minor with respect the whole module that handle the chunker transfer rules. And I think adding these modifications will not take long time.
So in addition to this extension, I think introducing new ideas or modifications that could enhance the accuracy and efficiency of the whole module could be the best thing to do. Also I think I may work in related or not related ideas to this one to make full use of the 3 months period.


Latest updates on WTR module

The module is now finished and is in the testing phase, it does well with Kazakh-Turkish pair and we hope it does as well with other pairs like Spanish-English pair which have more transfer rules than any other pair in apertium.
The latest code is uploaded in this repo[4]. The module is separated from apertium core, that is installing apertium only is not enough as one should download and install our module separately to use it along with apertium.


Coding Challenge

The coding challenge was to set up a pair and train the existing weighted transfer rule code. Which I have done several times while testing and debugging the code.
Since I didn't have a coding challenge and also the module was separated from apertium core as metioned before, Francis Tyers(spectei) told me integrate the module -without the training part- with apertium-transfer, and I did, in that pull-request[5].
Then he told me to make the module depends on libraries already used in apertium and not external ones, as I used 2 libraries pugixml to handle xml files and icu library to handle upper and lower cases, which are not used in apertium. Also Kevin Unhammer(unhammer) gave me some helpful review on the code, and I am currently resolving all these issues.


Additional thoughts

There are additional thoughts and modifications to the weighted transfer rules proposed in the aforementioned documentation[6].
And if some of them are valid, They could be applied along with the extension too.


Why google and apertium should sponsor it ?

- The project enhances apertium translation of all pairs making it closer to human translation.
- I have the right experience and qualifications to complete it successfully. And since I participated in building the module, I will easily be able to extend it.


How and who will it benefit in society ?

As the project will enhance apertium translation and make it closer to human translation, apertium will be more reliable and efficient to use in daily life and especially for document translation, which -in the long term- will enrich the data of languages with data scarcity, and hence help the speakers of such languages enriching and preserving their languages from extinction.


Other ideas ?

I would love to work on "Light alternative format for all XML files in an Apertium language pair"[7] idea along with weighted transfer rules idea too, if there was enough time. As there is an intersection between the two ideas which is the xml transfer files, and since I am already familiar with the documentation of these files, and has written module to handle, match and apply the rules, I think I could design another lighter format than xml, and write converters scripts between the two formats.
I hope in the in the next few days, I would be to able to finish the coding challenge of this idea so I could be considered working on it too if no other one applied to it.


Work plan

Exams and community bounding

I am having my final exams from May 27 to June 20 and it's almost exactly the same as the first phase of GSoC this year, and since I will not be able to work in my exams duration and even I want at least one free week before the first exam, I will start earlier, even before the announcement of accepted students, and that's not because I am sure that I will be accepted, but because I will continue contribution to the module anyways, if I got accepted or not. So I will start official coding on April 19 to May 16. And from May 17 to July 20 I will be taking my exams and I will still be able to do minor changes if necessary.


Schedule

First milestone

Week 1

(From April 19 - To April 25)

Read apertium2 document again, read deprecated or out of date parts from different sources and collect all the up to date documentation in one document.

Week 2

(From April 26 - To May 2)

Update and modify the ambiguous transfer file code to handle both inter- and post-chunk transfer files.

Week 3

(From May 3 - To May 9)

Continue code modifying and begin testing and debugging.

Week 4

(From May 10 - To May 16)

Finish coding, testing and debugging. Write documentation. Train one chosen pair and evaluate its accuracy.

Deliverable

Extended weighted transfer rules module.

After exams, spend days from June 21 to June 28 familiarize again with the code because my memory is not good enough :) . Also write the mentor evaluation, complete any unfinished documentation, tests or evaluations, and fix any reported issues or bugs.

Second milestone

Week 5

(From - To)

Week 6

(From - To)

Week 7

(From - To)

Week 8

(From - To)

Deliverable

Third milestone

Week 9

(From - To)

Week 10

(From - To)

Week 11

(From - To)

Week 12

(From - To)

Deliverable

Other summer plans