User:Pmodi/GSOC 2020 proposal: Hindi-Punjabi

From Apertium
Jump to navigation Jump to search

Contact Information

Name: Priyank Modi
Email: priyankmodi99@gmail.com
Current Designation: Undergraduate Researcher in the LTRC Lab, IIIT Hyderabad (completing 6th semester/3rd year in April '20) and a Teaching Assistant for Linguistics courses
IRC: pmodi
Timezone: GMT +0530 hrs
Linkedin: https://www.linkedin.com/in/priyank-modi-81584b175/
Github: https://github.com/priyankmodiPM
Website: https://priyankmodipm.github.io/

Why I am interested in Apertium

Apertium is an Open Source Rule-based machine translation system. Being an undergrad researcher at the LTRC lab in IIIT-H currently working on understanding the nuances of Indian languages and developing systems which improve our analysis of the same, Machine Translation interests me because it’s a complex problem which tries to achieve something most people believe is only achievable by humans.

Translating data to other languages, and especially low - resource languages gives the speakers of those languages access to valuable data and can help in several domains, such as education, news, judiciary, etc. The dictionaries made in the process are crucial for low resource languages and can even help making spell checkers etc.

The most striking factor for me is the fact that while recent trends to find a solution to MT lean towards Neural Networks and Deep Learning, which fall short when it comes to resource-poor languages, Apertium looks to tackle this using a rule based approach. Not only is this beneficial because of the level of understanding it provides instead of simply blaming data for poor results, it actually shows that it can perform better for low resource languages(even for the pair I present in my proposal).

A tool which is rule-based and open source really helps the community with language pairs that are resource - poor and gives them free translations for their needs and that is why I want to work on improving on it. I want to work with Apertium and GSoC so I can contribute to an important Open Source Tool while also honing my own skills, and I hope to become a part of this amazing community of developers!

Which of the published tasks are you interested in? What do you plan to do?

Adopt an unreleased language pair. I plan on developing the Hindi-Punjabi language pair in both directions i.e. hin-pan and pan-hin. This'll involve improving the monolingual dictionaries for both languages, the hin-pan bilingual dictionary and writing suitable transfer rules to bring this pair to a releasable state.

My Proposal

Mentors in Contact

Francis Tyers, Hèctor Alòs i Font

Brief of deliverables

  • A morph based dictionary of Punjabi with ~16,000 words
  • Improvements(current rules and word pairs) and additions to hin-pan bidictionary
  • Lexical selection and transfer rules for the pair
  • Translator for hin-pan and pan-hin with WER <20%
  • Morphological disambiguator for the pair

I plan on achieving coverage close to hin-urd pair. In the ideal case, I plan on getting better results than this pair since I feel enough data is available and given some dedicated work is done for 3 months, the predicted results aren't very difficult to achieve.

Why Google and Apertium should sponsor it

  • Both Hindi and Punjabi are widely spoken languages, both by number of speakers and geographic spread. Despite that, Punjabi especially has very limited online resources.
  • Services like Google Translate give unsatisfactory results when it comes to translation of this pair(see Section 4.4) On the contrary, I was able to achieve close to human translation for some sentences using minimal rules and time(see Section 6 : Coding Challenge).
  • I believe the Apertium architecture is suited perfectly for this pair and can replace the current state-of-art translator for this pair.
  • This is an important project(since it adds diversity to Apertium and translation systems in general) which requires at least 2-3 months of dedicated work and will be an important resource. In addition to this since it'll be publicly available, it'll drive research in vernacular languages, even in my own case(see Section 5 : Skills)

How and who it will benefit in society

As mentioned above, the Apertium community is strongly committed to under-resourced and minoritised/marginalised languages and Google helps its own way via programs like GSoC and GCI. There exists a good amount of vernacular literature and scriptures which could be circulated to a larger group of people if this project is successful. It'll be an important open source dictionary resource for both languages. My larger aim from this project is to develop a chain of pairs covering Indian languages. Since Urdu and Punjabi share their roots, at least one more pair can be developed with minimum effort. I plan of working towards the Hindi-English pair next year since by then I'll have a good understanding of the architecture a cross-language-family pair.

Google Translate : Analysis and comparison

Google Translate provides an interface to translate the pair in question. I have analysed the results of the translation into Punjabi from Google. The numerical results(computed on a small set of sentences from the coding challenge. The human translation which has been reviewed by 3 annotators is also available in the repo) are given below(source-target):

  • hin-pan: 79.23% WER
  • hin-pan: 56.56% PER
  • pan-hin: 82.23% WER
  • pan-hin: 57.83% WER

The results are simply poor, especially when it comes to longer sentences with less frequently used words. It is rather easy to see that Google Translate doesn't try to capture the case or tense in sentences, rather picks the most commonly used form of that particular root. NER is very limited, in contrast to the Apertium module which captures it well(because of it's 3 stage transfer mechanism I believe). The use of English as intermediate(which seems to be the case here as well because some words translate to English and fail to convert to Punjabi maybe because of some errors in parsing,as pointed by Hector) causes problems, such as the incorrect choice of tense in the verbs, wrong choice/dissappearance of some pronouns and the inability to handle copula constructions as well as verbal clauses(especially when other words occur between two sub-clauses). Here is an example of some of these form the Hindi test text:

Original source text (Hindi):

गिरजा आज फिर उस औरत को साथ लाया था.वही दुबली पतली मोटी-मोटी आंखें तीखी नाक और सांवले रंग वाली औरत.
Girija brought that woman with him again today. The same thin, big-eyed, pointy nosed and dusky woman.

Google translation (Punjabi):

ਚਰਚ ਨੇ todayਰਤ ਨੂੰ ਅੱਜ ਵਾਪਸ ਲਿਆਇਆ, ਉਹੀ ਪਤਲੀ womanਰਤ ਜਿਹੜੀ ਸੰਘਣੀ ਅੱਖਾਂ, ਤਿੱਖੀ ਨੱਕ ਅਤੇ ਹਨੇਰਾ ਰੰਗ.
The Church brought back todayਰਤ today. The same thin woman which big-eyed, pointy nose and dark colored.

Translation achieved using Apertium model(Punjabi):

ਗਿਰਜਾ ਅਜ੍ਜ ਫਿਰ ਉਸ ਔਰਤ ਨੂੰ ਨਾਲ ਲਾਇਆ ਸੀ.ਉਹੀ ਦੁਬਲੀ ਪਤਲੀ ਮੋਟੀ-ਮੋਟੀ ਅੱਖਾਂ ਤਿਖੀ ਨੱਕ ਅਤੇ ਸਾਉਲੇ #ਰਂਗ ਵਾਲੀ ਔਰਤ.
Girija brought that woman with him again today. The same thin, big-eyed, pointy nosed and dusky woman.

It is not difficult to see that most translations provided by Google Translate lead to a change in meaning. One clear reason that can be seen is that Google Translate relies on the n-grams available to it, and in case of rarely used words, it fails to translate those and worse, fails to capture the tense. In complex sentences, the chunking(stage 1 and 2 as per apertium model) fails hence leading to a failure in capturing meaning and very often, even generating any syntactically correct sentence.

Current state of dictionaries

A released module already exists for Hindi(as part of the urd-hin pair). However there still exist a lot of anomalies in the Hindi mono-dictionary. I've compiled a preliminary version of list of some these here. Apart from these, the existing state of the hin-pan bi-dictionary also needs massive improvement. The first step of this project will be to revise these lists of issues and come up with a sustainable solution. It'll be crucial that the changes made, especially to the Hindi mono-dictionary do no affect the urd-hin pair(and the hindi-begali, hindi-marathi and hindi-gujarati pairs which also have little but some work done) in a negative way.

Resources

[to be added - under confirmation for public use]
Hindi-Punjabi Dictionary - wiktionary
Punjabi-Hindi dictionary - Glosbe (awaiting confirmation)
Punjabi Articles - Wikipedia
Punjabi Dictionary - Wiktionary
Wikidumps-punjabi 1
Wikidumps-punjabi 2
Wikidumps-punjabi 3
Wikidumps-hindi 1
Wikidumps-hindi 2
Wikidumps-hindi 3

Workplan

PHASE DURATION GOALS OF THE WEEK BIDIX WER Coverage
Post Application Period
  • START:April 6th
  • END:May 3rd
  • List and discuss implementation choices of hin-pan bidix and urd-hin pair
  • Reading up on the details of Transfer rules(whether or not a 3-stage transfer is the best way for this pair) and assigning weights
  • Finding Language Resources
  • Making Frequency lists
COMMUNITY BONDING PERIOD : CLOSED CATEGORIES
  • START:May 4th
  • END:May 24th
  • Function words(voc prn, cnj, det, prn, post, gen_endings)
  • Transfer rules for post-positions
COMMUNITY BONDING PERIOD : Adjectives
  • START:May 25th
  • END:May 31st
  • Punjabi mono-dictionary : adj and adv
  • Expanding bilingual dictionary
  • Lexical selection rules for adj and adv
Week ONE: Verbal Paradigms
  • START:June 1st
  • END:June 7th
  • Punjabi mono-dictionary : Verbal paradigms(vblex, vbser, vaux)
  • Expanding bilingual dictionary
  • Lexical selection rules for verbs
  • testvoc : adj, adv
~ 3,000
Week TWO: Dictionary Expansion
  • START:June 8th
  • END:June 14th
  • Expanding bilingual dictionary
  • Lexical selection rules
~ 5,000
Week THREE: Dictionary Expansion
  • START:June 15th
  • END:June 21st
  • Expanding bilingual dictionary
  • Lexical selection rules
~ 6,500 < 25% (hin-pan) > 65% (hin-pan)
>60% (pan-hin)
Week FOUR: More works on verbs and testing
  • START:June 15th
  • END:June 21st
  • Expanding bilingual dictionary
  • Lexical selection rules
  • Manual Disambiguation of rules hin-pan(src-trg)
~ 7,500
Week FIVE : focus on Nouns
  • START:June 22nd
  • END:June 28th
  • Expanding bilingual dictionary
  • Lexical selection rules
~ 9,000
Week SIX : Expanding Dictionaries
  • START:June 29th
  • END:July 5th
  • Expanding bilingual dictionary
  • Lexical selection rules

First Evaluation(June 29th - July 3rd)

~ 10,500
Week SEVEN : Expanding Dictionaries
  • START:July 6th
  • END:July 12th
  • Expanding bilingual dictionary
  • Lexical selection rules
  • Manual disambiguation of rules(pan-hin)
~ 12,000
Week EIGHT : Transfer rules(hin-pan)
  • START:July 13th
  • END:July 19th
  • Expanding bilingual dictionary
  • Lexical selection rules
  • Transfer rules(hin-pan)
~ 13,000
Week NINE : Transfer rules
  • START:July 20th
  • END:July 26th
  • Expanding bilingual dictionary
  • Lexical selection rules
  • Transfer rules : pan-hin
~ 14,000
Week TEN
  • START:July 27th
  • END:August 2nd
  • Expanding bilingual dictionary
  • Lexical selection rules

Second Evaluation(July 27th - July 31st)

~ 15,000 <20% (hin-pan)
<25% (pan-hin)
>75% (hin-pan)
>70% (pan-hin)
Week ELEVEN
  • START:August 3rd
  • END:August 9th
  • Expanding bilingual dictionary
  • Lexical selection rules
  • Disambiguation rules
  • Transfer rules
~ 16,000
Week TWELVE : Testvoc
  • START:August 10th
  • END:August 16th
  • Testvoc hin-pan
  • Add rules, words
~ 16,500
Week THIRTEEN : Finishing up
  • START:August 17th
  • END:August 23rd
  • Testvoc pan-hin
  • Add rules, words

Personal Code freeze : August 22nd

~ 17,000
Week FOURTEEN : Review
  • START:August 24th
  • END:August 30th
  • Review and documentation

Final evaluation(August 24th - August 31st)

~ 17,000 ~15% (hin-pan)
<20% (pan-hin)
~90% (hin-pan)
~83% (pan-hin)

Skills

I'm currently a third year(commencing start of April '20 ) student at IIIT Hyderabad where I'm studying Computational Linguistics. It is a dual degree course where we study Computer Science, Linguistics, NLP and more. I am also a teaching assistant for courses on Language Typology, Universals and Historical Linguistics this semester, so I understand linguistic concepts very well along with the handling of linguistic data.

I've been interested in linguistics from the very beginning and due to the rigorous programming courses, I'm also adept at several programming languages like Python, C++, XML, Bash Scripting, etc. I'm skilled in writing Algorithms. Data Structures, and Machine Learning Algorithms as well.

I also have a lot of experience studying and generating data which I feel is essential in solving any problem, especially the one mentioned in this proposal. My paper on 'Hindi TimeBank: An ISO-TimeML Annotated Reference Corpus' recently got accepted in 16th Joint ACL - ISO Workshop on Interoperable Semantic Annotation at LREC 2020. I am working on extending the same for Punjabi using Transfer learning. (ISA list of accepted papers, Link to paper)

I am also closely involved with the committee conducting Asia-Pacific Linguistics Olympiad(which holds a camp, mentors and prepares students for the International Linguistics Olympiad) and help with the organisation and judging for the same.

Due to the focused nature of our courses, I have worked in several projects, such as building Anaphora Resolution systems, Abstractive Summarizers(using Pointer-generators, hierarchical attention and transformers), POS Taggers, Named Entity Recognisers, simple Q-A systems, a Linux based shell etc. all of which required a working understanding of Natural Language Processing scripting. Some of these projects aren't available on GitHub because of the privacy settings but can be provided if required.

I am fluent in English, Hindi and Punjabi.

Coding challenge

I've completed the coding challenge for translation from Hindi into Punjabi. You can find my work here : Coding challenge repository
Original corpus : source lang-hin
Translated output : target lang-pan
Human Translation : target lang-pan(human)

Results : Source - Hindi, Target - Punjabi (evaluator output included in repo)
(to be checked and revised since WER and PER before and after removing unknown words remains same even though the error on not identifying unrecognized words was fixed after consulting @TinoDidriksen)
WER achieved : 15.30 %
PER achieved : 15.03 %

Currently I'm working on finishing my list on the errors I could find in the existing files(See Section 4.5 : Current state of dictionaries). Once this is complete, I'll go ahead exploring and discussing the AnnCorra scheme for covering some of these (link to paper) This scheme captures dependency relations in much more detail than UD(Universal Dependency). (See section 4.6 for details on why it's required). While I'm more than familiar with AnnCorra, I'll have to check how to integrate it in the apertium pipeline, that is if the mentors think it is useful.
Once this is complete, I'll finish the compilation of texts from the dumps to get statistical usage of words. I plan to finish all this before the community bonding period is midway, so that I can meet the deliverables as planned in the Workplan.

Non-Summer-of-Code plans for the Summer

Since I'll be having my college summer vacations for almost the entire duration of the project, I can easily spend 35-40 hours per week on the project. Since, the academic schedule might vary a little bit due to lock downs for prevention of COVID-19, I'll be starting work early and cover the problems in the post-application period. I've also kept workload slightly heavier in the first 2 weeks to cover up any unlikely, uncertain extensions in academics that might show up. Even then, I can spend around 20 hours a week in any case(note that this is a very unlikely situation and even then this period won't last more than a week since the coursework is already underway online and is expected to be over before start of the project).