Difference between revisions of "Narimann/GSOC 2019 proposal: Kazakh-Turkish and Turkish-Kazakh"

From Apertium
Jump to navigation Jump to search
Line 18: Line 18:
 
I am interested in Apertium because it pays attention not only to common languages which have a lot of speakers around the world but also to these minority of languages which are not so popular and sometimes do not even have enough data to build a valuable translator.
 
I am interested in Apertium because it pays attention not only to common languages which have a lot of speakers around the world but also to these minority of languages which are not so popular and sometimes do not even have enough data to build a valuable translator.
   
Nowadays statistical machine translation(SMT) is very popular around the globe comparing with rule-based machine translation(RBMT). But the problem is that SMT requires a lot of data in the form of parallel languages corpora, since they very addicted to data, and many languages cannot afford it. While RBMT does not require so much data but requires a lot of effort to put in. From this point, we can conclude that Apertium is a good approach for machine translation purposes of small languages. Another point is that with a good and full implementation of a specific pair Apertium can reach accuracy comparable to big giants in this field.
+
Nowadays statistical machine translation(SMT) is very popular around the globe comparing with rule-based machine translation(RBMT). But the problem is that SMT requires a lot of data in the form of parallel languages corpora, since they very addicted to data, and many languages cannot afford it. While RBMT does not require so much data but requires a lot of effort to put in. From this point, we can conclude that Apertium is a good approach for machine translation purposes of small languages. Another point is that with a good and full implementation of a specific pair Apertium can reach accuracy comparable to big giants in this field such as Google or Yandex.
   
 
== Which of the published tasks are you interested in? What do you plan to do? ==
 
== Which of the published tasks are you interested in? What do you plan to do? ==

Revision as of 09:28, 1 April 2019

Contact Information

Name: Daniyar Nariman

Location: Kazan, Tatarstan

E-mail: n.daniyar@innopolis.ru, nariman9119@gmail.com

IRC: nariman

Github: https://github.com/nariman9119

Telegram: nariman9119

Why is it that you are interested in Apertium?

I am studying Computer Science at my university, Data Science track. I am very interested in machine translation and other stuff related to the NLP.

I am interested in Apertium because it pays attention not only to common languages which have a lot of speakers around the world but also to these minority of languages which are not so popular and sometimes do not even have enough data to build a valuable translator.

Nowadays statistical machine translation(SMT) is very popular around the globe comparing with rule-based machine translation(RBMT). But the problem is that SMT requires a lot of data in the form of parallel languages corpora, since they very addicted to data, and many languages cannot afford it. While RBMT does not require so much data but requires a lot of effort to put in. From this point, we can conclude that Apertium is a good approach for machine translation purposes of small languages. Another point is that with a good and full implementation of a specific pair Apertium can reach accuracy comparable to big giants in this field such as Google or Yandex.

Which of the published tasks are you interested in? What do you plan to do?

Proposal

List any non-Summer-of-Code plans you have for the Summer