Difference between revisions of "Ideas for Google Summer of Code"

From Apertium
Jump to navigation Jump to search
m (Text replacement - "(chat|irc)\.freenode\.net" to "irc.oftc.net")
 
(135 intermediate revisions by 16 users not shown)
Line 2: Line 2:
 
This is the ideas page for [[Google Summer of Code]], here you can find ideas on interesting projects that would make Apertium more useful for people and improve or expand our functionality. If you have an idea please add it below, if you think you could mentor someone in a particular area, add your name to "Interested mentors" using <nowiki>~~~</nowiki>
 
This is the ideas page for [[Google Summer of Code]], here you can find ideas on interesting projects that would make Apertium more useful for people and improve or expand our functionality. If you have an idea please add it below, if you think you could mentor someone in a particular area, add your name to "Interested mentors" using <nowiki>~~~</nowiki>
   
The page is intended as an overview of the kind of projects we have in mind. If one of them particularly piques your interest, please come and discuss with us on <code>#apertium</code> on <code>irc.freenode.net</code>, mail the [[Contact|mailing list]], or draw attention to yourself in some other way.
+
The page is intended as an overview of the kind of projects we have in mind. If one of them particularly piques your interest, please come and discuss with us on <code>#apertium</code> on <code>irc.oftc.net</code>, mail the [[Contact|mailing list]], or draw attention to yourself in some other way.
   
 
Note that, if you have an idea that isn't mentioned here, we would be very interested to hear about it.
 
Note that, if you have an idea that isn't mentioned here, we would be very interested to hear about it.
Line 13: Line 13:
 
* Resources that could be converted or expanded in the [[incubator]]. Consider doing or improving a language pair (see [[incubator]], [[nursery]] and [[staging]] for pairs that need work)
 
* Resources that could be converted or expanded in the [[incubator]]. Consider doing or improving a language pair (see [[incubator]], [[nursery]] and [[staging]] for pairs that need work)
 
* Unhammer's [[User:Unhammer/wishlist|wishlist]]
 
* Unhammer's [[User:Unhammer/wishlist|wishlist]]
  +
* The open issues [https://github.com/search?q=org%3Aapertium&state=open&type=Issues on Github] - especially the [https://github.com/search?q=org%3Aapertium+label%3A%22good+first+issue%22&state=open&type=Issues Good First Issues].
* The [http://sourceforge.net/p/apertium/tickets/search/?q=!status%3Awont-fix+%26%26+!status%3Aclosed open tickets] page on SourceForge
 
   
 
__TOC__
 
__TOC__
   
  +
If you're a student trying to propose a topic, the recommended way is to request a wiki account and then go to <pre>http://wiki.apertium.org/wiki/User:[[your username]]/GSoC2021Proposal</pre> and click the "create" button near the top of the page. It's also nice to include <code><nowiki>[[Category:GSoC_2021_student_proposals]]</nowiki></code> to help organize submitted proposals.
== List ==
 
   
  +
== Ideas ==
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
=== <u>Bring a released language pair up to state-of-the-art quality</u> ===
 
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
* '''Required skills''':<br>XML, a scripting language (Python, Perl), good knowledge of the language pair adopted.
 
* '''Description''':<br>Take a released language pair, and drastically improve the performance both in terms of coverage, and in terms of translation quality. This will involve working with dictionaries, transfer rules, scripting, corpora. The objective is to make an Apertium language pair state-of-the-art, or close to state-of-the-art in terms of translation quality. This will involve improving coverage to 95-98% on a range of corpora and decreasing [[word error rate]] by 30-50%. For example if the current word error rate is 30%, then it should be reduced to 15-20%.
 
* '''Rationale''':<br>Apertium has quite a broad coverage of language pairs, but few of these pairs offer state-of-the-art translation quality. We think broad is important, but deep coverage is important too.
 
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]], [[User:Mlforcada|Mikel Forcada]], [[User:Xavivars|Xavi Ivars]]
 
* '''[[/Make a language pair state-of-the-art|read more...]]'''
 
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Python API for Apertium
  +
| difficulty = medium
  +
| skills = C++, Python
  +
| description = Update the Python API for Apertium to expose all Apertium modes and test with all major OSes
  +
| rationale = The current Python API misses out on a lot of functionality, like phonemicisation, segmentation, and transliteration, and doesn't work for some OSes <s>like Debian</s>.
  +
| mentors = [[User:Francis Tyers|Francis Tyers]]
  +
| more = /Python API
  +
}}
   
  +
{{IdeaSummary
=== <u>Adopt an unreleased language pair</u> ===
 
  +
| name = OmniLingo and Apertium
* '''Difficulty''':<br><span style="background-color: #cdcdef">3. Entry level</span>
 
  +
| difficulty = medium
* '''Required skills''':<br>XML, a scripting language (Python, Perl), good knowledge of the language pair adopted.
 
  +
| skills = JS, Python
* '''Description''':<br>Take on an orphaned unreleased language pair, and bring it up to release quality results. What this quality will be will depend on the language pair adopted, and will need to be discussed with the prospective mentor. This will involve writing linguistic data (including morphological rules and transfer rules &mdash; which are specified in a declarative language &mdash; and possibly [[Constraint Grammar]] rules if that is relevant)
 
  +
| description = OmniLingo is a language learning system for practising listening comprehension using Apertium data. There is a lot of text processing involved (for example tokenisation) that could be aided by Apertium tools.
* '''Rationale''':<br>Apertium has a few pairs of languages (e.g. mt-he, ga-gd, ur-hi, pl-cs, sh-ru, etc...) that are orphaned, they don't have active maintainers. A lot of these pairs have a lot of work already put in, just need another few months to get them to release quality. See also [[Incubator]]
 
  +
| rationale =
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]], [[User:Jimregan|Jimregan]], [[User:Kevin Scannell|Kevin Scannell]], [[User:Trondtr|Trondtr]], [[User:Unhammer|Unhammer]], [[User:Darthxaher|Darthxaher]], [[User:Firespeaker|Firespeaker]], [[User:Hectoralos|Hectoralos]], [[User:Krvoje|Hrvoje Peradin]], [[User:Jacob Nordfalk|Jacob Nordfalk]], [[User:Mlforcada|Mikel Forcada]], [[User:Vin-ivar|Vinit Ravishankar]], [[User:Aida|Aida Sundetova]], [[User:Xavivars|Xavi Ivars]]
 
  +
| mentors = [[User:Francis Tyers|Francis Tyers]]
* '''[[/Adopt a language pair|read more...]]'''
 
  +
| more = /OmniLingo
</div>
 
  +
}}
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Web API extensions
  +
| difficulty = medium
  +
| skills = Python
  +
| description = Update the web API for Apertium to expose all Apertium modes
  +
| rationale = The current Web API misses out on a lot of functionality, like phonemicisation, segmentation, and transliteration
  +
| mentors = [[User:Francis Tyers|Francis Tyers]], [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
  +
| more = /Apertium APY
  +
}}
   
  +
{{IdeaSummary
=== <u>Extend lttoolbox to have the power of HFST</u> ===
 
  +
| name = Develop a morphological analyser
* '''Difficulty''':<br><span style="background-color: #efcdcd">1. Hard</span>
 
  +
| difficulty = easy
* '''Required skills''':<br>C++, XSLT, XML
 
  +
| skills = XML or HFST or lexd
* '''Description''':<br>Extend lttoolbox (perhaps writing a preprocessor for it) so that it can be used to do the morphological transformations currently done with HFST. And yes, of course, writing something that translates the current HFST format to the new lttolbox format. Proof of concept: Come up with a new format that can express all of the features found in the Kazakh transducer; implement this format in Apertium; Implement the Kazakh transducer in this format and integrate it in the English--Kazakh pair.
 
  +
| description = Write a morphological analyser and generator for a language that does not yet have one
* '''Rationale''':<br>Some language pairs in Apertium use HFST where most language pairs use Apertium's own lttoolbox. This is due to the fact that writing morphologies for languages that have features such as the vowel harmony found in Turkic languages is very hard with the current format supported by lttoolbox. The mixture of HFST and lttoolbox makes it harder for people to develop some language pairs.
 
  +
| rationale = A key part of an Apertium machine translation system is a morphological analyser and generator. The objective of this task is to create an analyser for a language that does not yet have one.
* '''Mentors''':<br>[[User:mlforcada|Mikel Forcada]], [[User:TommiPirinen|Tommi A Pirinen]], [[User:Unhammer]], [[User:Mlforcada|Mikel Forcada]], mentors wanted
 
  +
| mentors = [[User:Francis Tyers|Francis Tyers]], [[User:Firespeaker|Jonathan Washington]], [[User: Sevilay Bayatlı|Sevilay Bayatlı]], Hossep, nlhowell
* '''[[/Extend lttoolbox to have the power of HFST|read more]]'''
 
  +
| more = /Morphological analyser
</div>
 
  +
}}
   
  +
{{IdeaSummary
<!--
 
  +
| name = Support for Enhanced Dependencies in UD Annotatrix
  +
| difficulty = medium
  +
| skills = NodeJS
  +
| description = UD Annotatrix is an annotation interface for Universal Dependencies, but does not yet support all functionality
  +
| rationale =
  +
| mentors = [[User:Francis Tyers|Francis Tyers]]
  +
| more = /Morphological analyser
  +
}}
   
  +
{{IdeaSummary
DANGER TERROR HORROR !!!!!!
 
  +
| name = User-friendly lexical selection training
  +
| difficulty = Medium
  +
| skills = Python, C++, shell scripting
  +
| description = Make it so that training/inference of lexical selection rules is a more user-friendly process
  +
| rationale = Our lexical selection module allows for inferring rules from corpora and word alignments, but the procedure is currently a bit messy, with various scripts involved that require lots of manual tweaking, and many third party tools to be installed. The goal of this task is to make the procedure as user-friendly as possible, so that ideally only a simple config file would be needed, and a driver script would take care of the rest.
  +
| mentors = [[User:Unhammer|Unhammer]], [[User:Mlforcada|Mikel Forcada]]
  +
| more = /User-friendly lexical selection training
  +
}}
   
  +
{{IdeaSummary
The task above has subsumed these two
 
  +
| name = Robust tokenisation in lttoolbox
  +
| difficulty = Medium
  +
| skills = C++, XML, Python
  +
| description = Improve the longest-match left-to-right tokenisation strategy in [[lttoolbox]] to be fully Unicode compliant.
  +
| rationale = One of the most frustrating things about working with Apertium on texts "in the wild" is the way that the tokenisation works. If a letter is not specified in the alphabet, it is dealt with as whitespace, so e.g. you get unknown words split in two so you can end up with stuff like ^G$ö^k$ı^rmak$ which is terrible for further processing.
  +
| mentors = [[User:Francis Tyers|Francis Tyers]], [[User:TommiPirinen|Flammie]]
  +
| more = /Robust tokenisation
  +
}}
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = apertium-separable language-pair integration
=== <u>Flag diacritics in lttoolbox</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #efcdcd">1. Hard</span>
 
  +
| skills = XML, a scripting language (Python, Perl), some knowledge of linguistics and/or at least one relevant natural language
* '''Required skills''':<br>C++ or Java, XML, Knowledge of FSTs
 
  +
| description = Choose a language you can identify as having a good number of "multiwords" in the lexicon. Modify all language pairs in Apertium to use the [[Apertium-separable]] module to process the multiwords, and clean up the dictionaries accordingly.
* '''Description''':<br>Adapt [[lttoolbox]] to elegantly use flag diacritics. Flag diacritics are a way of avoiding transducer size blow-up by discarding impossible paths at runtime as opposed to compile time. Some work has already been done, see [[Flag diacritics]].
 
  +
| rationale = Apertium-separable is a newly developed module to process lexical items with discontinguous dependencies, an area where Apertium has traditionally fallen short. Despite all the module has to offer, it has only been put to use in small test cases, and hasn't been integrated into any translation pair's development cycle.
* '''Rationale''':<br>This will involve designing some changes to our XML dictionary format (see [[lttoolbox]], and implementing the associated changes in the FST compiling processing code. The reason behind this is that many languages have prefix inflection, and we cannot currently deal with this without either making paradigms useless, or overanalysing (e.g. returning analyses where none exist). Flag diacritics (or constraints) would allow us to restrict overanalysis without blowing up the size of our dictionaries.
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]]
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]] (C++), [[User:Jacob Nordfalk|Jacob Nordfalk]] (Java)
 
  +
| more = /Apertium separable
* '''[[/Flag diacritics in lttoolbox|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
  +
| name = UD and Apertium integration
  +
| difficulty = Entry level
  +
| skills = python, javascript, HTML, (C++)
  +
| description = Create a range of tools for making Apertium compatible with Universal Dependencies
  +
| rationale = Universal dependencies is a fast growing project aimed at creating a unified annotation scheme for treebanks. This includes both part-of-speech and morphological features. Their annotated corpora could be extremely useful for Apertium for training models for translation. In addition, Apertium's rule-based morphological descriptions could be useful for software that relies on Universal dependencies.
  +
| mentors = [[User:Francis Tyers]] [[User:Firespeaker| Jonathan Washington]]
  +
| more = /UD and Apertium integration
  +
}}
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = rule visualization tools
=== <u>Weights in lttoolbox</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
  +
| skills = python? javascript? XML
* '''Required skills''':<br>C++, XML, FSTs
 
  +
| description = make tools to help visualize the effect of various rules
* '''Description''':<br>[[lttoolbox]] is a set of tools for building finite-state transducers. As part of Apertium's long-term strategy we would like to include probabilistic information into more stages of the pipeline to allow generic tools to be optimised for machine translation. This task involves adding the possibility of weighting lexemes and analyses in our finite-state transducer toolbox.
 
  +
| rationale = TODO see https://github.com/Jakespringer/dapertium for an example
* '''Rationale''':<br>Weighting information for lexical forms will be useful for morphological disambiguation, and for work on [[spellchecking]].
 
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]], [[User:TommiPirinen|Tommi Pirinen]]
+
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Sevilay Bayatlı|Sevilay Bayatlı]]
  +
| more = /Visualization tools
* '''[[/Add weights to lttoolbox|read more...]]'''
 
  +
}}
</div>
 
-->
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = dictionary induction from wikis
=== <u>Discontiguous multiwords</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
  +
| skills = MySQL, mediawiki syntax, perl, maybe C++ or Java; Java, Scala, RDF, and DBpedia to use DBpedia extraction
* '''Required skills''':<br>C++, Knowledge of FSTs
 
  +
| description = Extract dictionaries from linguistic wikis
* '''Description''':<br>The task will be to develop, or adapt a module to deal with these kind of contiguous multiword expressions, for example, taking 'liggja ekki fyrir' and reordering it as 'liggja# fyrir ekki'.
 
  +
| rationale = Wiki dictionaries and encyclopedias (e.g. omegawiki, wiktionary, wikipedia, dbpedia) contain information (e.g. bilingual equivalences, morphological features, conjugations) that could be exploited to speed up the development of dictionaries for Apertium. This task aims at automatically building dictionaries by extracting different pieces of information from wiki structures such as interlingual links, infoboxes and/or from dbpedia RDF datasets.
* '''Rationale''':<br>In many languages, such as English, Norwegian and Icelandic, there are discontiguous multiwords, e.g. phrasal verbs, that we cannot easily support. For example 'liggja ekki fyrir' in Icelandic should be translated in English as 'to be not clear', but we cannot have 'liggja fyrir' as a traditional multiword because of the extra 'adverb', or it could even be a whole NP.
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]]
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]]
 
  +
| more = /Dictionary induction from wikis
* '''[[/Discontiguous multiwords|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = unit testing framework
=== <u>Robust recursive transfer</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #efcdcd">1. Hard</span>
 
  +
| skills = perl
* '''Required skills''':<br>Python, XML, linguistics
 
  +
| description = adapt https://github.com/TinoDidriksen/regtest for general Apertium use. [https://github.com/TinoDidriksen/regtest/wiki Screenshots of regtest action]
* '''Description''':<br>The purpose of this task would be to create a module to replace the apertium-transfer module(s) which will parse and allow transfer operations on an input.
 
  +
| rationale = We are gradually improving our quality control, with (semi-)automated tests, but these are done on the Wiki on an ad-hoc basis. Having a unified testing framework would allow us to be able to more easily track quality improvements over all language pairs, and more easily deal with regressions.
* '''Rationale''':<br>Currently we have a problem with very distantly related languages that have long-distance constituent reordering, because we can only do finite-state chunking.
 
  +
| mentors = [[User:Xavivars|Xavi Ivars]]
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]], [[User:Sortiz|Sortiz]], [[User:Mlforcada|Mikel Forcada]], [[User:Japerez|Juan Antonio Pérez]]
 
  +
| more = /Unit testing
* '''[[/Robust recursive transfer|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Bring an unreleased translation pair to releasable quality
=== <u>Extend weighted transfer rules</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #efcdcd">1. Hard</span>
 
  +
| skills = shell scripting
* '''Required skills''':<br>Python, C++, linguistics
 
  +
| description = Take an unstable language pair and improve its quality, focusing on testvoc
* '''Description''':<br>The purpose of this task is to extend weighted transfer rules to all transfer files and to allow conflicting rule patterns to be handled by combining (lexicalised) weights.
 
  +
| rationale = Many Apertium language pairs have large dictionaries and have otherwise seen much development, but are not of releasable quality. The point of this project would be bring one translation pair to releasable quality. This would entail obtaining good naïve coverage and a clean [[testvoc]].
* '''Rationale''':<br>Currently our transfer rules are applied longest-match left-to-right ([[LRLM]]). When two rule patterns conflict the first one is chosen. We have a prototype for selecting based on lexicalised weights, but it only applies to the first stage of transfer.
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Seviay Bayatlı|Sevilay Bayatlı]], [[User:Hectoralos|Hèctor Alòs i Font]]
* '''Mentors''':<br>[[User:Francis Tyers|Francis Tyers]], [[User:TommiPirinen|Tommi Pirinen]]
 
  +
| more = /Make a language pair state-of-the-art
* '''[[/Weighted transfer rules|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Develop a prototype MT system for a strategic language pair
  +
| difficulty = Medium
  +
| skills = XML, some knowledge of linguistics and of one relevant natural language
  +
| description = Create a translation pair based on two existing language modules, focusing on the dictionary and structural transfer
  +
| rationale = Choose a strategic set of languages to develop an MT system for, such that you know the target language well and morphological transducers for each language are part of Apertium. Develop an Apertium MT system by focusing on writing a bilingual dictionary and structural transfer rules. Expanding the transducers and disambiguation, and writing lexical selection rules and multiword sequences may also be part of the work. The pair may be an existing prototype, but if it's a heavily developed but unreleased pair, consider applying for "Bring an unreleased translation pair to releasable quality" instead.
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Sevilay Bayatlı| Sevilay Bayatlı]], [[User:Hectoralos|Hèctor Alòs i Font]]
  +
| more = /Adopt a language pair
  +
}}
   
  +
{{IdeaSummary
=== <u>Automatic blank handling</u> ===
 
  +
| name = Website Improvements: Misc
* '''Difficulty''':<br><span style="background-color: #cdefcd">3. Hard</span>
 
  +
| difficulty = Medium
* '''Required skills''':<br>C++, XML
 
  +
| skills = html, css, js, python
* '''Description''':<br>Let the C++ modules deal with formatting, treating some tags as word-bound.
 
  +
| description = Improve elements of Apertium's web infrastructure
* '''Rationale''':<br>Our current handling of formatting/markup (html, odt, docx, latex) is brittle, requiring transfer rules to explicitly deal with blanks (e.g. markup), and some times inevitably outputting them in the wrong order. This project is implement ''automatic'' blank-handling, by treating some blanks as "glued to words" (e.g. italics, emphasis), and others as paragraph-level blanks.
 
  +
| rationale = Apertium's website infrastructure [[Apertium-html-tools]] and its supporting API [[APy|Apertium APy]] have numerous open issues. This project would entail choosing a subset of open issues and features that could realistically be completed in the summer. You're encouraged to speak with the Apertium community to see which features and issues are the most pressing.
* '''Mentors''':<br>[[User:Unhammer|Unhammer]], [[User:Tino Didriksen|Tino Didriksen]], [[User:Mlforcada|Mikel Forcada]]
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
* '''[[/superblank_handling_algorithm|read more #1]], [[/Automatic blank handling|read more #2]]'''
 
  +
| more = /Website improvements
</div>
 
  +
}}
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Website Improvements: Dictionary Lookup
  +
| difficulty = Medium
  +
| skills = html, css, js, python
  +
| description = Finish implementing dictionary lookup mode in Apertium's web infrastructure
  +
| rationale = Apertium's website infrastructure [[Apertium-html-tools]] and its supporting API [[APy|Apertium APy]] have numerous open issues, including half-completed features like dictionary lookup. This project would entail completing the dictionary lookup feature. Some additional features which would be good to work would include automatic reverse lookups (so that a user has a better understanding of the results), grammatical information (such as the gender of nouns or the conjugation paradigms of verbs), and information about MWEs. See [https://github.com/apertium/apertium-html-tools/issues/105 the open issue on GitHub].
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
  +
| more = /Website improvements
  +
}}
   
  +
{{IdeaSummary
=== <u>Improvements to the Apertium website</u> ===
 
  +
| name = Website Improvements: Spell checking
* '''Difficulty''':<br><span style="background-color: #cdcdef">3. Entry level</span>
 
  +
| difficulty = Medium
* '''Required skills''':<br>Python, HTML, JS
 
  +
| skills = html, js, css, python
* '''Description''':<br>Our web site is pretty cool already, but it's missing things like dictionary/synonym lookup, support for several variants of one language, reliability visualisation, (reliable) webpage translation, feedback, etc.
 
  +
| description = Add a spell-checking interface to Apertium's web tools
* '''Rationale''':<br>[https://apertium.org https://apertium.org] is what most people know us by, it should show off more of the things we are capable of :-)
 
  +
| rationale = [[Apertium-html-tools]] has seen some prototypes for spell-checking interfaces (all in stale PRs and branches on GitHub), but none have ended up being quite ready to integrate into the tools. This project would entail polishing up or recreating an interface, and making sure [[APy]] has a mode that allows access to Apertium voikospell modules. The end result should be a slick, easy-to-use interface for proofing text, with intuitive underlining of text deemed to be misspelled and intuitive presentation and selection of alternatives. [https://github.com/apertium/apertium-html-tools/issues/390 the open issue on GitHub]
* '''Mentors''':<br>[[User:Firespeaker|Jonathan]]
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
* '''[[/Apertium website improvements|read more...]]'''
 
  +
| more = /Spell checker web interface
</div>
 
  +
}}
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Website Improvements: Suggestions
=== <u>User-friendly lexical selection training</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
* '''Required skills''':<br>Python, C++, shell scripting
+
| skills = html, css, js, python
  +
| description = Finish implementing a suggestions interface for Apertium's web infrastructure
* '''Description''':<br>Make it so that training/inference of lexical selection rules is a more user-friendly process
 
  +
| rationale = Some work has been done to add a "suggestions" interface to Apertium's website infrastructure [[Apertium-html-tools]] and its supporting API [[APy|Apertium APy]], whereby users can suggest corrected translations. This project would entail finishing that feature. There are some related [https://github.com/apertium/apertium-html-tools/issues/55 issues] and [https://github.com/apertium/apertium-html-tools/pull/252 PRs] on GitHub.
* '''Rationale''':<br>Our lexical selection module allows for inferring rules from corpora and word alignments, but the procedure is currently a bit messy, with various scripts involved that require lots of manual tweaking, and many third party tools to be installed. The goal of this task is to make the procedure as user-friendly as possible, so that ideally only a simple config file would be needed, and a driver script would take care of the rest.
 
* '''Mentors''':<br>[[User:Unhammer|Unhammer]], [[User:Francis Tyers|Francis Tyers]], [[User:Mlforcada|Mikel Forcada]]
+
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
  +
| more = /Website improvements
* '''[[/User-friendly lexical selection training|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Website Improvements: Orthography conversion interface
=== <u>Light alternative format for all XML files in an Apertium language pair</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #efcdcd">1. Hard</span>
 
  +
| skills = html, js, css, python
* '''Required skills''':<br>Python, C++, shell scripting, XSLT, flex
 
  +
| description = Add an orthography conversion interface to Apertium's web tools
* '''Description''':<br>Make it possible to edit and develop language data using a format that is lighter than XML
 
  +
| rationale = Several Apertium language modules (like Kazakh, Kyrgyz, Crimean Tatar, and Hñähñu) have orthography conversion modes in their mode definition files. This project would be to expose those modes through [[APy|Apertium APy]] and provide a simple interface in [[Apertium-html-tools]] to use them.
* '''Rationale''':<br>In most Apertium language pairs, monolingual dictionaries, bilingual dictionaries, post-generation rule files and structural transfer rule files are all written in XML. While XML is easy to process due to explicit tagging of every element, it is tedious to deal with, particularly when it comes to structural transfer rules. Apertium's precursor, interNOSTRUM, had lighter text based formats. The task involves: (a) designing and documenting an interNOSTRUM-style format for all of the XML language data files in a language pair; (b) writing converters to XML and from XML that are fully roundtrip-compliant: (c) designing a way to synchronize changes when both the XML and the non-XML format are used simultaneously in a specific language pair.
 
* '''Mentors''':<br>[[User:Mlforcada|Mikel Forcada]], [[User:Japerez|Juan Antonio Pérez]], pair.
+
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]]
  +
| more = /Website improvements
* '''[[/Plain-text_formats_for_Apertium_data|read more...]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Apertium Browser Plugin
=== <u>Eliminate dictionary trimming</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #ffbdbd">0. Very Hard</span>
 
  +
| skills = html, css, js, python
* '''Required skills''':<br>C++, Finite-State Transducers
 
  +
| description = Expand functionality of Geriaoueg vocabulary assistant
* '''Description''':<br>Eliminate the need for trimming the monolingual dictionaries, in order to preserve and take advantage of maximal source language analysis.
 
  +
| rationale = [[Geriaoueg]] is a vocabulary assistant with Firefox/Chrom[e/ium] plugins. These plugins interface with Apertium's web API, [[APy|Apertium APy]], and allow a user to look up (in Apertium's dictionaries) word forms from a web page they're viewing. A Firefox/Chrom[e/ium] plugin should also be able to provide in-browser website translation. This project is to clean up the dictionary lookup functionality and add translation support to the plugins. Some APy features may need to be tweaked, but most of the work in this project will be solely in the plugins.
* '''Rationale''':<br>[[Why we trim]] mentions several technical reasons for why trimming away monolingual information is currently needed. Unfortunately, this limitation means that a lot of useful contextual information is lost. It would be ideal if the source language could be fully analyzed independent of target language, with any untranslated part fed back into the source language generator.
 
  +
| mentors = [[User:Firespeaker|Jonathan Washington]], [[User:Xavivars|Xavi Ivars]], [[User:Tino_Didriksen|Tino Didriksen]]
* '''Mentors''':<br>? ([[User:TommiPirinen|Tommi Pirinen a.k.a. Flammie]] interested, but low on time if alone)
 
  +
| more = /Geriaoueg browser plugin
* '''Work around everything in [[Why we trim]]'''
 
  +
}}
</div>
 
   
  +
{{IdeaSummary
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
  +
| name = Extend Weighted transfer rules
=== <u>Bilingual dictionary enrichment via graph completion</u> ===
 
  +
| difficulty = Medium
* '''Difficulty''':<br><span style="background-color: #ffbdbd">0. Very hard</span>
 
* '''Required skills''':<br>shell scripting, python, XSLT, XML
+
| skills = C++, python
  +
| description = The weighted transfer module is already applied to the chunker transfer rules. And the idea here is to extend that module to be applied to interchunk and postchunk transfer rules too.
* '''Description''':<br>Generate new entries for existing or new bilingual dictionaries using graphic representations of bilingual correspondences as found in all existing dictionaries (note that this idea defines a rather open-ended task to be discussed in detail with mentors).
 
  +
| rationale = As a resource see https://github.com/aboelhamd/Weighted-transfer-rules-module
* '''Rationale''':<br>Apertium bilingual dictionaries establish correspondences between lexical forms in a number of language pairs. Connections among them may be used to infer new entries for existing or new language pairs using graphs. The graphs may be directly generated from Apertium bidixes and exploiting using [[Bilingual_dictionary_discovery|ideas that had already been proposed in Apertium]] or using existing [http://linguistic.linkeddata.es/apertium/ RDF representations] of parts of their content, which may benefit from the information coming from being linked to other resources.
 
  +
| mentors = [[User: Sevilay Bayatlı|Sevilay Bayatlı]]
* '''Mentors''':<br>[[User:Mlforcada|Mikel Forcada]], [[User:Francis Tyers|Francis Tyers]], [[User:Jorge Gracia|Jorge Gracia]]
 
  +
| more = /Make a module
* '''[[Bilingual_dictionary_discovery|read more...]]''' '''[[/Bilingual_dictionary_enrichment_via_graph_completion|read even more...]]'''
 
  +
}}
</div>
 
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
=== <u>Transfer rule induction from comparable parsed corpora</u> ===
 
* '''Difficulty''':<br><span style="background-color: #ffbdbd">1. Hard</span>
 
* '''Required skills''':<br>shell scripting, python, XSLT, XML
 
* '''Description''':<br>A system to infer transfer rules from comparable corpora that have both been deeply parsed (with e.g. CG)
 
* '''Rationale''':<br>Many languages have good CG's and fairly large monolingual corpora, but little parallel material. Given a small bidix, fairly large monolingual corpora and good analysers/CG's, we should be able to parse both corpora, translate lemmas and look for similar sentences, turning the differences in their parses into transfer rules.
 
* '''Mentors''':<br>[[User:Unhammer]]
 
* '''[[Transfer_induction_from_comparable_parsed_corpora|read more...]]'''
 
</div>
 
<div>
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
=== <u>UD and Apertium integration</u> ===
 
* '''Difficulty''':<br><span style="background-color: #cdcdef">3. Entry level</span>
 
* '''Required skills''':<br>python, javascript, HTML, (C++)
 
* '''Description''':<br>Create a range of tools for making Apertium compatible with Universal Dependencies
 
* '''Rationale''':<br>Universal dependencies is a fast growing project aimed at creating a unified annotation scheme for treebanks. This includes both part-of-speech and morphological features. Their annotated corpora could be extremely useful for Apertium for training models for translation. In addition, Apertium's rule-based morphological descriptions could be useful for software that relies on Universal dependencies.
 
* '''Mentors''':<br>[[User:Francis Tyers]] [[User:Firespeaker]]
 
* '''[[/UD and Apertium integration|read more...]]'''
 
</div>
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
=== <u>Add weights to lttoolbox</u> ===
 
* '''Difficulty''':<br><span style="background-color: #ffbdbd">1. Hard</span>
 
* '''Required skills''':<br>c++
 
* '''Description''':<br>Add support for weighted transducers to lttoolbox
 
* '''Rationale''':<br>This will either involve implementing it from scratch or adding OpenFST as a backend. We would like to be able to use it both in the bilingual dictionaries, and in the morphological analysers, to be able to order analyses/translations by their probability/weight instead of by the random topological order.
 
* '''Mentors''':<br>[[User:Francis Tyers]] [[User:Unhammer]]
 
* '''[[/Add weights to lttoolbox|read more...]]'''
 
</div>
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
=== <u>Improving language pairs mining Mediawiki Content Translation postedits</u> ===
 
* '''Difficulty''':<br><span style="background-color: #ffbdbd">1. Hard</span>
 
* '''Required skills''':<br>Python, shell scripting, some statistics
 
* '''Description''':<br>Implement a toolkit that allows mining existing machine translation postediting data in [Mediawiki Content Translation https://www.mediawiki.org/wiki/Content_translation] to generate (as automatically as possible, and as complete as possible) monodix and bidix entries to improve the performance of an Apertium language pair. Data is available from Wikimedia content translation through an [API https://www.mediawiki.org/wiki/Content_translation/Published_translations#API] or in the form of [Dumps https://dumps.wikimedia.org/other/contenttranslation/] available in JSON and TMX format. This project is rather experimental and involves some research in addition to coding.
 
* '''Rationale''':<br>Apertium is used to generate new Wikipedia content: machine-translated content is postedited (and perhaps adapted) before publishing. Postediting information may contain information that can be used to help improve the lexical components of an Apertium language pair.
 
* '''Mentors''':<br> [[User:Mlforcada|Mikel Forcada]], (more mentors to be added)
 
* '''(more soon)'''<!---[[/Shallow-function labeller|read more...]]'''-->
 
</div>
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
 
=== <u>Shallow-function labeller</u> ===
 
 
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
* '''Required skills''':<br>Python, shell scripting
 
* '''Description''':<br>Implement a prototype shallow syntactic function labeller for Apertium
 
* '''Rationale''':<br>In many pairs it is useful to know in addition to the morphological tags of a word, syntactic function tags in order to make an adequate translation. For instance, you might want to know in an ergative language if an absolutive is subject or object while translating. A shallow function labeller takes an annotated corpus and produces a model which can annotate new text.
 
* '''Mentors''':<br>[[User:Unhammer|Unhammer]], [[User:Francis Tyers|Francis Tyers]], [[User:Mlforcada|Mikel Forcada]]
 
* '''[[/Shallow-function labeller|read more...]]'''
 
 
</div>
 
 
<div style="background-color: #f9f9f9; border: 1px solid black; padding: 1ex; margin-bottom: 2ex;">
 
=== <u>Unsupervised weighting of automata</u> ===
 
* '''Difficulty''':<br><span style="background-color: #cdefcd">2. Medium</span>
 
* '''Required skills''':<br>Python, shell scripting, statistics, finite-state transducers
 
* '''Description''':<br>Implement a collection of methods for weighting finite-state transducers, the methods should include an implementation of a simple method of supervised training, and a number of methods for unsupervised training. The objective being to get the analysis ranking given a set of a analyses for a given surface form as close to the result given by supervised training as possible.
 
* '''Rationale''':<br>Apertium struggles with ambiguity, we have had many attempts to write better part of speech taggers. This would complement those attempts by providing a generic method to weight automata.
 
* '''Mentors''':<br> [[User:Francis Tyers|Francis Tyers]], [[User:TommiPirinen|Flammie]], [[User:Unhammer|Unhammer]]
 
* '''[[/Unsupervised weighting of automata|read more...]]'''
 
</div>
 
   
  +
{{IdeaSummary
  +
| name = Automatic Error-Finder / Backpropagation
  +
| difficulty = Medium
  +
| skills = python?
  +
| description = Develop a tool to locate the approximate source of translation errors in the pipeline.
  +
| rationale = Being able to generate a list of probable error sources automatically makes it possible to prioritize issues by frequency, frees up developer time, and is a first step towards automated generation of better rules.
  +
| mentors = ???
  +
| more = /Backpropagation
  +
}}
   
  +
{{IdeaSummary
  +
| name = Add support for NMT to web API
  +
| difficulty = Medium
  +
| skills = python, NMT
  +
| description = Add support for a popular NMT engine to Apertium's web API
  +
| rationale = Currently Apertium's web API [[APy|Apertium APy]], supports only Apertium language modules. But the front end could just as easily interface with an API that supports trained NMT models. The point of the project is to add support for one popular NMT package (e.g., OpenNMT or JoeyNMT) to the APy.
  +
| mentors = [[User:Firespeaker|Jonathan Washington]]
  +
| more =
  +
}}
   
  +
{{IdeaSummary
  +
| name = Localization (l10n/i18n) of Apertium tools
  +
| difficulty = Medium
  +
| skills = C++
  +
| description = All our command line tools are currently hardcoded as English-only and it would be good if this were otherwise. [https://github.com/apertium/organisation/issues/28#issuecomment-803474833 Coding Challenge]
  +
| rationale = ...
  +
| mentors = [[User:Tino_Didriksen|Tino Didriksen]]
  +
| more = https://github.com/apertium/organisation/issues/28 Github
  +
}}
   
 
[[Category:Development]]
 
[[Category:Development]]

Latest revision as of 06:27, 27 May 2021

This is the ideas page for Google Summer of Code, here you can find ideas on interesting projects that would make Apertium more useful for people and improve or expand our functionality. If you have an idea please add it below, if you think you could mentor someone in a particular area, add your name to "Interested mentors" using ~~~

The page is intended as an overview of the kind of projects we have in mind. If one of them particularly piques your interest, please come and discuss with us on #apertium on irc.oftc.net, mail the mailing list, or draw attention to yourself in some other way.

Note that, if you have an idea that isn't mentioned here, we would be very interested to hear about it.

Here are some more things you could look at:


If you're a student trying to propose a topic, the recommended way is to request a wiki account and then go to

http://wiki.apertium.org/wiki/User:[[your username]]/GSoC2021Proposal

and click the "create" button near the top of the page. It's also nice to include [[Category:GSoC_2021_student_proposals]] to help organize submitted proposals.

Ideas[edit]

Python API for Apertium[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    C++, Python
  • Description:
    Update the Python API for Apertium to expose all Apertium modes and test with all major OSes
  • Rationale:
    The current Python API misses out on a lot of functionality, like phonemicisation, segmentation, and transliteration, and doesn't work for some OSes like Debian.
  • Mentors:
    Francis Tyers
  • read more...


OmniLingo and Apertium[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    JS, Python
  • Description:
    OmniLingo is a language learning system for practising listening comprehension using Apertium data. There is a lot of text processing involved (for example tokenisation) that could be aided by Apertium tools.
  • Rationale:
  • Mentors:
    Francis Tyers
  • read more...


Web API extensions[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    Python
  • Description:
    Update the web API for Apertium to expose all Apertium modes
  • Rationale:
    The current Web API misses out on a lot of functionality, like phonemicisation, segmentation, and transliteration
  • Mentors:
    Francis Tyers, Jonathan Washington, Xavi Ivars
  • read more...


Develop a morphological analyser[edit]

  • Difficulty:
    3. Entry level
  • Required skills:
    XML or HFST or lexd
  • Description:
    Write a morphological analyser and generator for a language that does not yet have one
  • Rationale:
    A key part of an Apertium machine translation system is a morphological analyser and generator. The objective of this task is to create an analyser for a language that does not yet have one.
  • Mentors:
    Francis Tyers, Jonathan Washington, Sevilay Bayatlı, Hossep, nlhowell
  • read more...


Support for Enhanced Dependencies in UD Annotatrix[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    NodeJS
  • Description:
    UD Annotatrix is an annotation interface for Universal Dependencies, but does not yet support all functionality
  • Rationale:
  • Mentors:
    Francis Tyers
  • read more...


User-friendly lexical selection training[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    Python, C++, shell scripting
  • Description:
    Make it so that training/inference of lexical selection rules is a more user-friendly process
  • Rationale:
    Our lexical selection module allows for inferring rules from corpora and word alignments, but the procedure is currently a bit messy, with various scripts involved that require lots of manual tweaking, and many third party tools to be installed. The goal of this task is to make the procedure as user-friendly as possible, so that ideally only a simple config file would be needed, and a driver script would take care of the rest.
  • Mentors:
    Unhammer, Mikel Forcada
  • read more...


Robust tokenisation in lttoolbox[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    C++, XML, Python
  • Description:
    Improve the longest-match left-to-right tokenisation strategy in lttoolbox to be fully Unicode compliant.
  • Rationale:
    One of the most frustrating things about working with Apertium on texts "in the wild" is the way that the tokenisation works. If a letter is not specified in the alphabet, it is dealt with as whitespace, so e.g. you get unknown words split in two so you can end up with stuff like ^G$ö^k$ı^rmak$ which is terrible for further processing.
  • Mentors:
    Francis Tyers, Flammie
  • read more...


apertium-separable language-pair integration[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    XML, a scripting language (Python, Perl), some knowledge of linguistics and/or at least one relevant natural language
  • Description:
    Choose a language you can identify as having a good number of "multiwords" in the lexicon. Modify all language pairs in Apertium to use the Apertium-separable module to process the multiwords, and clean up the dictionaries accordingly.
  • Rationale:
    Apertium-separable is a newly developed module to process lexical items with discontinguous dependencies, an area where Apertium has traditionally fallen short. Despite all the module has to offer, it has only been put to use in small test cases, and hasn't been integrated into any translation pair's development cycle.
  • Mentors:
    Jonathan Washington
  • read more...


UD and Apertium integration[edit]

  • Difficulty:
    3. Entry level
  • Required skills:
    python, javascript, HTML, (C++)
  • Description:
    Create a range of tools for making Apertium compatible with Universal Dependencies
  • Rationale:
    Universal dependencies is a fast growing project aimed at creating a unified annotation scheme for treebanks. This includes both part-of-speech and morphological features. Their annotated corpora could be extremely useful for Apertium for training models for translation. In addition, Apertium's rule-based morphological descriptions could be useful for software that relies on Universal dependencies.
  • Mentors:
    User:Francis Tyers Jonathan Washington
  • read more...


rule visualization tools[edit]


dictionary induction from wikis[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    MySQL, mediawiki syntax, perl, maybe C++ or Java; Java, Scala, RDF, and DBpedia to use DBpedia extraction
  • Description:
    Extract dictionaries from linguistic wikis
  • Rationale:
    Wiki dictionaries and encyclopedias (e.g. omegawiki, wiktionary, wikipedia, dbpedia) contain information (e.g. bilingual equivalences, morphological features, conjugations) that could be exploited to speed up the development of dictionaries for Apertium. This task aims at automatically building dictionaries by extracting different pieces of information from wiki structures such as interlingual links, infoboxes and/or from dbpedia RDF datasets.
  • Mentors:
    Jonathan Washington
  • read more...


unit testing framework[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    perl
  • Description:
    adapt https://github.com/TinoDidriksen/regtest for general Apertium use. Screenshots of regtest action
  • Rationale:
    We are gradually improving our quality control, with (semi-)automated tests, but these are done on the Wiki on an ad-hoc basis. Having a unified testing framework would allow us to be able to more easily track quality improvements over all language pairs, and more easily deal with regressions.
  • Mentors:
    Xavi Ivars
  • read more...


Bring an unreleased translation pair to releasable quality[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    shell scripting
  • Description:
    Take an unstable language pair and improve its quality, focusing on testvoc
  • Rationale:
    Many Apertium language pairs have large dictionaries and have otherwise seen much development, but are not of releasable quality. The point of this project would be bring one translation pair to releasable quality. This would entail obtaining good naïve coverage and a clean testvoc.
  • Mentors:
    Jonathan Washington, Sevilay Bayatlı, Hèctor Alòs i Font
  • read more...


Develop a prototype MT system for a strategic language pair[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    XML, some knowledge of linguistics and of one relevant natural language
  • Description:
    Create a translation pair based on two existing language modules, focusing on the dictionary and structural transfer
  • Rationale:
    Choose a strategic set of languages to develop an MT system for, such that you know the target language well and morphological transducers for each language are part of Apertium. Develop an Apertium MT system by focusing on writing a bilingual dictionary and structural transfer rules. Expanding the transducers and disambiguation, and writing lexical selection rules and multiword sequences may also be part of the work. The pair may be an existing prototype, but if it's a heavily developed but unreleased pair, consider applying for "Bring an unreleased translation pair to releasable quality" instead.
  • Mentors:
    Jonathan Washington, Sevilay Bayatlı, Hèctor Alòs i Font
  • read more...


Website Improvements: Misc[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, css, js, python
  • Description:
    Improve elements of Apertium's web infrastructure
  • Rationale:
    Apertium's website infrastructure Apertium-html-tools and its supporting API Apertium APy have numerous open issues. This project would entail choosing a subset of open issues and features that could realistically be completed in the summer. You're encouraged to speak with the Apertium community to see which features and issues are the most pressing.
  • Mentors:
    Jonathan Washington, Xavi Ivars
  • read more...


Website Improvements: Dictionary Lookup[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, css, js, python
  • Description:
    Finish implementing dictionary lookup mode in Apertium's web infrastructure
  • Rationale:
    Apertium's website infrastructure Apertium-html-tools and its supporting API Apertium APy have numerous open issues, including half-completed features like dictionary lookup. This project would entail completing the dictionary lookup feature. Some additional features which would be good to work would include automatic reverse lookups (so that a user has a better understanding of the results), grammatical information (such as the gender of nouns or the conjugation paradigms of verbs), and information about MWEs. See the open issue on GitHub.
  • Mentors:
    Jonathan Washington, Xavi Ivars
  • read more...


Website Improvements: Spell checking[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, js, css, python
  • Description:
    Add a spell-checking interface to Apertium's web tools
  • Rationale:
    Apertium-html-tools has seen some prototypes for spell-checking interfaces (all in stale PRs and branches on GitHub), but none have ended up being quite ready to integrate into the tools. This project would entail polishing up or recreating an interface, and making sure APy has a mode that allows access to Apertium voikospell modules. The end result should be a slick, easy-to-use interface for proofing text, with intuitive underlining of text deemed to be misspelled and intuitive presentation and selection of alternatives. the open issue on GitHub
  • Mentors:
    Jonathan Washington, Xavi Ivars
  • read more...


Website Improvements: Suggestions[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, css, js, python
  • Description:
    Finish implementing a suggestions interface for Apertium's web infrastructure
  • Rationale:
    Some work has been done to add a "suggestions" interface to Apertium's website infrastructure Apertium-html-tools and its supporting API Apertium APy, whereby users can suggest corrected translations. This project would entail finishing that feature. There are some related issues and PRs on GitHub.
  • Mentors:
    Jonathan Washington, Xavi Ivars
  • read more...


Website Improvements: Orthography conversion interface[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, js, css, python
  • Description:
    Add an orthography conversion interface to Apertium's web tools
  • Rationale:
    Several Apertium language modules (like Kazakh, Kyrgyz, Crimean Tatar, and Hñähñu) have orthography conversion modes in their mode definition files. This project would be to expose those modes through Apertium APy and provide a simple interface in Apertium-html-tools to use them.
  • Mentors:
    Jonathan Washington, Xavi Ivars
  • read more...


Apertium Browser Plugin[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    html, css, js, python
  • Description:
    Expand functionality of Geriaoueg vocabulary assistant
  • Rationale:
    Geriaoueg is a vocabulary assistant with Firefox/Chrom[e/ium] plugins. These plugins interface with Apertium's web API, Apertium APy, and allow a user to look up (in Apertium's dictionaries) word forms from a web page they're viewing. A Firefox/Chrom[e/ium] plugin should also be able to provide in-browser website translation. This project is to clean up the dictionary lookup functionality and add translation support to the plugins. Some APy features may need to be tweaked, but most of the work in this project will be solely in the plugins.
  • Mentors:
    Jonathan Washington, Xavi Ivars, Tino Didriksen
  • read more...


Extend Weighted transfer rules[edit]


Automatic Error-Finder / Backpropagation[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    python?
  • Description:
    Develop a tool to locate the approximate source of translation errors in the pipeline.
  • Rationale:
    Being able to generate a list of probable error sources automatically makes it possible to prioritize issues by frequency, frees up developer time, and is a first step towards automated generation of better rules.
  • Mentors:
    ???
  • read more...


Add support for NMT to web API[edit]

  • Difficulty:
    2. Medium
  • Required skills:
    python, NMT
  • Description:
    Add support for a popular NMT engine to Apertium's web API
  • Rationale:
    Currently Apertium's web API Apertium APy, supports only Apertium language modules. But the front end could just as easily interface with an API that supports trained NMT models. The point of the project is to add support for one popular NMT package (e.g., OpenNMT or JoeyNMT) to the APy.
  • Mentors:
    Jonathan Washington
  • [[|read more...]]


Localization (l10n/i18n) of Apertium tools[edit]