Difference between revisions of "Wikipedia dumps"
Jump to navigation
Jump to search
Hectoralos (talk | contribs) (+ wikiextractor) |
|||
Line 12: | Line 12: | ||
* [[Calculating_coverage#More_involved_scripts]] – an ugly shell script that does the job |
* [[Calculating_coverage#More_involved_scripts]] – an ugly shell script that does the job |
||
* [http://wp2txt.rubyforge.org/ wp2txt] – some ruby thing (does this work?) |
* [http://wp2txt.rubyforge.org/ wp2txt] – some ruby thing (does this work?) |
||
There are also dumps of the articles translated with the Content Translation tool (which uses apertium (and other MT engines) under the hood): |
|||
* https://www.mediawiki.org/wiki/Content_translation/Published_translations |
|||
[[Category:Resources]] |
[[Category:Resources]] |
Revision as of 17:42, 6 August 2018
Wikipedia dumps are useful for quickly getting a corpus. They are also the best corpora for making your language pair are useful for Wikipedia's Content Translation tool :-)
You download them from
There are several tools for turning dumps into useful plaintext, e.g.
- Wikipedia Extractor – a python script that tries to remove all formatting
- wikiextractor – another python script that removes all formatting (with different options), putting XML marks just to know when begins and ends everty single article
- mwdump-to-pandoc – shell wrapper around pandoc (see the usage.sh below the script for how to use)
- Calculating_coverage#More_involved_scripts – an ugly shell script that does the job
- wp2txt – some ruby thing (does this work?)
There are also dumps of the articles translated with the Content Translation tool (which uses apertium (and other MT engines) under the hood):