2016-02-04 18:23:40 +08:00
# WikiExtractor
2015-03-22 20:41:39 +08:00
[WikiExtractor.py ](http://medialab.di.unipi.it/wiki/Wikipedia_Extractor ) is a Python script that extracts and cleans text from a [Wikipedia database dump ](http://download.wikimedia.org/ ).
2016-06-18 11:25:51 +08:00
The tool is written in Python and requires Python 2.7 or Python 3.3+ but no additional library.
2015-03-22 20:41:39 +08:00
2015-03-22 20:59:58 +08:00
For further information, see the [project Home Page ](http://medialab.di.unipi.it/wiki/Wikipedia_Extractor ) or the [Wiki ](https://github.com/attardi/wikiextractor/wiki ).
2015-03-22 20:58:50 +08:00
2016-02-04 18:23:40 +08:00
# Wikipedia Cirrus Extractor
2016-02-11 16:55:55 +08:00
`cirrus-extractor.py` is a version of the script that performs extraction from a Wikipedia Cirrus dump.
Cirrus dumps contain text with already expanded templates.
2016-02-04 18:23:40 +08:00
2016-02-11 16:55:55 +08:00
Cirrus dumps are available at:
[cirrussearch ](http://dumps.wikimedia.org/other/cirrussearch/ ).
2016-02-04 18:23:40 +08:00
# Details
WikiExtractor performs template expansion by preprocesssng the whole dump and extracting template definitions.
2015-10-26 00:03:17 +08:00
2016-02-13 01:16:54 +08:00
In order to speed up processing:
2015-10-26 00:03:17 +08:00
2016-02-13 01:16:54 +08:00
- multiprocessing is used for dealing with articles in parallel
- a cache is kept of parsed templates (only useful for repeated extractions).
## Installation
The script may be invoked directly, however it can be installed by doing:
(sudo) python setup.py install
2015-03-22 20:41:39 +08:00
## Usage
The script is invoked with a Wikipedia dump file as an argument.
2015-10-26 00:03:17 +08:00
The output is stored in several files of similar size in a given directory.
2015-03-22 20:41:39 +08:00
Each file will contains several documents in this [document format ](http://medialab.di.unipi.it/wiki/Document_Format ).
2016-06-19 19:10:36 +08:00
usage: WikiExtractor.py [-h] [-o OUTPUT] [-b n[KMG]] [-c] [--html] [-l] [-s]
[--lists] [-ns ns1,ns2] [-xns ns1,ns2]
[--templates TEMPLATES] [--no-templates] [--escapedoc]
[-r] [--min_text_length MIN_TEXT_LENGTH]
[--filter_disambig_pages] [--processes PROCESSES] [-q]
[--debug] [-a] [-v]
input
Wikipedia Extractor:
Extracts and cleans text from a Wikipedia database dump and stores output in a
number of files of similar size in a given directory.
Each file will contain several documents in the format:
< doc id = "" revid = "" url = "" title = "" >
...
< / doc >
Template expansion requires preprocesssng first the whole dump and
collecting template definitions.
2015-03-22 20:41:39 +08:00
2015-04-20 12:56:29 +08:00
positional arguments:
2015-10-26 00:03:17 +08:00
input XML wiki dump file
2015-04-15 20:30:55 +08:00
2015-03-22 20:41:39 +08:00
optional arguments:
-h, --help show this help message and exit
2015-11-20 07:06:23 +08:00
--processes PROCESSES number of processes to use (default: number of CPU cores)
2015-04-20 12:56:29 +08:00
Output:
2015-03-22 20:41:39 +08:00
-o OUTPUT, --output OUTPUT
2016-06-19 19:10:36 +08:00
directory for extracted files (or '-' for dumping to
2015-11-20 07:06:23 +08:00
stdout)
2015-04-15 20:30:55 +08:00
-b n[KMG], --bytes n[KMG]
2015-11-20 07:06:23 +08:00
maximum bytes per output file (default 1M)
2015-03-22 20:41:39 +08:00
-c, --compress compress output files using bzip
2015-04-20 12:56:29 +08:00
Processing:
2015-11-20 07:34:23 +08:00
--html produce HTML output, subsumes --links
2015-03-22 20:41:39 +08:00
-l, --links preserve links
2016-06-19 19:10:36 +08:00
-s, --sections preserve sections
2016-02-13 06:31:21 +08:00
--lists preserve lists
2015-03-22 20:41:39 +08:00
-ns ns1,ns2, --namespaces ns1,ns2
2016-06-19 19:10:36 +08:00
accepted link namespaces
-xns ns1,ns2, --xml_namespaces ns1,ns2
accepted page xml namespaces -- 0 for main/articles
2015-03-22 20:41:39 +08:00
--templates TEMPLATES
2016-06-19 19:10:36 +08:00
use or create file containing templates
2015-04-12 17:05:52 +08:00
--no-templates Do not expand templates
2015-11-14 00:32:32 +08:00
--escapedoc use to escape the contents of the output
< doc > ...< / doc >
2016-06-19 19:10:36 +08:00
-r, --revision Include the document revision id (default=False)
--min_text_length MIN_TEXT_LENGTH
Minimum expanded text length required to write
document (default=0)
--filter_disambig_pages
Remove pages from output that contain disabmiguation
markup (default=False)
2015-04-20 12:56:29 +08:00
Special:
-q, --quiet suppress reporting progress info
--debug print debug info
2016-06-19 19:10:36 +08:00
-a, --article analyze a file containing a single article (debug
option)
2015-03-22 20:41:39 +08:00
-v, --version print program version
2016-06-19 19:10:36 +08:00
2015-11-20 07:06:23 +08:00
Saving templates to a file will speed up performing extraction the next time,
assuming template definitions have not changed.
2015-03-22 20:41:39 +08:00
2015-11-20 07:06:23 +08:00
Option --no-templates significantly speeds up the extractor, avoiding the cost
of expanding [MediaWiki templates ](https://www.mediawiki.org/wiki/Help:Templates ).
2015-04-12 17:18:19 +08:00
2015-04-26 14:57:25 +08:00
For further information, visit [the documentation ](http://attardi.github.io/wikiextractor ).