Build a hyperdrive feed of Wikipedia articles, including historical revisions.
Wikipedia is an incredibly important collection of knowledge on the internet. It is free to read and edit for everyone. Though, it is important to know that it is stored on only a handful of servers in a handful of countries controlled by a single organisation. This mainly causes two problems:
- Currently, it is too easy to censor the Wikipedia. We need a system that supports redundancy without any additional effort.
- It does not work offline. Making an offline copy is complicated. Also, you usually have to download all articles for a language.
Let's store Wikipedia's content in a peer-to-peer (P2P) system. By leveraging software from the Dat project, we won't have to reinvent the wheel. The Dat protocol efficiently only syncs changes between to versions of data, allows for sparse & live replication and is completely distributed.
This tool can extract articles from a Wikipedia dump or download it directly, and store it in a Dat archive. See below for more details.
npm install -g build-wikipedia-feed
This module exposes several command line building blocks.
Pipe a stub-meta-history*
XML file into wiki-revisions-list
. You will get an ndjson list of page revisions.
curl -s 'https://dumps.wikimedia.org/enwiki/20181001/enwiki-20181001-stub-meta-history1.xml.gz' | gunzip | wiki-revisions-list >revisions.ndjson
Pipe a stub-meta-current*
XML file file into wiki-revisions-list
. You will get an ndjson list of page revisions.
curl -s 'https://dumps.wikimedia.org/enwiki/20181001/enwiki-20181001-stub-meta-current1.xml.gz' | gunzip | wiki-revisions-list >revisions.ndjson
Use wiki-live-revisions
. You will get an ndjson list of page revisions.
wiki-live-revisions >revisions.ndjson
fetch & store revisions in a hyperdrive
Use wiki-store-revisions
to write the HTML content of all revisions in revisions.ndjson
into a hyperdrive. The archive will be created under p2p-wiki
in your system's data directory.
cat revisions.ndjson | wiki-store-revisions
- distributed-wikipedia-mirror – Putting wikipedia on IPFS
fetch-wikipedia-page-revision
– Fetch a revision of a Wikipedia page as mobile HTML.wikipedia-edits-stream
– A live stream of page edits on Wikipedia.commons-photo-url
– Download Wikimedia Commons photos.wiki-article-name-encoding
– Encode & decode Wiki(pedia) article names/slugs.
If you have a question or have difficulties using build-wikipedia-feed
, please double-check your code and setup first. If you think you have found a bug or want to propose a feature, refer to the issues page.