Simply customize the SPARQL query to your needs
using a dump and Couchdb
This was the bot's first implementation: more rigid and requiring more resources but I let it here as an inspiration:
- install wikidata-filter
- download Wikidata latest dump and filter on the desired claim: here P106:Q36180 keeps only entities with occupation (P106) writter (Q36180):
curl https://dumps.wikimedia.org/wikidatawiki/entities/latest-all.json.gz |gzip -d | wikidata-filter --claim P106:Q36180 > isWriter.json
load all the remaining entities lines in a new CouchDB database. See the couch-wikidata-dump-importer for how I did that. It's a bit hacky but it works.
create the views you need using CouchDB's map functions. Here, the important view is
byBirthday, and voila:
a view ready to be queried by day and month by our grumpy Victor Hugo!
- on twitter:
- on facebook:
npm run authors-tomorrow fetches the data of the authors of the next day and help you match them with twitter users in order to improve wikidata data before the bot comes to use it.
To edit wikidata using the
validate button, a rightly configured instance of wikidata-agent is expected to be running on port 4115 of your machine.