Skip to content


Repository files navigation

Bangumi Spider

This is a scrapy based spider used to scrape data from Bangumi 番组计划.

What information it can scrape?

Spider name Purpose Items
user-api User information retrived via API. Aligned with User in
subject-api Subject information retrived via API. Aligned with Subject in
collections-api User records retrived via API. Aligned with UserSubjectCollection in
user User information retrived via web. uid, name, nickname, joindate
record* User-subject information retrived via web. uid(user id), iid(item id), typ(item type), state(user's favorite state), adddate, rate, tags, comment
subject Subject information retrived via web. subjectid, subjecttype, subjectname, order(same as subjectid except for redirected subjects, for which original id is kept as order), alias(alias of subjectname), staff, relations(subject's relation with other subjects)

* record spider: it would scrape user simultaneously.

How to use the spider?

Prerequisit of this spider is scrapy, so one need to install scrapy first.

You also need to obtain your Bangumi user token via

To use web spiders locally, one should git clone this repository first then run

scrapy crawl subject-api -a id_max=100 -a token={TOKEN}

TOKEN is the Bangumi token you retrieved via

id_max is a parameter specifying the maximun subject id it should be scraped. Meanwhile there is another parameter naming id_min.

For spiders scraping from web page, one need to add environmental variable SCRAPY_PROJECT=bgm before running. For example, on Linux it should be

SCRAPY_PROJECT=bgm scrapy crawl subject -a id_max=100

Then you can check the scraped items under main folder.

How to deploy?

To setup a scrapyd server, one can certainly do that by following Scrapyd document. However, we are providing a docker image to help you achieve that goal more quickly. The docker image is has a nginx served as authentication server. To start the docker image, one need to specify the USERNAME and PASSWORD as environment variable.

docker run -d -p 6810:6810 -e USERNAME=username -e PASSWORD=password wattlebird/scrapyd:latest
curl --user username:password http://localhost:6810/schedule.json -d project=bgm -d spider=record -d id_max=100

Then you can visit http://localhost:6810 to watch your jobs.

The source code of that docker image is under the folder scrapyd.

To deploy bgmapi and bgm properly, one have to execute the following commands:

python bdist_egg
curl --user {USERNAME}:{PASSWORD} http://localhost:6810/addversion.json -F project=bgmapi -F version=1.0 -F egg=@dist/project-1.0-py3.10.egg # The actual egg file generated may have a differ
python bdist_egg
curl --user {USERNAME}:{PASSWORD} http://localhost:6810/addversion.json -F project=bgm -F version=1.0 -F egg=@dist/project-1.0-py3.10.egg # The actual egg file generated may have a differ

Known issues

Due to sensitive content restriction, spider subject cannot scrape subjects that are marked as R-18.


MIT Licensed.