This is the data used by the Enhanced Bangumi API, yet another API to collect anime-related information such as on which websites people can stream animes.
schemas/schemas.sql
is the formal definition of the relational database schemas that the API uses. The format of CSV files described below are also in compliance with the tables defined here.schemas/*.json
describes the schemas in JSON Schema format and are solely for validation purpose.
services.csv
contains the records of anime streaming websites.locales.csv
contains the subtitle locale information.subjects/<year>.csv
contains the records of anime series.sources/<quarter>.csv
contain the records of which anime websites are streaming which animes.episodes/<quarter>/<subject>.csv
contain the records for each episode.
catelog/<year>.md
records for each anime series, the data of which streaming websites have been collected.
Run sh import.sh
to bulkily load schemas and data into a Postgres database.
The data crawler is a tool that crawls anime data from various streaming websites and generates the source and episode records automatically. Records in the other CSV files need to be collected manually.
To validate the data files, run npm install -g csval
to install csval
, and then run sh validate.sh
.
This sub-project is inspired by the following projects:
© 101对双生儿 2020. All Rights Reserved.