Scrapers for German school data. As Germany is a federal country and education is a state level matter every state publishes their data differently. The goal of this project is to scraper all data and to make it available under a common data format.
A version of these scrapers is deployed at https://jedeschule.codefor.de/. You can use it one of three ways:
- Using the API. Documentation is available at https://jedeschule.codefor.de/docs
- Using the CSV dump provided at https://jedeschule.codefor.de/csv-data/
- Exploring the data using a small interactive dashboard at: https://knutator2.github.io/jedeschule-explore/#/dashboard
Note that at least Python 3.7 is required.
pip install -r requirements.txt
Make your you have a postgres database with postgis support running if you want to use the database pipeline and expor t an environment variable pointing to it like so:
export DATABASE_URL=postgres://[email protected]:5432/jedeschule
Then, run the migrations to bring you database up to date
alembic upgrade head
To run some very basic integration tests, you will need another postgres
database for test data. After creating it, run the following steps
(assuming you called your database jedeschule_scraper_test
):
export DATABASE_URL=postgres://[email protected]:5432/jedeschule_scraper_test
alembic upgrade head
python test_models.py
If you made changes to scrapers, you can also run a script to check the changes between your new scraper results and the data that is currently hosted on the server at jedeschule.codefor.de.
To do so, run the ./test_changes.sh
script on your feature branch. This
will find the scrapers that have been changed on your branch and will then run
them to get some sample data. For each fetched result, it will call the upstream API
and compare the results for this ID. Note that the script always returns
with a 0 exit code, hence you need to evaluate the changes manually.
To update data for all states, run
bash ./scrape_all.sh
To get updates for a single state, run
scrapy crawl <state_name>
# use `scrapy list` to get a list of all available states
You can deploy the scrapers directly on a server or using Docker. The resulting Docker
image needs to be run with the DATABASE_URL
environment variable pointing to a Postgres
instance. You can for example use docker-compose to achieve this. See the
docker-compose file at jedeschule-api
for inspiration.
There is a version of these scrapers (and an instance of jedeschule-api) that is deployed at
jedeschule.codefor.de.
The deployment uses docker-compose. To update the data there, login as the
jedeschule
user and run the commands mentioned int the previous section
prefixed with sudo docker-compose run scrapers
. This means that if you
want to for example get the newest data for Berlin you would run sudo docker-compose run scrapers scrapy crawl berlin