Find scraped Wikipedia content in output/
directory.
- install Node.js (find correct node version in .nvmrc file)
- run
yarn install
This will execute the ETL pipeline:
- download Wikipedia pages with politician/cabinet info tables (Bund und Länder)
- get relevant infos from html tables (names, parties, ministerial position, etc.)
- write infos to json and markdown files
yarn install
yarn start
yarn install
yarn test