Concretely this means algorithms for
harvest/: Code for making the dataset. It works by crawling and filtering Wikipedia for bias-driven
src/: Code for training models and using trained models to run inference. The models
implemented here are referred to as
CONCURRENT in the paper.
These commands will download data and a pretrained model before running inference.
$ cd src/ $ python3 -m venv venv $ source venv/bin/activate $ pip install -r requirements.txt $ python >> import nltk; nltk.download("punkt") $ sh download_data_ckpt_and_run_inference.sh
You can also run
sh integration_test.sh to further verify that everything is installed correctly
and working as it should be.