Project by Kaj Bostrom, Jifan Chen, and Greg Durrett. Code by Kaj Bostrom and Jifan Chen.
Modified by Pengfei Yan.
You'll need Python >= 3.6 to run the code in this repo.
First, clone the repository:
git clone [email protected]:gregdurrett/fp-dataset-artifacts.git
Then install the dependencies:
pip install --upgrade pip
pip install -r requirements.txt
If you're running on a shared machine and don't have the privileges to install Python packages globally, or if you just don't want to install these packages permanently, take a look at the "Virtual environments" section further down in the README.
To make sure pip is installing packages for the right Python version, run pip --version
and check that the path it reports is for the right Python interpreter.
To train an ELECTRA-small model on the SNLI natural language inference dataset, you can run the following command:
python3 run.py --do_train --task nli --dataset snli --output_dir ./trained_model/
Checkpoints will be written to sub-folders of the trained_model
output directory.
To evaluate the final trained model on the SNLI dev set, you can use
python3 run.py --do_eval --task nli --dataset snli --model ./trained_model/ --output_dir ./eval_output/
To prevent run.py
from trying to use a GPU for training, pass the argument --no_cuda
.
To train/evaluate a question answering model on SQuAD instead, change --task nli
and --dataset snli
to --task qa
and --dataset squad
.
Descriptions of other important arguments are available in the comments in run.py
.
Data and models will be automatically downloaded and cached in ~/.cache/huggingface/
.
To change the caching directory, you can modify the shell environment variable HF_HOME
or TRANSFORMERS_CACHE
.
For more details, see this doc.
An ELECTRA-small based NLI model trained on SNLI for 3 epochs (e.g. with the command above) should achieve an accuracy of around 89%, depending on batch size. An ELECTRA-small based QA model trained on SQuAD for 3 epochs should achieve around 78 exact match score and 86 F1 score.
This repo uses Huggingface Datasets to load data.
The Dataset objects loaded by this module can be filtered and updated easily using the Dataset.filter
and Dataset.map
methods.
For more information on working with datasets loaded as HF Dataset objects, see this page.
Python 3 supports virtual environments with the venv
module. These will let you select a particular Python interpreter
to be the default (so that you can run it with python
) and install libraries only for a particular project.
To set up a virtual environment, use the following command:
python3 -m venv path/to/my_venv_dir
This will set up a virtual environment in the target directory. WARNING: This command overwrites the target directory, so choose a path that doesn't exist yet!
To activate your virtual environment (so that python
redirects to the right version, and your virtual environment packages are active),
use this command:
source my_venv_dir/bin/activate
This command looks slightly different if you're not using bash
on Linux. The venv docs have a list of alternate commands for different systems.
Once you've activated your virtual environment, you can use pip
to install packages the way you normally would, but the installed
packages will stay in the virtual environment instead of your global Python installation. Only the virtual environment's Python
executable will be able to see these packages.
1.1
- train
python3 run.py --do_train --task qa --dataset ./squad-retrain-data/train-v1.1.json --output_dir ./trained_model_v_1/
- evaluate
# squad
python3 run.py --do_eval --task qa --dataset squad --model ./trained_model_v_1/ --output_dir ./eval_output_v_1/
# squad_adversarial:AddSent
python3 run.py --do_eval --task qa --dataset squad_adversarial:AddSent --model ./trained_model_v_1/ --output_dir ./eval_output_v_1_adversarial_AddSent/
# squad_adversarial:AddOneSent
python3 run.py --do_eval --task qa --dataset squad_adversarial:AddOneSent --model ./trained_model_v_1/ --output_dir ./eval_output_v_1_adversarial_AddOneSent/
2.0
- train
# using adversarial retrain
python3 run.py --do_train --task qa --dataset ./squad-retrain-data/train-v2.0.json --output_dir ./trained_model_v_2
- evaluate
# squad
python3 run.py --do_eval --task qa --dataset squad --model ./trained_model_v_2/ --output_dir ./eval_output_v_2/
# squad_adversarial:AddSent
python3 run.py --do_eval --task qa --dataset squad_adversarial:AddSent --model ./trained_model_v_2/ --output_dir ./eval_output_v_2_adversarial_AddSent/
# squad_adversarial:AddOneSent
python3 run.py --do_eval --task qa --dataset squad_adversarial:AddOneSent --model ./trained_model_v_2/ --output_dir ./eval_output_v_2_adversarial_AddOneSent/
1.1
- start position
python3 -m selection.train_dy_filtering --plot --model_dir ./trained_model_v_1/start_pos/ --model ELECTRA-small-start-pos --burn_out 3
- end position
python3 -m selection.train_dy_filtering --plot --model_dir ./trained_model_v_1/end_pos/ --model ELECTRA-small-end-pos --burn_out 3
2.0
- start position
# plot start pos
python3 -m selection.train_dy_filtering --plots_dir cartography_v_2 --plot --model_dir ./trained_model_v_2/start_pos --model ELECTRA-small-start-pos --burn_out 3
- end position
# plot end pos
python3 -m selection.train_dy_filtering --plots_dir cartography_v_2 --plot --model_dir ./trained_model_v_2/end_pos --model ELECTRA-small-end-pos --burn_out 3