Eric-Wallace / dadc-limit

DADC in the Limit

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Analyzing Dynamic Adversarial Training Data in the Limit

This repository contains the official data and code for the ACL 2022 findings paper "Analyzing Dynamic Adversarial Training Data in the Limit".

Data

The data folders contains our final training datasets, broken down by the collection type: non-adversarial data collection, static adversarial data collection, and dynamic adversarial collection. The datasets differ slightly from the ones used in the paper, namely they contain ~10% more data, consisting of examples that were not ready by the paper. We also release our manually-crafted test set.

Analysis

See the analysis folder to recreate the analysis experiments from Section 5 of the paper. In particular:

  • diversity.py will recreate the diversity rows in Table 4.
  • complexity.py will recreate the complexity rows in Table 4.
  • artifacts.py will recreate the artifacts rows in Table 4.

Training Models

The training folder contains the static adversarial model and the code to reproduce the other training experiments using RoBERTa large models. The train-all-nli-100000 checkpoint is the static adversarial model used in the experiments. It is a binary NLI model trained for 100,000 steps on the union of SNLI, MNLI, and FEVER-NLI. You will need to download that model from the following link as the checkpoint is too big to fit onto github.

To recreate the training experiments using various amounts of round data, use the train.sh script. You will need to run the following installation inside the training folder.

conda create -n dadc python=3.7
source activate dadc
pip install -r requirements.txt

Now you should be ready to go!

The train-all-nli-100000 checkpoint is the static adversarial model used in the experiments. It is a binary NLI model trained for 100,000 steps on the union of SNLI, MNLI, and FEVER-NLI.

Citation

Please consider citing our work if you found this code or our paper beneficial to your research.

@inproceedings{Wallace2022Dynamic,  
    Title = {Analyzing Dynamic Adversarial Training Data in the Limit},
    Author = {Eric Wallace and Adina Williams and Robin Jia and Douwe Kiela}, 
    Booktitle={Findings of the Association for Computational Linguistics},
    Year = {2022}
}

Contributions and Contact

This code was developed by Eric Wallace, contact available at ericwallace@berkeley.edu.

If you'd like to contribute code, feel free to open a pull request. If you find an issue with the code, please open an issue.

About

DADC in the Limit

License:Other


Languages

Language:Python 100.0%