shichao-wang / multi_relational_script_learning

This repository contains code, data, and pre-trained models for the paper "Multi-Relation Script Learning for Discourse Relations"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multi-Relation Script Learning for Discourse Relations

This repository contains code, data, and pre-trained models for the following papers.

I-Ta Lee, and Dan Goldwasser, "Multi-Relational Script Learning for Discourse Relations", ACL 2019

bibtex

@inproceedings{lee2019multi,
  title={Multi-Relational Script Learning for Discourse Relations},
  author={Lee, I-Ta and Goldwasser, Dan},
  booktitle={Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics},
  year={2019}
}

If you use any resources within this repository, please cite the paper.

Dependencies

Our project was originally using Python2.7. However, when ELMo embeddings, which requires Python3, come into play, we migrate the code runnable with Python3. Please install dependencies for both.

pip install -r requirements.txt
pip3 install -r requirements3.txt

Data

Our core experiments (MCNC, MCNS, MCNE) use data splits of NYT section of English Gigawords, following the data splits given by Granroth-Wilding (https://mark.granroth-wilding.co.uk/papers/what_happens_next/) [1].

As our models need entity mention spans rather than entity head words only, which are not provided in Granroth-Wilding's code outputs, considering the stochastic factors in the pre-processing, we release a newly pre-processsed train/dev/test data for MCNC, MCNS, MCNE (check the experiments in the paper). Unlike Granroth-Wilding's code, we use Stanford CoreNLP as the text pipeline tool and follow the heuristic mentioned in the paper for retrieving all the events.

Download

The datasets are in token-indexed triplet format and the index files are provided.

data/skipthought_models/dictionary.txt
data/skipthought_models/utable.npy
data/skipthought_models/btable.npy
data/skipthought_models/uni_skip.npz
data/skipthought_models/uni_skip.npz.pkl
data/skipthought_models/bi_skip.npz
data/skipthought_models/bi_skip.npz.pkl

Pre-trained models

You can download the models from the link in the Download section

  • EventTransE:
    • model: pretrained/out_transe_v0.2.10_long9_tmp
    • config: train_config_transe_v0.2.10_long9.json
  • EventTransR: pretrained/out_transr_v0.2.10_long9_tmp
    • model: pretrained/out_transr_v0.2.10_long9_tmp
    • config: train_config_transr_v0.2.10_long9.json

Run experiments with pre-trained models

To begin with, download the test data and pretrain models in the Download section, and decompress them in the repo folder. You need two configuation files: train_config_{xxx}.json and relation_{xxx}.json. Check all the file paths in the configuration file are correct.

Note that for all the executable .py, you can use -h argument to see the help and -g <gpu_id> to specify running on a specific gpu.

MCNC

For EventTransE

python bin/evaluations/eval_mcnc.py -v pretrained/out_transe_v02.10_long9_tmp/model_2_3_2591.pt pretrained/out_transe_v0.2.10_long9_tmp/argw_enc_2_3_2591.pt data/mcnc_test_v0.3.4/mcnc_coref_next.pkl train_config_transe_v0.2.10_long9.json relation_9disc.json

For EventTransR, you simply need to replace the model file, argument encoder file, and config file.

MCNS and MCNE

Two experiments run in one command.

For EventTransE:

python bin/evaluations/eval_mcns.py -v pretrained/out_transe_v0.2.10_long9_tmp/model_2_3_2591.pt pretrained/out_transe_v0.2.0_long9_tmp/argw_enc_2_3_2591.pt glove.6B.300d.txt data/mcnc_test_v0.3.4/mcns_coref_next.pkl train_config_transe_v0.2.10_long9.json relation_9disc.json Viterbi

For EventTransR, you simply need to replace the model file, argument encoder file, and config file.

Intrinsic Discourse Relations

The evaluations here take long time to run. GPUs recommended.

Predict an event or an relation class

There are two set-up here:

  • Predict the next event given one event and relation
  • Predict the relation given two events

The following command will output results for both setups.

For EventTransE

python bin/evaluations/eval_disc.py pretrained/out_transe_v0.2.10_long9_tmp/model_2_3_2591.pt pretrained/out_transe_v0.2.10_ong9_tmp/argw_enc_2_3_2591.pt data/disc_test_v0.2.0.pkl train_config_transe_v0.2.10_long9.json relation_9disc.json -v

For EventTransR, you simply need to replace the model file, argument encoder file, and config file.

You can do this using pre-trained ELMo embeddings like what is described in the paper. To do so, download ELMo models from the Download section. Python3 is required by ELMo. This following commands also output results for the two setups.

python3 bin/evaluations/eval_disc_elmo.py -v data/disc_test_v0.2.0.pkl relation_9disc.json

Triplet classication

Another setup is to do a binary classification for a given triplet. Download ELMo models from the Download section. Python3 is required by ELMo.

python3 bin/evaluations/eval_disc_binary.py -v pretrained/out_transe_v0.2.10_long9_tmp/model_2_3_2591.pt pretrained/out_trane_v0.2.10_long9_tmp/argw_enc_2_3_2591.pt data/disc_dev_v0.2.0.pkl data/disc_test_v0.2.0.pkl train_config_transe_v0.2.10_long9.json relation_9disc.json

This command runs for ELMo+EventTransE. To run with ELMo-only, add -m to the commmand. The results are shown in the log file or stdout. The reported results are averaged over 5 runs

Implicit Discourse Sense Classifications

Download ELMo models from the Download section. Python3 is required by ELMo. Download the pre-trained classifier (in pretraind.tar.gz in the Download section) that takes in EventTransE+ELMo as input representations. The result reported in the paper is averaged over 5 runs.

python3 bin/evaluations/test_combined_features.py -v data/pdtb_ds/ds_dev_events.json data/pdtb_ds/ds_test_events.json data/ptb_ds/ds_blind_test_events.json pretrained/out_ds_transe_nonexplicit_t5/model_0_0_33.pt pretrained/out_ds_transe_nonexplicit_t5/argw_enc_0_0_33.pt train_config_ds_transe.json relation_pdtb.json pretrained/out_ds_comb_elmo_transe_e200_4/best_model.pt output_folder

This will output predictions in json format, which is supported by the CONLL 2016's official scorer.

Train from scratch

Download the training data from the Download section. Make sure the training file path in the config file is correct. Then run

python bin/train.py -v -r train_config_transe_v0.2.10_long9.json output_model

Again, for EventTransR, simply replace the config file.

References

[1] Granroth-Wilding, Mark, and Stephen Clark. "What happens next? event prediction using a compositional neural network model." Thirtieth AAAI Conference on Artificial Intelligence. 2016. [2] Kiros, Ryan, et al. "Skip-thought vectors." Advances in neural information processing systems. 2015.

About

This repository contains code, data, and pre-trained models for the paper "Multi-Relation Script Learning for Discourse Relations"

License:MIT License


Languages

Language:Python 100.0%