shawntan / SUT

Repository for Sparse Universal Transformers

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Sparse Universal Transformers (SUT)

Implementation for WMT'14 En-De for SUT.

Link to paper


Steps to run:

  1. Clone and follow the preprocessing steps in Transformer-Clinic

    export TRANSFORMER_CLINIC_ROOT=/path/to/Transformer-Clinic
  2. Initialise evironment variables for experiment (base or big):

    source scripts/params_de_base.sh
    export EXP_NAME=sut_base
    # Submit job to cluster (32 compute nodes)
    sbatch slurm/launch.slurm
  3. Generate and calculate BLEU:

    scripts/eval_wmt14ende.sh checkpoints/sut_base

About

Repository for Sparse Universal Transformers


Languages

Language:Python 93.7%Language:Shell 6.3%