mmarius's starred repositories
mesh-transformer-jax
Model parallel transformers in JAX and Haiku
FriendsDontLetFriends
Friends don't let friends make certain types of data visualization - What are they and why are they bad.
text-to-text-transfer-transformer
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
x-transformers
A simple but complete full-attention transformer with a set of promising experimental features from various papers
OpenPrompt
An Open-Source Framework for Prompt-Learning.
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
few-shot-learning
Few-shot Learning of GPT-3
Channel-LM-Prompting
An original implementation of "Noisy Channel Language Model Prompting for Few-Shot Text Classification"
prompt_semantics
This repository accompanies our paper “Do Prompt-Based Models Really Understand the Meaning of Their Prompts?”
composable-sft
A library for parameter-efficient and composable transfer learning for NLP with sparse fine-tunings.
tilt-transfer
Code to run the TILT transfer learning experiments
bayesian-mi
This code accompanies the paper "Bayesian Framework for Information-Theoretic Probing" published in EMNLP 2021.