Yao Lu's repositories
Multi-XScience
Multi-XScience: A Large-scale Dataset for Extreme Multi-document Summarization of Scientific Articles
random-prompt
Code and supplementary document
clifton
SSH connection manager
codespaces-jupyter
Explore machine learning and data science with Codespaces
CTranslate2
Fast inference engine for Transformer models
cvmfs-tutorial-hpc-best-practices
Contents for "Best Practices for CernVM-FS in HPC" tutorial
ELI5
Scripts and links to recreate the ELI5 dataset.
few-shot-learning
Few-shot Learning of GPT-3
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).
hiersumm
Code for paper Hierarchical Transformers for Multi-Document Summarization in ACL2019
lm-evaluation-harness
A framework for few-shot evaluation of language models.
Megatron-LM
Ongoing research training transformer models at scale
ml-engineering
Machine Learning Engineering Open Book
Multi-News
Large-scale multi-document summarization dataset and code
nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
NLP-progress
Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks.
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
Sentence-VAE
PyTorch Implementation of "Generating Sentences from a Continuous Space" by Bowman et al 2015 https://arxiv.org/abs/1511.06349
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
writing-code-for-nlp-research-emnlp2018
A companion repository for the "Writing code for NLP Research" Tutorial at EMNLP 2018
yaolu.github.io
Personal Website