Tianjian Li (tianjianl)

tianjianl

Geek Repo

Company:Johns Hopkins University

Location:Baltimore, MD

Home Page:tianjianl.github.io

Twitter:@tli104

Github PK Tool:Github PK Tool

Tianjian Li's starred repositories

cartography

Dataset Cartography: Mapping and Diagnosing Datasets with Training Dynamics

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:188Issues:0Issues:0

direct-preference-optimization

Reference implementation for DPO (Direct Preference Optimization)

Language:PythonLicense:Apache-2.0Stargazers:2056Issues:0Issues:0

pytorch-lightning

Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.

Language:PythonLicense:Apache-2.0Stargazers:28085Issues:0Issues:0

ParroT

The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.

Language:PythonStargazers:166Issues:0Issues:0

ALMA

State-of-the-art LLM-based translation models.

Language:RubyLicense:MITStargazers:395Issues:0Issues:0

Tk-Instruct

Tk-Instruct is a Transformer model that is tuned to solve many NLP tasks by following instructions.

Language:PythonLicense:MITStargazers:177Issues:0Issues:0

Glot500

Glot500: Scaling Multilingual Corpora and Language Models to 500 Languages -- ACL 2023

Language:PythonLicense:NOASSERTIONStargazers:96Issues:0Issues:0

TaiLr

ICLR2023 - Tailoring Language Generation Models under Total Variation Distance

Language:PythonLicense:MITStargazers:21Issues:0Issues:0

RL4LMs

A modular RL library to fine-tune language models to human preferences

Language:PythonLicense:Apache-2.0Stargazers:2183Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:1214Issues:0Issues:0

DITTO

The code of paper "Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation" published at NeurIPS 2022

Language:PythonLicense:MITStargazers:39Issues:0Issues:0

COMET

A Neural Framework for MT Evaluation

Language:PythonLicense:Apache-2.0Stargazers:492Issues:0Issues:0

EKFAC-pytorch

Repository containing Pytorch code for EKFAC and K-FAC perconditioners.

Language:PythonLicense:MITStargazers:139Issues:0Issues:0

LASER

Language-Agnostic SEntence Representations

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:3583Issues:0Issues:0

llama-recipes

Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.

Language:Jupyter NotebookStargazers:11937Issues:0Issues:0

flores

Facebook Low Resource (FLoRes) MT Benchmark

Language:PythonLicense:NOASSERTIONStargazers:694Issues:0Issues:0

mtdata

A tool that locates, downloads, and extracts machine translation corpora

Language:PythonLicense:Apache-2.0Stargazers:147Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:55781Issues:0Issues:0

stopes

A library for preparing data for machine translation research (monolingual preprocessing, bitext mining, etc.) built by the FAIR NLLB team.

Language:PythonLicense:MITStargazers:247Issues:0Issues:0
Language:PythonStargazers:91Issues:0Issues:0

awesome-RLHF

A curated list of reinforcement learning with human feedback resources (continually updated)

License:Apache-2.0Stargazers:3285Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:10144Issues:0Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

Language:PythonLicense:Apache-2.0Stargazers:6855Issues:0Issues:0

xmtf

Crosslingual Generalization through Multitask Finetuning

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:513Issues:0Issues:0

Intra-Distillation

This is the repository for our EMNLP 2022 paper "The Importance of Being Parameters: An Intra-Distillation Method for Serious Gains".

Language:PythonStargazers:10Issues:0Issues:0

mesh-transformer-jax

Model parallel transformers in JAX and Haiku

Language:PythonLicense:Apache-2.0Stargazers:6281Issues:0Issues:0

pytorch-pruning

PyTorch Implementation of [1611.06440] Pruning Convolutional Neural Networks for Resource Efficient Inference

Language:PythonStargazers:873Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:6576Issues:0Issues:0

XLM

PyTorch original implementation of Cross-lingual Language Model Pretraining.

Language:PythonLicense:NOASSERTIONStargazers:2877Issues:0Issues:0

unify-parameter-efficient-tuning

Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)

Language:PythonLicense:Apache-2.0Stargazers:514Issues:0Issues:0