Jonathan Tow (jon-tow)

jon-tow

Geek Repo

Location:New York, New York

Home Page:https://jon-tow.github.io

Github PK Tool:Github PK Tool


Organizations
EleutherAI

Jonathan Tow's repositories

cs224n

Solutions to CS224n: Natural Language Processing with Deep Learning assignments.

text-sed

Implementation of Self-conditioned Embedding Diffusion for Text Generation

Language:PythonLicense:MITStargazers:35Issues:2Issues:2

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:2Issues:1Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

accelerate

🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

jon-tow.github.io

My personal website

Language:HTMLStargazers:0Issues:1Issues:0

cc_net

Tools to download and cleanup Common Crawl data

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

contriever

Contriever: Unsupervised Dense Information Retrieval with Contrastive Learning

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

CPCargo

A simple package to upload DL checkpoints to remote storage

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:Jupyter NotebookLicense:NOASSERTIONStargazers:0Issues:0Issues:0

english-wordnet

The Open English WordNet

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

goodreads

code samples for the goodreads datasets

License:Apache-2.0Stargazers:0Issues:0Issues:0

helm

Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:RustLicense:Apache-2.0Stargazers:0Issues:0Issues:0

kernl

Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:MITStargazers:0Issues:1Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Megatron-LLM

distributed trainer for LLMs

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

ml-engineering

Machine Learning Engineering Guides and Tools

Language:PythonLicense:CC-BY-SA-4.0Stargazers:0Issues:0Issues:0

ok

Codex-based command line assistant

Language:C++Stargazers:0Issues:1Issues:0

rerope

Rectified Rotary Position Embeddings

Language:PythonStargazers:0Issues:0Issues:0

ring-flash-attention

Ring attention implementation with flash attention

Language:PythonStargazers:0Issues:0Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0

scattermoe

Triton-based implementation of Sparse Mixture of Experts.

License:Apache-2.0Stargazers:0Issues:0Issues:0

text-dedup

All-in-one text de-duplication

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

triton

Development repository for the Triton language and compiler

License:MITStargazers:0Issues:0Issues:0

zero-bubble-pipeline-parallelism

Zero Bubble Pipeline Parallelism

License:NOASSERTIONStargazers:0Issues:0Issues:0