Ziwen Han (zw123han)

zw123han

Geek Repo

Company:University of Toronto

Github PK Tool:Github PK Tool

Ziwen Han's starred repositories

picoGPT

An unnecessarily tiny implementation of GPT-2 in NumPy.

Language:PythonLicense:MITStargazers:3215Issues:0Issues:0

TTS

🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production

Language:PythonLicense:MPL-2.0Stargazers:34752Issues:0Issues:0

ml-engineering

Machine Learning Engineering Open Book

Language:PythonLicense:CC-BY-SA-4.0Stargazers:11432Issues:0Issues:0

streaming-llm

[ICLR 2024] Efficient Streaming Language Models with Attention Sinks

Language:PythonLicense:MITStargazers:6610Issues:0Issues:0
Language:TypeScriptLicense:MITStargazers:3825Issues:0Issues:0

LLaMA-Factory

Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:32695Issues:0Issues:0

text-generation-inference

Large Language Model Text Generation Inference

Language:PythonLicense:Apache-2.0Stargazers:8924Issues:0Issues:0

audiocraft

Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.

Language:PythonLicense:MITStargazers:20792Issues:0Issues:0

llm-attacks

Universal and Transferable Attacks on Aligned Language Models

Language:PythonLicense:MITStargazers:3390Issues:0Issues:0

so-vits-svc-fork

so-vits-svc fork with realtime support, improved interface and more features.

Language:PythonLicense:NOASSERTIONStargazers:8750Issues:0Issues:0

so-vits-svc

SoftVC VITS Singing Voice Conversion

Language:PythonLicense:AGPL-3.0Stargazers:25677Issues:0Issues:0

tinygrad

You like pytorch? You like micrograd? You love tinygrad! ❤️

Language:PythonLicense:MITStargazers:26619Issues:0Issues:0

deep-language-networks

We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a Deep Language Network - DLN

Language:PythonLicense:MITStargazers:91Issues:0Issues:0

whisper

Robust Speech Recognition via Large-Scale Weak Supervision

Language:PythonLicense:MITStargazers:69917Issues:0Issues:0

CodeGeeX

CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)

Language:PythonLicense:Apache-2.0Stargazers:8191Issues:0Issues:0

evals

Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.

Language:PythonLicense:NOASSERTIONStargazers:14879Issues:0Issues:0

alpaca-lora

Instruct-tune LLaMA on consumer hardware

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:18588Issues:0Issues:0

alpaca.cpp

Locally run an Instruction-Tuned Chat-Style LLM

Language:CLicense:MITStargazers:10252Issues:0Issues:0

llama.cpp

LLM inference in C/C++

Language:C++License:MITStargazers:66510Issues:0Issues:0

JAX_llama

Inference code for LLaMA models in JAX

Language:PythonLicense:MITStargazers:111Issues:0Issues:0

stanford_alpaca

Code and documentation to train Stanford's Alpaca models, and generate the data.

Language:PythonLicense:Apache-2.0Stargazers:29429Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:133641Issues:0Issues:0
License:MITStargazers:175Issues:0Issues:0

paxml

Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.

Language:PythonLicense:Apache-2.0Stargazers:457Issues:0Issues:0

levanter

Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax

Language:PythonLicense:Apache-2.0Stargazers:502Issues:0Issues:0

flash-attention-jax

Implementation of Flash Attention in Jax

Language:PythonLicense:MITStargazers:190Issues:0Issues:0

workshops

This is a repository for all workshop related materials.

Language:Jupyter NotebookStargazers:204Issues:0Issues:0

PiPPy

Pipeline Parallelism for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:722Issues:0Issues:0

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

License:NOASSERTIONStargazers:26899Issues:0Issues:0

jax-llm-benchmarking

Scripts for benchmarking LLM fine-tuning throughput.

Language:PythonStargazers:1Issues:0Issues:0