deepuncertainty's repositories
guidance
A guidance language for controlling large language models.
petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
web-llm
Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.
long-form-factuality
Benchmarking long-form factuality in large language models. Original code for our paper "Long-form factuality in large language models".
chat.petals.dev
💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client
trl
Train transformer language models with reinforcement learning.
dspy
DSPy: The framework for programming—not prompting—foundation models
NeMo-Guardrails
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
web-stable-diffusion
Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.
instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
jsonformer
A Bulletproof Way to Generate Structured JSON from Language Models
alpaca_farm
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
basaran
Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Transformers-based text generation models.
trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
web-ai
Run modern deep learning models in the browser.
LLM-Trainer
Train LLMs by just modifying config files!
gpt-tfjs
GPT in TensorFlow.js
petals.dev
🌸 Run 100B+ language models at home, BitTorrent-style. Landing page for the project
lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
FastChat
The release repo for "Vicuna: An Open Chatbot Impressing GPT-4"
pyllama
LLaMA: Open and Efficient Foundation Language Models
alpaca-lora
Instruct-tune LLaMA on consumer hardware
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
vanilla-llama
Plain pytorch implementation of LLaMA
alpaca.cpp
Locally run an Instruction-Tuned Chat-Style LLM
nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.