RossSong's repositories
alpaca-lora
Code for reproducing the Stanford Alpaca InstructLLaMA result on consumer hardware
AlpacaDataCleaned
Alpaca dataset from Stanford, cleaned and curated
awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT better.
bloomz.cpp
C++ implementation for BLOOM
CoLT5-attention
Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch
dolly
Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform
FastChat
The release repo for "Vicuna: An Open Chatbot Impressing GPT-4"
flash-attention
Fast and memory-efficient exact attention
FlexGen
Running large language models like GPT-3/OPT-175B on a single GPU. Up to 100x faster than other offloading systems.
gpt4all.cpp
Locally run an Assistant-Tuned Chat-Style LLM
guidance
A guidance language for controlling large language models.
KoChatGPT
ChatGPT의 RLHF를 학습을 위한 3가지 step별 한국어 데이터셋
LMOps
General technology for enabling AI capabilities w/ LLMs and Generative AI models
local-attention
An implementation of local windowed attention for language modeling
Long-context-transformers
Exploring finetuning public checkpoints on filter 8K sequences on Pile
musiclm-pytorch
Implementation of MusicLM, Google's new SOTA model for music generation using attention networks, in Pytorch
nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
NPM
The original implementation of Min et al. "Nonparametric Masked Language Modeling" (paper https//arxiv.org/abs/2212.01349)
OpenELM
Evolution Through Large Models
privateGPT
Interact privately with your documents using the power of GPT, 100% privately, no data leaks
prompt-engine
A library for helping developers craft prompts for Large Language Models
Prompt-Engineering-Guide
:octopus: Guide and resources for prompt engineering
Promptify
Prompt Engineering | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engineering, LLMs and other latest research
self-instruct
Aligning pretrained language models with instruction data generated by themselves.
setfit
Efficient few-shot learning with Sentence Transformers
spacy-llm
🦙 Integrating LLMs into structured NLP pipelines
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
torchscale
Transformers at any scale
unlimiformer
Public repo for the preprint "Unlimiformer: Long-Range Transformers with Unlimited Length Input"