Shahul ES's starred repositories
LLaMA-Factory
Unify Efficient Fine-Tuning of 100+ LLMs
streaming-llm
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
FlagEmbedding
Retrieval and Retrieval-augmented LLMs
OpenCopilot
🤖 🔥 Language-to-actions engine
1password-teams-open-source
Get a free 1Password Teams membership for your open source project
mixture-of-experts
PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538
awesome-mixture-of-experts
A collection of AWESOME things about mixture-of-experts
zeno-build
Build, evaluate, understand, and fix LLM-based apps
Awesome-Mixture-of-Experts-Papers
A curated reading list of research in Mixture-of-Experts(MoE).
pykoi-rlhf-finetuned-transformers
pykoi: Active learning in one unified interface
st-moe-pytorch
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
LLM-SLERP-Merge
Spherical Merge Pytorch/HF format Language Models with minimal feature loss.
soft-mixture-of-experts
PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)
decontamination
This repository contains code for cleaning your training data of benchmark data to help combat data snooping.
spade-experiments
Experiments to assess SPADE on different LLM pipelines.