Shahul ES (shahules786)

shahules786

Geek Repo

Location:Terra

Home Page:shahules786.github.io

Twitter:@shahules786

Github PK Tool:Github PK Tool


Organizations
explodinggradients

Shahul ES's starred repositories

LLaMA-Factory

Unify Efficient Fine-Tuning of 100+ LLMs

Language:PythonLicense:Apache-2.0Stargazers:23820Issues:162Issues:3752

axolotl

Go ahead and axolotl questions

Language:PythonLicense:Apache-2.0Stargazers:6459Issues:48Issues:579

streaming-llm

[ICLR 2024] Efficient Streaming Language Models with Attention Sinks

Language:PythonLicense:MITStargazers:6293Issues:61Issues:76

FlagEmbedding

Retrieval and Retrieval-augmented LLMs

Language:PythonLicense:MITStargazers:5519Issues:33Issues:778

openchat

OpenChat: Advancing Open-source Language Models with Imperfect Data

Language:PythonLicense:Apache-2.0Stargazers:5071Issues:51Issues:180

OpenCopilot

🤖 🔥 Language-to-actions engine

Language:TypeScriptLicense:MITStargazers:4949Issues:46Issues:99

LongLoRA

Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)

Language:PythonLicense:Apache-2.0Stargazers:2507Issues:13Issues:168

deepeval

The LLM Evaluation Framework

Language:PythonLicense:Apache-2.0Stargazers:2061Issues:15Issues:191

1password-teams-open-source

Get a free 1Password Teams membership for your open source project

fastmoe

A fast MoE impl for PyTorch

Language:PythonLicense:Apache-2.0Stargazers:1430Issues:12Issues:113

OpenMoE

A family of open-sourced Mixture-of-Experts (MoE) Large Language Models

mixture-of-experts

PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538

Language:PythonLicense:GPL-3.0Stargazers:870Issues:3Issues:25

punica

Serving multiple LoRA finetuned LLM as one

Language:PythonLicense:Apache-2.0Stargazers:862Issues:14Issues:37

SPIN

The official implementation of Self-Play Fine-Tuning (SPIN)

Language:PythonLicense:Apache-2.0Stargazers:853Issues:10Issues:26

awesome-mixture-of-experts

A collection of AWESOME things about mixture-of-experts

Language:PythonLicense:Apache-2.0Stargazers:765Issues:12Issues:34

vec2text

utilities for decoding deep representations (like sentence embeddings) back to text

Language:PythonLicense:NOASSERTIONStargazers:625Issues:13Issues:38

zeno-build

Build, evaluate, understand, and fix LLM-based apps

Language:Jupyter NotebookLicense:MITStargazers:480Issues:9Issues:69

Awesome-Mixture-of-Experts-Papers

A curated reading list of research in Mixture-of-Experts(MoE).

MS-AMP

Microsoft Automatic Mixed Precision Library

Language:PythonLicense:MITStargazers:471Issues:11Issues:57

pykoi-rlhf-finetuned-transformers

pykoi: Active learning in one unified interface

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:403Issues:10Issues:4

AnglE

Train and Infer Powerful Sentence Embeddings with AnglE | 🔥 SOTA on STS and MTEB Leaderboard

Language:PythonLicense:MITStargazers:380Issues:8Issues:40

st-moe-pytorch

Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch

Language:PythonLicense:MITStargazers:241Issues:5Issues:11

MoEBERT

This PyTorch package implements MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation (NAACL 2022).

Language:PythonLicense:Apache-2.0Stargazers:93Issues:1Issues:6

LLM-SLERP-Merge

Spherical Merge Pytorch/HF format Language Models with minimal feature loss.

MDEL

Multi-Domain Expert Learning

Language:PythonLicense:Apache-2.0Stargazers:68Issues:21Issues:29

soft-mixture-of-experts

PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)

Language:PythonLicense:MITStargazers:60Issues:3Issues:0

decontamination

This repository contains code for cleaning your training data of benchmark data to help combat data snooping.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:25Issues:3Issues:0

spade-experiments

Experiments to assess SPADE on different LLM pipelines.

Language:PythonStargazers:15Issues:0Issues:0