theblackcat102's starred repositories
CTranslate2
Fast inference engine for Transformer models
resource-stream
CUDA related news and material links
awesome-mixture-of-experts
A collection of AWESOME things about mixture-of-experts
adept-inference
Inference code for Persimmon-8B
self-correction-llm-papers
This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.
neurips_llm_efficiency_challenge
NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day
ModuleFormer
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
ToolkenGPT
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)
TransformerPrograms
[NeurIPS 2023] Learning Transformer Programs
ml-calibration
relplot: Utilities for measuring calibration and plotting reliability diagrams
GPU-Puzzles
Solve puzzles. Learn CUDA.
Beyond-Neural-Scaling
Implementation of Beyond Neural Scaling beating power laws for deep models and prototype-based models
arc-agents
Experiments with LLMs on the Abstraction and Reasoning Corpus (ARC)
json-schema-corpus
Corpus of over 80thousand JSON Schema documents, collected from open source GitHub repositories.