Dip_an 's repositories

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

accelerate

๐Ÿš€ A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

halutmatmul_for_windows

Stella Nera is the first Maddness accelerator achieving 15x higher area efficiency (GMAC/s/mm^2) and 25x higher energy efficiency (TMAC/s/W) than direct MatMul accelerators in the same technology

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Open-Llama

The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

License:BSD-3-ClauseStargazers:0Issues:0Issues:0

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

License:Apache-2.0Stargazers:0Issues:0Issues:0

grok-1

Grok open release

License:Apache-2.0Stargazers:0Issues:0Issues:0

lightning-attention

Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models

License:MITStargazers:0Issues:0Issues:0

linear_open_lm

A repository for research on medium sized language models.

License:MITStargazers:0Issues:0Issues:0

llama3

The official Meta Llama 3 GitHub site

License:NOASSERTIONStargazers:0Issues:0Issues:0

llamafile

Distribute and run LLMs with a single file.

License:NOASSERTIONStargazers:0Issues:0Issues:0

LLM-Agents-Papers

A repo lists papers related to LLM based agent

Stargazers:0Issues:0Issues:0

llm-foundry

LLM training code for Databricks foundation models

License:Apache-2.0Stargazers:0Issues:0Issues:0

LLMTest_NeedleInAHaystack

Doing simple retrieval from LLM models at various context lengths to measure accuracy

License:NOASSERTIONStargazers:0Issues:0Issues:0

MetaGPT

๐ŸŒŸ The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming

License:MITStargazers:0Issues:0Issues:0

minLlama3

a quick & complete guide to Llama 3's architecture

Language:PythonStargazers:0Issues:0Issues:0

nanoGPT-TK

The simplest, fastest repository for training/finetuning medium-sized GPTs. Now, with kittens!

Language:MakefileLicense:MITStargazers:0Issues:0Issues:0

ollama

Get up and running with Llama 3, Mistral, Gemma, and other large language models.

Language:GoLicense:MITStargazers:0Issues:0Issues:0

othello_mamba

Evaluating the Mamba architecture on the Othello game

Stargazers:0Issues:0Issues:0

pykan

Kolmogorov Arnold Networks

License:MITStargazers:0Issues:0Issues:0

pythia

The hub for EleutherAI's work on interpretability and learning dynamics

License:Apache-2.0Stargazers:0Issues:0Issues:0
License:MITStargazers:0Issues:0Issues:0

ThunderKittens

Tile primitives for speedy kernels

License:MITStargazers:0Issues:0Issues:0

tiny-gpu

A minimal GPU design in Verilog to learn how GPUs work from the ground up

Stargazers:0Issues:0Issues:0

torchscale

Foundation Architecture for (M)LLMs

License:MITStargazers:0Issues:0Issues:0

unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

License:MITStargazers:0Issues:0Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

X_net

a new transformer architecture

Language:PythonStargazers:0Issues:0Issues:0