MAKOTO (makotov)

makotov

Geek Repo

Company:Morgan Stanley Science Investment Co.,LTD

Location:London, UK

Github PK Tool:Github PK Tool

MAKOTO's repositories

accelerate

🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Chinese-LLaMA-Alpaca-2

中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Chinese-LLaMA-Alpaca-3

中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Chinese-Mixtral

中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

codellama

Inference code for CodeLlama models

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

DongwuLLM

This is the codebase for pre-training, compressing, extending, and distilling LLMs with Megatron-LM.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:GoLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llama.go

llama.go is like llama.cpp in pure Golang!

Language:GoLicense:NOASSERTIONStargazers:0Issues:0Issues:0

LLM-Pruner

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

LLM-Shearing

[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

Open-Sora

Open-Sora: Democratizing Efficient Video Production for All

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

promptulate_new

A large language model automation and Autonomous Language Agents development framework. Build your LLM Agent Application in a pythonic way!

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Pruning-LLMs

The framework to prune LLMs to any size and any config.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

k8s-vgpu-scheduler

OpenAIOS vGPU device plugin for Kubernetes is originated from the OpenAIOS project to virtualize GPU device memory, in order to allow applications to access larger memory space than its physical capacity. It is designed for ease of use of extended device memory for AI workloads.

License:Apache-2.0Stargazers:0Issues:0Issues:0

Llama2-Code-Interpreter

Make Llama2 use Code Execution, Debug, Save Code, Reuse it, Access to Internet

Stargazers:0Issues:0Issues:0

LLaVA-pp

🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)

Stargazers:0Issues:0Issues:0

mixture-of-experts

PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538

License:GPL-3.0Stargazers:0Issues:0Issues:0

MS-DOS

The original sources of MS-DOS 1.25, 2.0, and 4.0 for reference purposes

License:MITStargazers:0Issues:0Issues:0

xllm

🦖 X—LLM: Cutting Edge & Easy LLM Finetuning

License:Apache-2.0Stargazers:0Issues:0Issues:0