BaaBaa (JoeyYoung)

JoeyYoung

Geek Repo

Company:The University of HongKong

Github PK Tool:Github PK Tool

BaaBaa's starred repositories

system-design-primer

Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.

Language:PythonLicense:NOASSERTIONStargazers:259972Issues:6663Issues:283

gpt_academic

为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, moss等。

Language:PythonLicense:GPL-3.0Stargazers:60122Issues:254Issues:1492
Language:PythonLicense:NOASSERTIONStargazers:34456Issues:309Issues:348

Awesome-LLM

Awesome-LLM: a curated list of Large Language Model

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:9016Issues:156Issues:551

coder-kung-fu

开发内功修炼

Language:CLicense:Apache-2.0Stargazers:5977Issues:176Issues:6

bob-plugin-openai-translator

基于 ChatGPT API 的文本翻译、文本润色、语法纠错 Bob 插件,让我们一起迎接不需要巴别塔的新时代!Licensed under CC BY-NC-SA 4.0

Language:JavaScriptLicense:NOASSERTIONStargazers:5467Issues:28Issues:92

k8s-device-plugin

NVIDIA device plugin for Kubernetes

Language:GoLicense:Apache-2.0Stargazers:2487Issues:63Issues:440

Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Language:PythonLicense:NOASSERTIONStargazers:1687Issues:23Issues:168

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:1536Issues:34Issues:224

ByteTransformer

optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052

Language:C++License:Apache-2.0Stargazers:430Issues:10Issues:10

clearml-agent

ClearML Agent - ML-Ops made easy. ML-Ops scheduler & orchestration solution

Language:PythonLicense:Apache-2.0Stargazers:228Issues:16Issues:164

rccl

ROCm Communication Collectives Library (RCCL)

Language:C++License:NOASSERTIONStargazers:223Issues:32Issues:74

oneCCL

oneAPI Collective Communications Library (oneCCL)

Language:C++License:NOASSERTIONStargazers:177Issues:23Issues:33

nccl-rdma-sharp-plugins

RDMA and SHARP plugins for nccl library

Language:CLicense:BSD-3-ClauseStargazers:143Issues:22Issues:17

awesome-Auto-Parallelism

A baseline repository of Auto-Parallelism in Training Neural Networks

Language:PythonStargazers:131Issues:3Issues:0

csconferences

Major CS conference publication stats (including accepted and submitted) by year.

orion

An interference-aware scheduler for fine-grained GPU sharing

Language:PythonLicense:MITStargazers:72Issues:2Issues:16

kubernetes-scheduler-simulator

Kubernetes Scheduler Simulator

Language:ShellLicense:Apache-2.0Stargazers:65Issues:4Issues:3

TGS

Artifacts for our NSDI'23 paper TGS

Language:PythonLicense:Apache-2.0Stargazers:56Issues:3Issues:3

brainstorm

Compiler for Dynamic Neural Networks

teal

Codebase for Teal (SIGCOMM 2023)

Language:PythonLicense:MITStargazers:37Issues:1Issues:1

taccl

TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches

Language:PythonLicense:MITStargazers:33Issues:6Issues:2

AdaQP

Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training

Language:PythonLicense:MITStargazers:17Issues:1Issues:2

dear_pytorch

[ICDCS 2023] DeAR: Accelerating Distributed Deep Learning with Fine-Grained All-Reduce Pipelining

Language:PythonStargazers:10Issues:1Issues:0
Language:PythonStargazers:7Issues:0Issues:0
Language:PythonStargazers:5Issues:0Issues:0