lipengyue (epic2005)

epic2005

Geek Repo

Company:Lenovo Inc.

Location:beijing china.

Twitter:@epic2005

Github PK Tool:Github PK Tool

lipengyue's starred repositories

cuOpt-Resources

A collection of NVIDIA cuOpt samples and other resources

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:136Issues:0Issues:0
Language:PythonStargazers:177Issues:0Issues:0

kubekey

Install Kubernetes/K3s only, both Kubernetes/K3s and KubeSphere, and related cloud-native add-ons, it supports all-in-one, multi-node, and HA 🔥 ⎈ 🐳

Language:GoLicense:Apache-2.0Stargazers:2190Issues:0Issues:0

CV-CUDA

CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.

Language:C++License:NOASSERTIONStargazers:2279Issues:0Issues:0

DeepSeek-Coder

DeepSeek Coder: Let the Code Write Itself

Language:PythonLicense:MITStargazers:6134Issues:0Issues:0

gpu-operator

NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes

Language:GoLicense:Apache-2.0Stargazers:1681Issues:0Issues:0

open-webui

User-friendly WebUI for LLMs (Formerly Ollama WebUI)

Language:SvelteLicense:MITStargazers:31926Issues:0Issues:0

clash-for-linux-backup

基于Clash Core 制作的Clash For Linux备份仓库 A Clash For Linux Backup Warehouse Based on Clash Core

Language:ShellLicense:GPL-3.0Stargazers:1799Issues:0Issues:0

inference

Reference implementations of MLPerf™ inference benchmarks

Language:PythonLicense:Apache-2.0Stargazers:1149Issues:0Issues:0
Language:Jupyter NotebookLicense:MITStargazers:197Issues:0Issues:0

clash-for-AutoDL

AutoDL平台服务器适配梯子, 使用 Clash 作为代理工具

Language:ShellStargazers:38Issues:0Issues:0

inference

Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.

Language:PythonLicense:Apache-2.0Stargazers:3789Issues:0Issues:0

ColossalAI

Making large AI models cheaper, faster and more accessible

Language:PythonLicense:Apache-2.0Stargazers:38382Issues:0Issues:0

transformers-benchmarks

real Transformer TeraFLOPS on various GPUs

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:833Issues:0Issues:0

triton

Development repository for the Triton language and compiler

Language:C++License:MITStargazers:12028Issues:0Issues:0

generative-ai-for-beginners

18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/

Language:Jupyter NotebookLicense:MITStargazers:57019Issues:0Issues:0

veScale

A PyTorch Native LLM Training Framework

Language:PythonLicense:Apache-2.0Stargazers:504Issues:0Issues:0

ray

Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.

Language:PythonLicense:Apache-2.0Stargazers:32165Issues:0Issues:0

DCGM

NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs

Language:C++License:Apache-2.0Stargazers:355Issues:0Issues:0

k8s-device-plugin

NVIDIA device plugin for Kubernetes

Language:GoLicense:Apache-2.0Stargazers:2575Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:9459Issues:0Issues:0

llama-recipes

Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.

Language:Jupyter NotebookStargazers:10528Issues:0Issues:0

LLaMA-Factory

A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:27194Issues:0Issues:0

Langchain-Chatchat

Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain

Language:TypeScriptLicense:Apache-2.0Stargazers:30030Issues:0Issues:0

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:1662Issues:0Issues:0

kubesphere

The container platform tailored for Kubernetes multi-cloud, datacenter, and edge management ⎈ 🖥 ☁️

Language:GoLicense:Apache-2.0Stargazers:14722Issues:0Issues:0

slurm

Slurm: A Highly Scalable Workload Manager

Language:CLicense:NOASSERTIONStargazers:2464Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:33933Issues:0Issues:0

drawio-desktop

Official electron build of draw.io

Language:JavaScriptLicense:Apache-2.0Stargazers:48193Issues:0Issues:0

LLM-quickstart

Quick Start for Large Language Models (Theoretical Learning and Practical Fine-tuning) 大语言模型快速入门(理论学习与微调实战)

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:390Issues:0Issues:0