NVIDIA Corporation (NVIDIA)

NVIDIA Corporation

NVIDIA

Organization data from Github https://github.com/NVIDIA

Location:2788 San Tomas Expressway, Santa Clara, CA, 95051

Home Page:https://nvidia.com

GitHub:@NVIDIA

NVIDIA Corporation's repositories

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:14129Issues:171Issues:1173

TensorRT-LLM

TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

Language:C++License:Apache-2.0Stargazers:12065Issues:120Issues:2951

cuda-python

CUDA Python: Performance meets Productivity

Language:PythonLicense:NOASSERTIONStargazers:3023Issues:46Issues:599

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:2891Issues:36Issues:556

nv-ingest

NeMo Retriever extraction is a scalable, performance-oriented document content and metadata extraction microservice. NeMo Retriever extraction uses specialized NVIDIA NIM microservices to find, contextualize, and extract text, tables, charts and images that you can use in downstream generative applications.

Language:PythonLicense:Apache-2.0Stargazers:2760Issues:28Issues:183

gpu-operator

NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes

Language:GoLicense:Apache-2.0Stargazers:2387Issues:50Issues:984

stdexec

`std::execution`, the proposed C++ framework for asynchronous and parallel programming.

Language:C++License:Apache-2.0Stargazers:2074Issues:58Issues:606

cccl

CUDA Core Compute Libraries

Language:C++License:NOASSERTIONStargazers:2016Issues:32Issues:2445

aistore

AIStore: scalable storage for AI applications

TensorRT-Model-Optimizer

A unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed.

Language:PythonLicense:Apache-2.0Stargazers:1519Issues:24Issues:237

NeMo-Agent-Toolkit

The NVIDIA NeMo Agent toolkit is an open-source library for efficiently connecting and optimizing teams of AI agents.

Language:PythonLicense:Apache-2.0Stargazers:1497Issues:24Issues:237

KAI-Scheduler

KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

Language:GoLicense:Apache-2.0Stargazers:896Issues:18Issues:86

cuda-quantum

C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows

Language:C++License:NOASSERTIONStargazers:843Issues:22Issues:1001

NeMo-Skills

A project to improve skills of large language models

Language:PythonLicense:Apache-2.0Stargazers:568Issues:19Issues:112

bionemo-framework

BioNeMo Framework: For building and adapting AI models in drug discovery at scale

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:563Issues:41Issues:110

cuopt

GPU accelerated decision optimization

Language:CudaLicense:Apache-2.0Stargazers:436Issues:11Issues:182

tilus

Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.

Language:PythonLicense:Apache-2.0Stargazers:367Issues:4Issues:11

JAX-Toolbox

JAX-Toolbox

Language:PythonLicense:Apache-2.0Stargazers:359Issues:20Issues:184

Fuser

A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")

Language:C++License:NOASSERTIONStargazers:356Issues:16Issues:978

mig-parted

MIG Partition Editor for NVIDIA GPUs

Language:GoLicense:Apache-2.0Stargazers:223Issues:11Issues:38

nim-deploy

A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:204Issues:9Issues:32

recsys-examples

Examples for Recommenders - easy to train and deploy on accelerated infrastructure.

Language:PythonLicense:NOASSERTIONStargazers:165Issues:0Issues:0

vgpu-device-manager

NVIDIA vGPU Device Manager manages NVIDIA vGPU devices on top of Kubernetes

Language:GoLicense:Apache-2.0Stargazers:142Issues:9Issues:0

cudaqx

Accelerated libraries for quantum-classical computing built on CUDA-Q.

Language:C++License:NOASSERTIONStargazers:65Issues:10Issues:47

NV-Kernels

Ubuntu kernels which are optimized for NVIDIA server systems

Language:CLicense:NOASSERTIONStargazers:64Issues:5Issues:1

gontainer

Simple but powerful dependency injection container for Go projects!

Language:GoLicense:Apache-2.0Stargazers:57Issues:4Issues:6

doca-platform

DOCA Platform manages provisioning and service orchestration for Bluefield DPUs

Language:GoLicense:Apache-2.0Stargazers:51Issues:11Issues:0

spark-rapids-jni

RAPIDS Accelerator JNI For Apache Spark

Language:CudaLicense:Apache-2.0Stargazers:51Issues:19Issues:318

cloud-native-docs

Documentation repository for NVIDIA Cloud Native Technologies

Language:PowerShellLicense:Apache-2.0Stargazers:29Issues:10Issues:10

doca-sosreport

A unified tool for collecting system logs and other debug information

Language:PythonLicense:GPL-2.0Stargazers:4Issues:0Issues:0