SriKrishna Paparaju (spaparaju)

spaparaju

Geek Repo

Company:@RedHatOfficial

Github PK Tool:Github PK Tool

SriKrishna Paparaju's repositories

accelerate

πŸš€ A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DALI

A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

dbrx

Code examples and resources for DBRX, a large language model developed by Databricks

License:NOASSERTIONStargazers:0Issues:0Issues:0

dcgm-exporter

NVIDIA GPU metrics exporter for Prometheus leveraging DCGM

Language:GoLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeed-MII

MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.

License:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeedExamples

Example models using DeepSpeed

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

diffusers

πŸ€— Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch

License:Apache-2.0Stargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

dspy

DSPy: The framework for programmingβ€”not promptingβ€”foundation models

License:MITStargazers:0Issues:0Issues:0

faiss

A library for efficient similarity search and clustering of dense vectors.

License:MITStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

gpu-operator

NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes

Language:GoLicense:Apache-2.0Stargazers:0Issues:0Issues:0

graphrag

A modular graph-based Retrieval-Augmented Generation (RAG) system

License:MITStargazers:0Issues:0Issues:0

ignite

High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently.

License:BSD-3-ClauseStargazers:0Issues:0Issues:0

jupyterlab-nvdashboard

A JupyterLab extension for displaying dashboards of GPU usage.

Language:TypeScriptLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

kuberay

A toolkit to run Ray applications on Kubernetes

Language:GoLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

License:NOASSERTIONStargazers:0Issues:0Issues:0

mlflow

Open source platform for the machine learning lifecycle

License:Apache-2.0Stargazers:0Issues:0Issues:0

nim-anywhere

Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

nim-deploy

A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

optimum

πŸš€ Accelerate training and inference of πŸ€— Transformers and πŸ€— Diffusers with easy to use hardware optimization tools

License:Apache-2.0Stargazers:0Issues:0Issues:0

ray

Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

streaming

A Data Streaming Library for Efficient Neural Network Training

License:Apache-2.0Stargazers:0Issues:0Issues:0

TensorRT-LLM

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

TensorRT-Model-Optimizer

TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization and sparsity. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

torchtune

A Native-PyTorch Library for LLM Fine-tuning

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

transformers

πŸ€— Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

triton

Development repository for the Triton language and compiler

Language:C++License:MITStargazers:0Issues:0Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0