Oleg Ovcharenko's starred repositories
ndc_dapt_playbook
Scalable toolkit for data curation
Perplexica
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
Scrapegraph-ai
Python scraper based on AI
llama-recipes
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.
open-gpu-kernel-modules
NVIDIA Linux open GPU with P2P support
NeMo-Curator
Scalable toolkit for data curation
EAGE-Hackathon-2024-Instructions
Here you will find all the info you need to know to participate in the 2024 EAGE Annual Hackathon in Oslo!
NeMo-Aligner
Scalable toolkit for efficient model alignment
lm-evaluation-harness
A framework for few-shot evaluation of language models.
LLMs-from-scratch
Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step
tensorrtllm_backend
The Triton TensorRT-LLM Backend
json_repair
A python module to repair invalid JSON, commonly used to parse the output of LLMs
modulus-sym
Framework providing pythonic APIs, algorithms and utilities to be used with Modulus core to physics inform model training as well as higher level abstraction for domain experts
jetson-inference
Hello AI World guide to deploying deep-learning inference networks and deep vision primitives with TensorRT and NVIDIA Jetson.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
modulus-makani
Massively parallel training of machine-learning based weather and climate models
chatbot-ui
AI chat for every model.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
cartography
Dataset Cartography: Mapping and Diagnosing Datasets with Training Dynamics
annotated_deep_learning_paper_implementations
๐งโ๐ซ 60 Implementations/tutorials of deep learning papers with side-by-side notes ๐; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), ๐ฎ reinforcement learning (ppo, dqn), capsnet, distillation, ... ๐ง
text-generation-inference
Large Language Model Text Generation Inference