Ankit Shah's repositories
awesome-terminal-hacks
A repository consisting of useful terminal commands required in daily tasks to reduce stackoverflow searches.
AAR
[Official Implementation] Acoustic Autoregressive Modeling 🔥
agihouse-werewolf-template
Template repository for the Werewolf hackathon
ATST-SED
This repo includes the official implementations of "Fine-tune the pretrained ATST model for sound event detection".
autogen
A programming framework for agentic AI 🤖
aws-elb-best-practices
ELB Best Practices Guides
blog.eleven-labs.com
Eleven-labs blog
bolt.new
Prompt, run, edit, and deploy full-stack web applications
cleanlab
The standard package for machine learning with noisy labels and finding mislabeled data. Works with most datasets and models.
dspy
DSPy: The framework for programming—not prompting—foundation models
gpu-operator
NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes
groq-moa
Mixture of Agents using Groq
litellm
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
morphic
An AI-powered search engine with a generative UI
msclap_copy
msclap_copy version
nim-deploy
A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.
openagents
An open agents platform (v1; deprecated)
OpenDevin
🐚 OpenDevin: Code Less, Make More
OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
prometheus
The Prometheus monitoring system and time series database.
SpeechTranslation
SpeechTranslation
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs