kamillle's starred repositories
segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
awesome-japanese-llm
日本語LLMまとめ - Overview of Japanese LLMs
DeepFilterNet
Noise supression using deep filtering
vits2_pytorch
unofficial vits2-TTS implementation in pytorch
sticky-pull-request-comment
create comment on pull request, if exists update that comment.
karpenter-provider-aws
Karpenter is a Kubernetes Node Autoscaler built for flexibility, performance, and simplicity.
stable-diffusion-webui-colab
stable diffusion webui colab
fastembed-rs
Library for generating vector embeddings, reranking in Rust
ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
awsome-distributed-training
Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.
guidance-for-machine-learning-inference-on-aws
This Guidance demonstrates how to deploy a machine learning inference architecture on Amazon Elastic Kubernetes Service (Amazon EKS). It addresses the basic implementation requirements as well as ways you can pack thousands of unique PyTorch deep learning (DL) models into a scalable architecture and evaluate performance at scale
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
intel-extension-for-pytorch
A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
guardrails
Adding guardrails to large language models.