Amanda-Barbara's repositories
autogen
Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ
blora-text-generation-inference
Batched LORA + Continuous Batching
BLoRA-TGI-with-python-server
Batched Lora + Continuous Batching
ComputeLibrary
The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologies.
CPlusPlus-Tutorial
C++ Tutorial
CTC-loss-introduction
介绍ctc算法原理以及numpy简单实现
flash-attention
Fast and memory-efficient exact attention
generative-ai-for-beginners
12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
gpu-profiling
GPU Profiling
kohya_ss
train stable diffusion models
langchain
⚡ Building applications with LLMs through composability ⚡
Latte
The official implementation of Latte: Latent Diffusion Transformer for Video Generation.
leetcode-master
《代码随想录》LeetCode 刷题攻略:200道经典题目刷题顺序,共60w字的详细图解,视频难点剖析,50余张思维导图,支持C++,Java,Python,Go,JavaScript等多语言版本,从此算法学习不再迷茫!🔥🔥 来看看,你会发现相见恨晚!🚀
llm-benchmark-test
include LLM open framework measurement
MetaGPT-agent
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
multimodal-ai-jina
☁️ Build multimodal AI applications with cloud-native stack
opencl-intercept-layer
Intercept Layer for Debugging and Analyzing OpenCL Applications
openmlsys-zh
《Machine Learning Systems: Design and Implementation》- Chinese Version
pocl
pocl - Portable Computing Language
stable-diffusion
A latent text-to-image diffusion model
text2video-generative-models
Generative Models by Stability AI
tgi-benchmarking
Benchmarking LLMs on GPUs
tvm
Open deep learning compiler stack for cpu, gpu and specialized accelerators
tvm-mlc-llm
Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
yolov5-5.x-annotations
一个基于yolov5-5.0的中文注释版本!