Juncheng's starred repositories
LLocalSearch
LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.
algebraic-nnhw
AI acceleration using matrix multiplication with half the multiplications
opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
FriendsDontLetFriends
Friends don't let friends make certain types of data visualization - What are they and why are they bad.
screenshot-to-code
Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
lm-format-enforcer
Enforce the output format (JSON Schema, Regex etc) of a language model
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
text-embeddings-inference
A blazing fast inference solution for text embeddings models
uptime-kuma
A fancy self-hosted monitoring tool
openstatus
🏓 The open-source synthetic & real user monitoring platform 🏓
KVM-Opencore
OpenCore disk image for running macOS VMs on Proxmox/QEMU
alaz
Alaz: Advanced eBPF Agent for Kubernetes Observability – Effortlessly monitor K8s service interactions and performance metrics in your K8s environment. Gain in-depth insights with service maps, metrics, distributed tracing, and more, while staying alert to crucial system anomalies 🐝
whisper.api
This project provides an API with user level access support to transcribe speech to text using a finetuned and processed Whisper ASR model.
llm-engine
Scale LLM Engine public repository
inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
openmlsys-zh
《Machine Learning Systems: Design and Implementation》- Chinese Version