Uranus's repositories
llama_generative_agent
A generative agent implementation for LLaMA based models, derived from langchain's implementation.
qing_bureau_of_construction
This project employs Optical Character Recognition (OCR) to digitize historical records from the Qing manufacturing office.
Effective-LLM-Inference-Evaluation
A project aimed at measuring the real-world performance of Large Language Model (LLM) inference frameworks, inspired by the concepts in deepspeed-fastgen.
github-readme-stats
:zap: Dynamically generated stats for your github readmes
llama-recipes
Examples and recipes for Llama 2 model
UranusSeven
Config files for my GitHub profile.
alpa
Training and serving large-scale neural networks
arrow-datafusion-python
Apache Arrow DataFusion Python Bindings
CnOCR
CnOCR: Awesome Chinese/English OCR toolkits based on PyTorch/MXNet, It comes with 20+ well-trained models for different application scenarios and can be used directly after installation. 【基于 PyTorch/MXNet 的中文/英文 OCR Python 包。】
examples
Xorbits Example Notebooks
file-gpt
Start a chat with any document with Ada Embedding and Davinci Completion
mars
Mars is a tensor-based unified framework for large-scale data computation which scales numpy, pandas, scikit-learn and Python functions.
pandas
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
velox
A C++ vectorized database acceleration library aimed to optimizing query engines and data processing systems.
cursor
An editor made for programming with AI 🤖
dify
One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.
EAGLE
EAGLE: Lossless Acceleration of LLM Decoding by Feature Extrapolation
langchain
⚡ Building applications with LLMs through composability ⚡
lightllm
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
llama.cpp
Port of Facebook's LLaMA model in C/C++
LLMSpeculativeSampling
Fast inference from large lauguage models via speculative decoding
open-interpreter
OpenAI's Code Interpreter in your terminal, running locally
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
xoscar
Python actor framework for heterogeneous computing.