Bing Bo (bbo0924)

bbo0924

Geek Repo

Github PK Tool:Github PK Tool

Bing Bo's starred repositories

Language:PythonLicense:NOASSERTIONStargazers:43Issues:0Issues:0

datatrove

Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.

Language:PythonLicense:Apache-2.0Stargazers:1855Issues:0Issues:0

sglang

SGLang is yet another fast serving framework for large language models and vision language models.

Language:PythonLicense:Apache-2.0Stargazers:4158Issues:0Issues:0

mem0

The memory layer for Personalized AI

Language:PythonLicense:Apache-2.0Stargazers:19599Issues:0Issues:0

LLM101n

LLM101n: Let's build a Storyteller

Stargazers:26991Issues:0Issues:0

AIOS

AIOS: LLM Agent Operating System

Language:PythonLicense:MITStargazers:3130Issues:0Issues:0

OpenAGI

OpenAGI: When LLM Meets Domain Experts

Language:PythonLicense:MITStargazers:1886Issues:0Issues:0

JioNLP

中文 NLP 预处理、解析工具包,准确、高效、易用 A Chinese NLP Preprocessing & Parsing Package www.jionlp.com

Language:PythonLicense:Apache-2.0Stargazers:3209Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:393Issues:0Issues:0

EasyNLP

EasyNLP: A Comprehensive and Easy-to-use NLP Toolkit

Language:PythonLicense:Apache-2.0Stargazers:2017Issues:0Issues:0

lmdeploy

LMDeploy is a toolkit for compressing, deploying, and serving LLMs.

Language:PythonLicense:Apache-2.0Stargazers:3887Issues:0Issues:0

AutoAWQ

AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:

Language:PythonLicense:MITStargazers:1555Issues:0Issues:0

dive-into-llms

《动手学大模型Dive into LLMs》系列编程实践教程

Stargazers:2990Issues:0Issues:0

llama2.c

Inference Llama 2 in one file of pure C

Language:CLicense:MITStargazers:17019Issues:0Issues:0

qserve

QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving

Language:PythonLicense:Apache-2.0Stargazers:374Issues:0Issues:0

TensorRT-LLM

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

Language:C++License:Apache-2.0Stargazers:7923Issues:0Issues:0

llm-action

本项目旨在分享大模型相关技术原理以及实战经验。

Language:HTMLLicense:Apache-2.0Stargazers:8564Issues:0Issues:0

llama-cpp-python

Python bindings for llama.cpp

Language:PythonLicense:MITStargazers:7483Issues:0Issues:0

LLM-Tuning

Tuning LLMs with no tears💦; Sample Design Engineering (SDE) for more efficient downstream-tuning.

Language:HTMLStargazers:952Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:382Issues:0Issues:0

dify

Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.

Language:TypeScriptLicense:NOASSERTIONStargazers:41906Issues:0Issues:0

llm-course

Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:35839Issues:0Issues:0

ColBERT

ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)

Language:PythonLicense:MITStargazers:2774Issues:0Issues:0

ragflow

RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.

Language:PythonLicense:Apache-2.0Stargazers:14155Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:22657Issues:0Issues:0

Open-Sora

Open-Sora: Democratizing Efficient Video Production for All

Language:PythonLicense:Apache-2.0Stargazers:21241Issues:0Issues:0

NexusRaven

NexusRaven-13B, a new SOTA Open-Source LLM for function calling. This repo contains everything for reproducing our evaluation on NexusRaven-13B and baselines.

Language:PythonLicense:Apache-2.0Stargazers:294Issues:0Issues:0

ChatGPTX-Uni

实现一种多Lora权值集成切换+Zero-Finetune零微调增强的跨模型技术方案,LLM-Base+LLM-X+Alpaca,初期,LLM-Base为Chatglm6B底座模型,LLM-X是LLAMA增强模型。该方案简易高效,目标是使此类语言模型能够低能耗广泛部署,并最终在小模型的基座上发生“智能涌现”,力图最小计算代价达成ChatGPT、GPT4、ChatRWKV等人类友好亲和效果。当前可以满足总结、提问、问答、摘要、改写、评论、扮演等各种需求。

Language:PythonLicense:GPL-3.0Stargazers:118Issues:0Issues:0

ChatGLM3

ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型

Language:PythonLicense:Apache-2.0Stargazers:13258Issues:0Issues:0

direct-preference-optimization

Reference implementation for DPO (Direct Preference Optimization)

Language:PythonLicense:Apache-2.0Stargazers:1952Issues:0Issues:0