onenight's starred repositories

AutoGPT

AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.

Language:PythonLicense:MITStargazers:165936Issues:1553Issues:2543

gpt_academic

为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, moss等。

Language:PythonLicense:GPL-3.0Stargazers:63351Issues:265Issues:1555

gpt4free

The official gpt4free repository | various collection of powerful language models

Language:PythonLicense:GPL-3.0Stargazers:59722Issues:464Issues:1298

bert

TensorFlow code and pre-trained models for BERT

Language:PythonLicense:Apache-2.0Stargazers:37696Issues:998Issues:1142

LLaMA-Factory

Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:29319Issues:191Issues:4611

coder2gwy

互联网首份程序员考公指南,由3位已经进入体制内的前大厂程序员联合献上。

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

ragflow

RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.

Language:PythonLicense:Apache-2.0Stargazers:14894Issues:97Issues:956

RWKV-LM

RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

Language:PythonLicense:Apache-2.0Stargazers:12130Issues:134Issues:199

awesome-knowledge-distillation

Awesome Knowledge Distillation

MTBook

《机器翻译:基础与模型》肖桐 朱靖波 著 - Machine Translation: Foundations and Models

hugging-llm

HuggingLLM, Hugging Future.

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:2688Issues:39Issues:11

TextBrewer

A PyTorch-based knowledge distillation toolkit for natural language processing

Language:PythonLicense:Apache-2.0Stargazers:1582Issues:27Issues:104

lm-human-preferences

Code for the paper Fine-Tuning Language Models from Human Preferences

Language:PythonLicense:MITStargazers:1194Issues:25Issues:15

awesome_LLMs_interview_notes

LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案

data-centric-AI

A curated, but incomplete, list of data-centric AI resources.

RAG-Retrieval

Unify Efficient Fine-tuning of RAG Retrieval, including Embedding, ColBERT,Cross Encoder

Language:PythonLicense:MITStargazers:402Issues:6Issues:22
Language:PythonLicense:MITStargazers:311Issues:8Issues:8

NBCE

Naive Bayes-based Context Extension

legal-ml-datasets

A collection of datasets and tasks for legal machine learning

Language:PythonStargazers:291Issues:15Issues:0

self-speculative-decoding

Code associated with the paper **Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding**

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:122Issues:3Issues:17

dlcl

The implementation of "Learning Deep Transformer Models for Machine Translation"

Language:PythonLicense:NOASSERTIONStargazers:114Issues:6Issues:6

llmeval-1

中文大语言模型评测第一期

Pruning-LLMs

The framework to prune LLMs to any size and any config.

Language:PythonLicense:Apache-2.0Stargazers:94Issues:1Issues:1

FudanNLP_Begginer

这是复旦NLP实验室五个任务的代码仓库,每个代码以ipynb形式展开,可以通过colab在线运行代码

Language:Jupyter NotebookStargazers:20Issues:1Issues:1

hibrids_summ

Code for ACL 2022 paper "HIBRIDS: Attention with Hierarchical Biases for Structure-aware Long Document Summarization".

Language:PythonLicense:MITStargazers:12Issues:1Issues:3