Reacubeth's starred repositories
Open-Assistant
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
HowToLiveLonger
程序员延寿指南 | A programmer's guide to live longer
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
LLaMA-Factory
Unify Efficient Fine-Tuning of 100+ LLMs
alpaca-lora
Instruct-tune LLaMA on consumer hardware
Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
Megatron-LM
Ongoing research training transformer models at scale
alignment-handbook
Robust recipes to align language models with human and AI preferences
Alpaca-CoT
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
NLP-Interview-Notes
该仓库主要记录 NLP 算法工程师相关的面试题
LLMDataHub
A quick guide (especially) for trending instruction finetuning datasets
Orion
Orion-14B is a family of models includes a 14B foundation LLM, and a series of models: a chat model, a long context model, a quantized model, a RAG fine-tuned model, and an Agent fine-tuned model. Orion-14B 系列模型包括一个具有140亿参数的多语言基座大模型以及一系列相关的衍生模型,包括对话模型,长文本模型,量化模型,RAG微调模型,Agent微调模型等。
representation-engineering
Representation Engineering: A Top-Down Approach to AI Transparency
EvaluationPapers4ChatGPT
Resource, Evaluation and Detection Papers for ChatGPT
Awesome-Scientific-Language-Models
A Curated List of Language Models in Scientific Domains
AutoCompressors
[EMNLP 2023] Adapting Language Models to Compress Long Contexts
nosync-icloud
避免 iCloud 同步 node_modules(Avoid node_modules to sync with iCloud)
geogalactica
Code and datasets for paper "GeoGalactica: A Scientific Large Language Model in Geoscience"
GPT2-Knowledge-Distillation
Knowledge Distillation for student model of GPT from GPT-medium on tiny Shakespeare dataset