reborm's repositories
agentx
AgentX is an experiment to develop an autonomous agent that delegates well to Auto-GPT, babyagi, and other agents using LangChain
alpaca_chinese_dataset
人工精调的中文对话数据集和一段chatglm的微调代码
Auto-GPT
An experimental open-source attempt to make GPT-4 fully autonomous.
BillyGPT
可自动优化提示、免费开源、全平台傻瓜式 ChatGPT 本地客户端,支持断点续聊、修改历史对话、本地聊天记录存储导入导出、添加自己的 apikey
ChatGLM-Finetuning
基于ChatGLM-6B模型,进行下游具体任务微调,涉及Freeze、Lora、P-tuning等
ChatGLM-LLaMA-chinese-insturct
探索中文instruct数据在ChatGLM, LLaMA上的微调表现
ChatGLM-LoRA-RLHF-PyTorch
A full pipeline to finetune ChatGLM LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the ChatGLM architecture. Basically ChatGPT but with ChatGLM
Chatglm_lora_multi-gpu
chatglm多gpu用deepspeed和
ChatGLM_LoRA_zh
在ChatGLM大模型上利用LoRA方法进行小参数学习,训练语料库选择中文的[alpaca-zh](https://huggingface.co/datasets/shibing624/alpaca-zh)
chatgpt-llamaindex-demo
ChatGPT and LlamaIndex demo
Chinese-LangChain
中文langchain项目|小必应,Q.Talk,强聊,QiangTalk
ChineseNLPCorpus
中文自然语言处理数据集,平时做做实验的材料。欢迎补充提交合并。
ColossalAI
Making large AI models cheaper, faster and more accessible
DeepSpeed-Chat-ChatGLM
包含了RLHF
FinGLM
FinGLM: 致力于构建一个开放的、公益的、持久的金融大模型项目,利用开源开放来促进「AI+金融」。
hcgf
Humanable ChatGLM/GPT Fine-tuning | ChatGLM微调
InstructGLM
ChatGLM-6B 指令学习|指令数据|Instruct
japanese-alpaca-lora
A japanese finetuned instruction LLaMA
llama-trl
LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA
Llama-X
Open Academic Research on Improving LLaMA to SOTA LLM
LLMs_interview_notes
该仓库主要记录 大模型(LLMs) 算法工程师相关的面试题
minChatGPT
A minimum example of aligning language models with RLHF similar to ChatGPT
promptlib
A collection of prompts for use with GPT-4 via ChatGPT, OpenAI API w/ Gradio frontend & notebook
RLHF
Implementation of Chinese ChatGPT
text-to-sql-wizardcoder
Leveraging large language models for text-to-SQL synthesis, this project fine-tunes WizardLM/WizardCoder-15B-V1.0 with QLoRA on a custom Spider training dataset. The resultant model, achieves 61% execution accuracy, incorporating database context for validation.
textgen
textgen, Text Generation models. 文本生成,包括:LLAMA、ChatGLM、UDA,GPT2,Seq2Seq,BART,T5等模型实现,开箱即用。
transformers_tasks
⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
unstructured
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
Vicuna-LoRA-RLHF-PyTorch
A full pipeline to finetune Vicuna LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the Vicuna architecture. Basically ChatGPT but with Vicuna