laocheche's starred repositories
transformers_tasks
⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
uie_pytorch
PaddleNLP UIE模型的PyTorch版实现
FinanceChatGLM
SMP 2023 ChatGLM金融大模型挑战赛 60 分baseline思路介绍
Chinese-Vicuna
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
chatgpt_all
学习开源chatGPT类模型的指南,汇总各种训练数据获取、模型微调、模型服务的方法,以及记录自己操作总遇到的各种常见坑,欢迎收藏、转发,希望能帮你省一些时间
direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
baichuan-speedup
纯c++的全平台llm加速库,支持python调用,支持baichuan, glm, llama, moss基座,手机端流畅运行chatglm-6B级模型单卡可达10000+token / s,
Baichuan-Chat-Tuning
实现了Baichuan-Chat微调,Lora、QLora等各种微调方式,一键运行。
ChatGLM2-6B-Explained
ChatGLM2-6B-Explained
GLM-Explained
GLM-Explained
pytorch-explained
pytorch-explained annatated
tranformers-expalined
tranformers-code-explained--line-by-line
pdf_to_txt
change pdf to txt
contriever
Contriever: Unsupervised Dense Information Retrieval with Contrastive Learning
Fengshenbang-LM
Fengshenbang-LM(封神榜大模型)是IDEA研究院认知计算与自然语言研究中心主导的大模型开源体系,成为中文AIGC和认知智能的基础设施。
ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
alpaca-lora
Instruct-tune LLaMA on consumer hardware
self-instruct
Aligning pretrained language models with instruction data generated by themselves.
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
edgar-crawler
The only open-source toolkit that can download EDGAR financial reports and extract textual data from specific item sections into nice and clean JSON files.