kingfan1998's starred repositories
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Document-Plugin
Plug-and-Play Document Modules for Pre-trained Models
tree-of-thought-llm
[NeurIPS 2023] Tree of Thoughts: Deliberate Problem Solving with Large Language Models
CSPostgraduate-408
💯 CSPostgraduate 计算机考研 408 专业课资料及真题资源
AlignScore
ACL2023 - AlignScore, a metric for factual consistency evaluation.
text-to-text-transfer-transformer
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
sentencepiece
Unsupervised text tokenizer for Neural Network-based text generation.
Distinct-N
Compute Distinct-N metric proposed by Jiwei Li et al.
ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
SeqDiffuSeq
Text Diffusion Model with Encoder-Decoder Transformers for Sequence-to-Sequence Generation [NAACL 2024]
ChatGLM-Finetuning
基于ChatGLM-6B、ChatGLM2-6B、ChatGLM3-6B模型,进行下游具体任务微调,涉及Freeze、Lora、P-tuning、全参微调等
awesome-pretrained-chinese-nlp-models
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
MT5ForGeneration
基于精简mt5预训练模型的seq2seq结构的实现
MT5_chinese_simplify
pytorch版本MT5模型的中文精简代码
turkish-question-generation
Automated question generation and question answering from Turkish texts using text-to-text transformers
pytorch_med_T5-large_scale_pretraining_and_fientune-
基于T5 和 mt5 模型的医学nlp大规模预训练模型的训练和验证,测试