There are 38 repositories under gpt-2 topic.
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
This repository contains demos I made with the Transformers library by HuggingFace.
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
Chinese version of GPT2 training code, using BERT tokenizer.
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI**)
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project: http://casl-project.ai/
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Simple UI for LLM Model Finetuning
llama and other large language models on iOS and MacOS offline using GGML library.
A Large-scale Chinese Short-Text Conversation Dataset and Chinese pre-training dialog models
Guide to using pre-trained large language models of source code
🦄 State-of-the-Art Conversational AI with Transfer Learning
GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
[CVPR 2025 Oral]Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
Visual Studio Code client for Tabnine. https://marketplace.visualstudio.com/items?itemName=TabNine.tabnine-vscode
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
This Discord chatbot is incredibly versatile. Powered incredibly fast Groq API
This series will take you on a journey from the fundamentals of NLP and Computer Vision to the cutting edge of Vision-Language Models.
A curated list of NLP resources focused on Transformer networks, attention mechanism, GPT, BERT, ChatGPT, LLMs, and transfer learning.
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
This Word Does Not Exist
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
This is a repository that aims to provide updates on the status of jailbreaking the OpenAI GPT language model.
Integrating the Best of TF into PyTorch, for Machine Learning, Natural Language Processing, and Text Generation. This is part of the CASL project: http://casl-project.ai/
Vim client for TabNine. https://vimawesome.com/plugin/tabnine-vim
The fastest JavaScript BPE Tokenizer Encoder Decoder for OpenAI's GPT models (gpt-5, gpt-o*, gpt-4o, etc.). Port of OpenAI's tiktoken with additional features.
Medical Q&A with Deep Language Models