There are 38 repositories under gpt-2 topic.
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
This repository contains demos I made with the Transformers library by HuggingFace.
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
Chinese version of GPT2 training code, using BERT tokenizer.
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI**)
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project: http://casl-project.ai/
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Simple UI for LLM Model Finetuning
A Large-scale Chinese Short-Text Conversation Dataset and Chinese pre-training dialog models
Guide to using pre-trained large language models of source code
🦄 State-of-the-Art Conversational AI with Transfer Learning
GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
llama and other large language models on iOS and MacOS offline using GGML library.
Visual Studio Code client for Tabnine. https://marketplace.visualstudio.com/items?itemName=TabNine.tabnine-vscode
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
This Discord chatbot is incredibly versatile. Powered incredibly fast Groq API
[CVPR 2025 Oral]Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
This series will take you on a journey from the fundamentals of NLP and Computer Vision to the cutting edge of Vision-Language Models.
A curated list of NLP resources focused on Transformer networks, attention mechanism, GPT, BERT, ChatGPT, LLMs, and transfer learning.
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
This Word Does Not Exist
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
This is a repository that aims to provide updates on the status of jailbreaking the OpenAI GPT language model.
Integrating the Best of TF into PyTorch, for Machine Learning, Natural Language Processing, and Text Generation. This is part of the CASL project: http://casl-project.ai/
Vim client for TabNine. https://vimawesome.com/plugin/tabnine-vim
Medical Q&A with Deep Language Models
Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-176B/bloom/gpt/bart/T5/MetaICL)