There are 122 repositories under transformers topic.
21 Lessons, Get Started Building with Generative AI
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
CVPR 2025 论文和开源项目合集
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Private AI platform for agents, assistants and enterprise search. Built-in Agent Builder, Deep research, Document analysis, Multi-model support, and API connectivity for agents.
Machine Learning Engineering Open Book
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
Ongoing research training transformer models at scale
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
💡 All-in-one open-source AI framework for semantic search, LLM orchestration and language model workflows
This repository contains demos I made with the Transformers library by HuggingFace.
Semantic segmentation models with 500+ pretrained convolutional and transformer-based backbones.
A PyTorch-based Speech Toolkit
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
Open source real-time translation app for Android that runs locally
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, vLLM, DeepSpeed, Axolotl, etc.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
A concise but complete full-attention transformer with a set of promising experimental features from various papers
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Superduper: End-to-end framework for building custom AI applications and agents.
Robust recipes to align language models with human and AI preferences
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites