There are 25 repositories under transformers topic.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
🧑🏫 50! Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
CVPR 2022 论文和开源项目合集
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
:mag: Haystack is an open source NLP framework that leverages Transformer models. It enables developers to implement production-ready neural search, question answering, semantic document search and summarization for a wide range of applications.
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
Simple command line tool for text to image generation using OpenAI's CLIP and Siren (Implicit neural representation network). Technique was originally created by https://twitter.com/advadnoun
Easy-to-use and powerful NLP library with Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including Neural Search, Question Answering, Information Extraction and Sentiment Analysis end-to-end system.
A PyTorch-based Speech Toolkit
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
Transformers for Classification, NER, QA, Language Modelling, Language Generation, T5, Multi-Modal, and Conversational AI
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
State of the Art Natural Language Processing
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Leveraging BERT and c-TF-IDF to create easily interpretable topics.
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
TransmogrifAI (pronounced trăns-mŏgˈrə-fī) is an AutoML library for building modular, reusable, strongly typed machine learning workflows on Apache Spark with minimal hand-tuning
Reformer, the efficient Transformer, in Pytorch
A simple but complete full-attention transformer with a set of promising experimental features from various papers
jiant is an nlp toolkit
This repository contains demos I made with the Transformers library by HuggingFace.
MLeap: Deploy ML Pipelines to Production
Open-source offline translation library written in Python
Generative Adversarial Transformers
Over TypeScript tool to use custom transformers in the tsconfig.json
Scenic: A Jax Library for Computer Vision Research and Beyond
This Word Does Not Exist
🔥🔥🔥🔥 YOLO with Transformers and Instance Segmentation, with TensorRT acceleration! 🔥🔥🔥
Text-to-Image generation. The repo for NeurIPS 2021 paper "CogView: Mastering Text-to-Image Generation via Transformers".
Korean BERT pre-trained cased (KoBERT)
An implementation of Performer, a linear attention-based transformer, in Pytorch
Implementation of various self-attention mechanisms focused on computer vision. Ongoing repository.