There are 30 repositories under attention topic.
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
Natural Language Processing Tutorial for Deep Learning Researchers
🍀 Pytorch implementation of various Attention Mechanisms, MLP, Re-parameter, Convolution, which is helpful to further understand papers.⭐⭐⭐
A PyTorch implementation of the Transformer model in "Attention is All You Need".
Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.
该仓库主要记录 NLP 算法工程师相关的顶会论文研读笔记
Scenic: A Jax Library for Computer Vision Research and Beyond
Transformer: PyTorch Implementation of "Attention Is All You Need"
Draw a leader line in your web page.
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Quantized Attention achieves speedup of 2-5x and 3-11x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and video models.
Aspect Based Sentiment Analysis, PyTorch Implementations. 基于方面的情感分析,使用PyTorch实现。
Gathers machine learning and Tensorflow deep learning models for NLP problems, 1.13 < Tensorflow < 2.0
This is a pytorch repository of YOLOv4, attentive YOLOv4 and mobilenet YOLOv4 with PASCAL VOC and COCO
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute
Improving Convolutional Networks via Attention Transfer (ICLR 2017)
Generative Adversarial Transformers
Replication of simple CV Projects including attention, classification, detection, keypoint detection, etc.
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
Implementation of various self-attention mechanisms focused on computer vision. Ongoing repository.
An implementation of Performer, a linear attention-based transformer, in Pytorch
Code for our CVPR2021 paper coordinate attention
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
A Tensorflow implementation of Spatial Transformer Networks.
list of efficient attention modules
Simple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Graph Transformer Architecture. Source code for "A Generalization of Transformer Networks to Graphs", DLG-AAAI'21.
Voice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Official Pytorch Code for "Medical Transformer: Gated Axial-Attention for Medical Image Segmentation" - MICCAI 2021
多标签文本分类,多标签分类,文本分类, multi-label, classifier, text classification, BERT, seq2seq,attention, multi-label-classification
A PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.
中文实体关系抽取,pytorch,bilstm+attention
Long Range Arena for Benchmarking Efficient Transformers
Tensorflow implementation of attention mechanism for text classification tasks.
Implementation of papers for text classification task on DBpedia