There are 24 repositories under attention topic.
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
Natural Language Processing Tutorial for Deep Learning Researchers
🍀 Pytorch implementation of various Attention Mechanisms, MLP, Re-parameter, Convolution, which is helpful to further understand papers.⭐⭐⭐
A PyTorch implementation of the Transformer model in "Attention is All You Need".
Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.
该仓库主要记录 NLP 算法工程师相关的顶会论文研读笔记
Scenic: A Jax Library for Computer Vision Research and Beyond
Draw a leader line in your web page.
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Transformer: PyTorch Implementation of "Attention Is All You Need"
Aspect Based Sentiment Analysis, PyTorch Implementations. 基于方面的情感分析,使用PyTorch实现。
Gathers machine learning and Tensorflow deep learning models for NLP problems, 1.13 < Tensorflow < 2.0
This is a pytorch repository of YOLOv4, attentive YOLOv4 and mobilenet YOLOv4 with PASCAL VOC and COCO
Documents, papers and codes related to Natural Language Processing, including Topic Model, Word Embedding, Named Entity Recognition, Text Classificatin, Text Generation, Text Similarity, Machine Translation),etc. All codes are implemented intensorflow 2.0.
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute
Improving Convolutional Networks via Attention Transfer (ICLR 2017)
Generative Adversarial Transformers
Replication of simple CV Projects including attention, classification, detection, keypoint detection, etc.
Implementation of various self-attention mechanisms focused on computer vision. Ongoing repository.
🚀🚀🚀 A collection of some awesome public YOLO object detection series projects.
An implementation of Performer, a linear attention-based transformer, in Pytorch
list of efficient attention modules
A Tensorflow implementation of Spatial Transformer Networks.
Code for our CVPR2021 paper coordinate attention
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Simple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Voice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Graph Transformer Architecture. Source code for "A Generalization of Transformer Networks to Graphs", DLG-AAAI'21.
Official Pytorch Code for "Medical Transformer: Gated Axial-Attention for Medical Image Segmentation" - MICCAI 2021
A PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.
Tensorflow implementation of attention mechanism for text classification tasks.
Implementation of papers for text classification task on DBpedia
中文实体关系抽取,pytorch,bilstm+attention
Long Range Arena for Benchmarking Efficient Transformers