There are 0 repository under multi-head-self-attention topic.
Attention temporal convolutional network for EEG-based motor imagery classification
A Faster Pytorch Implementation of Multi-Head Self-Attention
BabyGPT: Build Your Own GPT Large Language Model from Scratch Pre-Training Generative Transformer Models: Building GPT from Scratch with a Step-by-Step Guide to Generative AI in PyTorch and Python
Transformer/Transformer-XL/R-Transformer examples and explanations
Transformer creation from scratch using Jax.