Phil Wang (lucidrains)

lucidrains

Geek Repo

Location:San Francisco

Home Page:lucidrains.github.io

Twitter:@lucidrains

Github PK Tool:Github PK Tool

Phil Wang's repositories

musiclm-pytorch

Implementation of MusicLM, Google's new SOTA model for music generation using attention networks, in Pytorch

Language:PythonLicense:MITStargazers:3008Issues:98Issues:52

reformer-pytorch

Reformer, the efficient Transformer, in Pytorch

Language:PythonLicense:MITStargazers:2053Issues:54Issues:120

make-a-video-pytorch

Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch

Language:PythonLicense:MITStargazers:1838Issues:72Issues:15

perceiver-pytorch

Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch

Language:PythonLicense:MITStargazers:1047Issues:31Issues:59

memorizing-transformers-pytorch

Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch

Language:PythonLicense:MITStargazers:608Issues:11Issues:13

MEGABYTE-pytorch

Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch

Language:PythonLicense:MITStargazers:590Issues:10Issues:13

siren-pytorch

Pytorch implementation of SIREN - Implicit Neural Representations with Periodic Activation Function

Language:PythonLicense:MITStargazers:452Issues:13Issues:6

memory-efficient-attention-pytorch

Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"

Language:PythonLicense:MITStargazers:338Issues:9Issues:5

conformer

Implementation of the convolutional module from the Conformer paper, for use in Transformers

Language:PythonLicense:MITStargazers:329Issues:9Issues:12

deformable-attention

Implementation of Deformable Attention in Pytorch from the paper "Vision Transformer with Deformable Attention"

Language:PythonLicense:MITStargazers:248Issues:9Issues:7

x-unet

Implementation of a U-net complete with efficient attention as well as the latest research findings

Language:PythonLicense:MITStargazers:248Issues:12Issues:7

electra-pytorch

A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch

Language:PythonLicense:MITStargazers:215Issues:9Issues:11

jax2torch

Use Jax functions in Pytorch

Language:PythonLicense:MITStargazers:209Issues:5Issues:3

block-recurrent-transformer-pytorch

Implementation of Block Recurrent Transformer - Pytorch

Language:PythonLicense:MITStargazers:204Issues:8Issues:6

Mega-pytorch

Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena

Language:PythonLicense:MITStargazers:201Issues:8Issues:2

mlm-pytorch

An implementation of masked language modeling for Pytorch, made as concise and simple as possible

Language:PythonLicense:MITStargazers:169Issues:5Issues:6

graph-transformer-pytorch

Implementation of Graph Transformer in Pytorch, for potential use in replicating Alphafold2

Language:PythonLicense:MITStargazers:163Issues:4Issues:2

ETSformer-pytorch

Implementation of ETSformer, state of the art time-series Transformer, in Pytorch

Language:PythonLicense:MITStargazers:141Issues:10Issues:10

mixture-of-attention

Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts

Language:PythonLicense:MITStargazers:93Issues:7Issues:0

discrete-key-value-bottleneck-pytorch

Implementation of Discrete Key / Value Bottleneck, in Pytorch

Language:PythonLicense:MITStargazers:87Issues:4Issues:2

TPDNE

Thispersondoesnotexist went down, so this time, while building it back up, I am going to open source all of it.

Language:PythonLicense:MITStargazers:82Issues:4Issues:0

VN-transformer

A Transformer made of Rotation-equivariant Attention using Vector Neurons

Language:PythonLicense:MITStargazers:73Issues:7Issues:2

rvq-vae-gpt

My attempts at applying Soundstream design on learned tokenization of text and then applying hierarchical attention to text generation

Language:PythonLicense:MITStargazers:72Issues:5Issues:0

product-key-memory

Standalone Product Key Memory module in Pytorch - for augmenting Transformer models

Language:PythonLicense:MITStargazers:67Issues:3Issues:1

flash-genomics-model

My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other hierarchical methods)

Language:PythonLicense:MITStargazers:52Issues:6Issues:3

coordinate-descent-attention

Implementation of an Attention layer where each head can attend to more than just one token, using coordinate descent to pick topk

Language:PythonLicense:MITStargazers:43Issues:3Issues:1

autoregressive-linear-attention-cuda

CUDA implementation of autoregressive linear attention, with all the latest research findings

Language:PythonLicense:MITStargazers:42Issues:4Issues:0

nim-tokenizer

Implementation of a simple BPE tokenizer, but in Nim

Language:NimLicense:MITStargazers:20Issues:3Issues:0

hyena-dna

Fork of HyenaDNA, a long-range genomic foundation model built with Hyena

Language:AssemblyLicense:Apache-2.0Stargazers:9Issues:2Issues:0

nitter

Alternative Twitter front-end

License:AGPL-3.0Stargazers:4Issues:0Issues:0