yxchng's starred repositories
LinearAttentionArena
Here we will test various linear attention designs.
objaverse-xl
🪐 Objaverse-XL is a Universe of 10M+ 3D Objects. Contains API Scripts for Downloading and Processing!
DeepSeek-VL
DeepSeek-VL: Towards Real-World Vision-Language Understanding
Awesome_Mamba
Computation-Efficient Era: A Comprehensive Survey of State Space Models in Medical Image Analysis
flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
inceptionnext
InceptionNeXt: When Inception Meets ConvNeXt (CVPR 2024)
EfficientVMamba
Code Implementation of EfficientVMamba
imagenet_d
[CVPR2024 Highlight] Official Code for "ImageNet-D: Benchmarking Neural Network Robustness on Diffusion Synthetic Object"
LLM-Inheritune
This is the official repository for Inheritune.
VideoMamba
[ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding
Diffusion-RWKV
Scaling RWKV-Like Architectures for Diffusion Models
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
frequency_determines_performance
Code for the paper: "No Zero-Shot Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance"