mathon's starred repositories
pytorch-lightning
Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.
immersive-translate
沉浸式双语网页翻译扩展 , 支持输入框翻译, 鼠标悬停翻译, PDF, Epub, 字幕文件, TXT 文件翻译 - Immersive Dual Web Page Translation Extension
flash-attention
Fast and memory-efficient exact attention
Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
Megatron-LM
Ongoing research training transformer models at scale
streaming-llm
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
DeepSeek-Coder
DeepSeek Coder: Let the Code Write Itself
GPU-Puzzles
Solve puzzles. Learn CUDA.
QuantsPlaybook
量化研究-券商金工研报复现
FinRL-Trading
For trading. Please star.
rq-vae-transformer
The official implementation of Autoregressive Image Generation using Residual Quantization (CVPR '22)
M5-methods
Data, Benchmarks, and methods submitted to the M5 forecasting competition
hive-third-functions
Some useful custom hive udf functions, especial array, json, math, string functions.
Reinforcement-Learning-for-Market-Making
Using tabular and deep reinforcement learning methods to infer optimal market making strategies
DisCo-CLIP
Official PyTorch implementation of the paper "DisCo-CLIP: A Distributed Contrastive Loss for Memory Efficient CLIP Training".