pauldanielconway's repositories
accelerate
π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Awesome-Code-LLM
π¨βπ» An awesome and curated list of best code-LLM for research.
Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
datasets
π€ The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
Deep-Live-Cam
real time face swap and one-click video deepfake with only a single image
DeepSeek-Coder-V2
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
ERNIE
Official implementations for various pre-training models of ERNIE-family, covering topics of Language Understanding & Generation, Multimodal Understanding & Generation, and beyond.
evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
flash-attention
Fast and memory-efficient exact attention
google-research
Google Research
gpt-2-simple
Python package to easily retrain OpenAI's GPT-2 text-generating model on new texts
GPT-SoVITS
1 min voice data can also be used to train a good TTS model! (few shot voice cloning)
lm-evaluation-harness
A framework for few-shot evaluation of language models.
MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
mt-dnn
Multi-Task Deep Neural Networks for Natural Language Understanding
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
peft
π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Qwen2.5-Coder
Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.
scikit-learn
scikit-learn: machine learning in Python
tensor2tensor
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
tqdm
:zap: A Fast, Extensible Progress Bar for Python and CLI
transformers
π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Transformers-for-NLP-2nd-Edition
Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL-E including jump starting GPT-4, speech-to-text, text-to-speech, text to image generation with DALL-E, Google Cloud AI,HuggingGPT, and more
trax
Trax β Deep Learning with Clear Code and Speed
triton
Development repository for the Triton language and compiler
trl
Train transformer language models with reinforcement learning.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs