Stephen Fernandes's repositories
attention_sinks
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
Awesome-Diffusion-Models
A collection of resources and papers on Diffusion Models and Score-based Models, a darkhorse in the field of Generative Models
conformer
PyTorch implementation of "Conformer: Convolution-augmented Transformer for Speech Recognition" (INTERSPEECH 2020)
deepspeed-test
Testing the DeepSpeed integration in Trainer and Accelerate
dora-from-scratch
LoRA and DoRA from Scratch Implementations
FastAPI-template
Feature rich robust FastAPI template.
instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
instruction-datasets
All available datasets for Instruction Tuning of Large Language Models
Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
mesh-transformer-jax
Model parallel transformers in JAX and Haiku
min-decision-transformer
Minimal implementation of Decision Transformer: Reinforcement Learning via Sequence Modeling in PyTorch for mujoco control tasks in OpenAI gym
Modified-DETR
The PyTorch re-implement of the official DETR.
NanoPeft
The simplest repository & Neat implementation of different Lora methods for training/fine-tuning Transformer-based models (i.e., BERT, GPTs). [ Research purpose ]
openai_trtllm
OpenAI compatible API for TensorRT LLM triton backend
phi-1
Plug in and play implementation of " Textbooks Are All You Need", ready for training, inference, and dataset generation
Promptify
Prompt Engineering | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engineering, LLMs and other latest research
python-bpe
Byte Pair Encoding for Python!
PyTorch-Elmo-BiLSTMCRF
PyTorch BiLSTMCRF w Elmo
RealtimeTTS
Converts text to speech in realtime
system-design-primer
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
Transformers-Tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
tree-of-thought-llm
[NeurIPS 2023] Tree of Thoughts: Deliberate Problem Solving with Large Language Models
unify-learning-paradigms
data collator for UL2 and U-PaLM, suprising no one wrote or tried this open source