Kim, Taehoon's starred repositories
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
flash-attention
Fast and memory-efficient exact attention
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
P-tuning-v2
An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks
lang-segment-anything
SAM with text prompt
pytorch_wavelets
Pytorch implementation of 2D Discrete Wavelet (DWT) and Dual Tree Complex Wavelet Transforms (DTCWT) and a DTCWT based ScatterNet
LovelyPlots
Matplotlib style sheets to nicely format figures for scientific papers, thesis and presentations while keeping them fully editable in Adobe Illustrator.
minDiffusion
Self-contained, minimalistic implementation of diffusion models with Pytorch.
cc2dataset
Easily convert common crawl to a dataset of caption and document. Image/text Audio/text Video/text, ...
py-bottom-up-attention
PyTorch bottom-up attention with Detectron2
compression-DCT
Implementation of Image compression using DCT
EXAONE-Atelier
Jupyter notebook examples for EXAONE Atelier in AWS Marketplace
Eval-Service
MoverScore: Text Generation Evaluation with Contextualized Embedding and Earth Mover Distance