Junseo Park's starred repositories
FlameGraph
Stack trace visualizer
TensorRT-Model-Optimizer
TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillation, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
applied-ml
📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.
awesome-knowledge-distillation
Awesome Knowledge Distillation
jetson-containers
Machine Learning Containers for NVIDIA Jetson and JetPack-L4T
KakaoPostCodeWeb
KakaoPostCodeWeb 호스팅용
tensorrtllm_backend
The Triton TensorRT-LLM Backend
2023-MatKor-Rust-Interpreter
2023년 고려대학교 MatKor 스터디 - Rust 기초 프로그래밍 + 인터프리터 만들기
Tensorrt-Deformable-Detr
Tensorrt-Deformable-Detr
Awesome-Pruning
A curated list of neural network pruning resources.
poly-match
Source for the "Making Python 100x faster with less than 100 lines of Rust" blog post
awesome-actions
A curated list of awesome actions to use on GitHub