Jatin Prakash's starred repositories
llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
Teacher-free-Knowledge-Distillation
Knowledge Distillation: CVPR2020 Oral, Revisiting Knowledge Distillation via Label Smoothing Regularization
Nasty-Teacher
[ICLR 2021 Spotlight Oral] "Undistillable: Making A Nasty Teacher That CANNOT teach students", Haoyu Ma, Tianlong Chen, Ting-Kuei Hu, Chenyu You, Xiaohui Xie, Zhangyang Wang
llm-workflow-engine
Power CLI and Workflow manager for LLMs (core package)
x-transformers
A concise but complete full-attention transformer with a set of promising experimental features from various papers
Dreambooth-Stable-Diffusion
Implementation of Dreambooth (https://arxiv.org/abs/2208.12242) with Stable Diffusion
companies-sponsoring-visas
A list of companies that sponsor employees from other countries.
vscode-dev-containers
NOTE: Most of the contents of this repository have been migrated to the new devcontainers GitHub org (https://github.com/devcontainers). See https://github.com/devcontainers/template-starter and https://github.com/devcontainers/feature-starter for information on creating your own!
torchdistill
A coding-free framework built on PyTorch for reproducible deep learning studies. 🏆25 knowledge distillation methods presented at CVPR, ICLR, ECCV, NeurIPS, ICCV, etc are implemented so far. 🎁 Trained models, training logs and configurations are available for ensuring the reproducibiliy and benchmark.
python-sortedcontainers
Python Sorted Container Types: Sorted List, Sorted Dict, and Sorted Set
Knowledge-Distillation-Zoo
Pytorch implementation of various Knowledge Distillation (KD) methods.
imagen-pytorch
Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch
torchtyping
Type annotations and dynamic checking for a tensor's shape, dtype, names, etc.
knowledge-distillation-pytorch
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
MDCA-Calibration
[CVPR 2022] Official code for the paper: "A Stitch in Time Saves Nine: A Train-Time Regularizing Loss for Improved Neural Network Calibration"