Hongsun_Jang (hongsunjang)

hongsunjang

Geek Repo

Company:@AIS_SNU, SNU ECE

Location:Seoul, Republic of Korea

Home Page:https://aisys.snu.ac.kr/hongsun.html

Github PK Tool:Github PK Tool

Hongsun_Jang's starred repositories

finetune-gpt2xl

Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed

Language:PythonLicense:MITStargazers:427Issues:0Issues:0

gpt-2

Code for the paper "Language Models are Unsupervised Multitask Learners"

Language:PythonLicense:NOASSERTIONStargazers:22146Issues:0Issues:0

metaseq

Repo for external large-scale work

Language:PythonLicense:MITStargazers:6439Issues:0Issues:0

Pretrained-Language-Model

Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

Language:PythonStargazers:3003Issues:0Issues:0

Ok-Topk

Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k communication volume which is asymptotically optimal) with the decentralized parallel Stochastic Gradient Descent (SGD) optimizer, and its convergence is proved theoretically and empirically.

Language:PythonLicense:GPL-3.0Stargazers:23Issues:0Issues:0

computer-science

:mortar_board: Path to a free self-taught education in Computer Science!

License:MITStargazers:167564Issues:0Issues:0

SimpleSSD-FullSystem

Open-Source Licensed Educational SSD Simulator for High-Performance Storage and Full-System Evaluations

Language:C++License:BSD-3-ClauseStargazers:87Issues:0Issues:0

torchrec

Pytorch domain library for recommendation systems

Language:PythonLicense:BSD-3-ClauseStargazers:1844Issues:0Issues:0
License:MITStargazers:365Issues:0Issues:0

SQuant

SQuant [ICLR22]

Language:PythonStargazers:158Issues:0Issues:0

lora

Using Low-rank adaptation to quickly fine-tune diffusion models.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:6904Issues:0Issues:0

LoRA

Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"

Language:PythonLicense:MITStargazers:10129Issues:0Issues:0

open-llms

πŸ“‹ A list of open LLMs available for commercial use.

License:Apache-2.0Stargazers:10804Issues:0Issues:0

improved-diffusion

Release for Improved Denoising Diffusion Probabilistic Models

Language:PythonLicense:MITStargazers:3100Issues:0Issues:0

ChatGPT-as-a-server

Using ChatGPT as a real backend

Language:GoStargazers:6Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:6195Issues:0Issues:0

KDiskMark

A simple open-source disk benchmark tool for Linux distros

Language:C++License:GPL-3.0Stargazers:1032Issues:0Issues:0

CrystalDiskInfo

CrystalDiskInfo

Language:C++License:MITStargazers:1572Issues:0Issues:0

ssd-benchmark-rs

Super Simple Disk Benchmark - benchmarks the writing performance of your disk

Language:RustLicense:GPL-3.0Stargazers:47Issues:0Issues:0
Language:PythonStargazers:13Issues:0Issues:0

Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Language:PythonLicense:NOASSERTIONStargazers:1801Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:34425Issues:0Issues:0
Language:VerilogLicense:GPL-3.0Stargazers:40Issues:0Issues:0

MQSim

MQSim is a fast and accurate simulator modeling the performance of modern multi-queue (MQ) SSDs as well as traditional SATA based SSDs. MQSim faithfully models new high-bandwidth protocol implementations, steady-state SSD conditions, and the full end-to-end latency of requests in modern SSDs. It is described in detail in the FAST 2018 paper by Arash Tavakkol et al., "MQSim: A Framework for Enabling Realistic Studies of Modern Multi-Queue SSD Devices" (https://people.inf.ethz.ch/omutlu/pub/MQSim-SSD-simulation-framework_fast18.pdf)

Language:C++License:MITStargazers:272Issues:0Issues:0

Vitis-Tutorials

Vitis In-Depth Tutorials

Language:CLicense:MITStargazers:1171Issues:0Issues:0
Language:C++Stargazers:35Issues:0Issues:0

GCoD

[HPCA 2022] GCoD: Graph Convolutional Network Acceleration via Dedicated Algorithm and Accelerator Co-Design

Language:PythonLicense:Apache-2.0Stargazers:32Issues:0Issues:0

Literatures-on-GNN-Acceleration

A reading list for deep graph learning acceleration.

License:MITStargazers:214Issues:0Issues:0

GenStore

GenStore is the first in-storage processing system designed for genome sequence analysis that greatly reduces both data movement and computational overheads of genome sequence analysis by exploiting low-cost and accurate in-storage filters. Described in the ASPLOS 2022 paper by Mansouri Ghiasi et al. at https://people.inf.ethz.ch/omutlu/pub/GenStore_asplos22-arxiv.pdf

Language:CLicense:MITStargazers:12Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:55178Issues:0Issues:0