Doohae Jung (wavy.ocean) (Doohae)

Doohae

Geek Repo

Company:@kakaobrain

Location:Seoul, Korea

Github PK Tool:Github PK Tool

Doohae Jung (wavy.ocean)'s repositories

DenseRetrieval

Implementation of DPR, GradCache, DSI etc.

Language:PythonStargazers:1Issues:1Issues:0

accelerate

🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeedExamples

Example models using DeepSpeed

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

GradCache

Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraint

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

RETRO-pytorch

Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Building-Python-Web-APIs-with-FastAPI

Building Python Web APIs with FastAPI, published by Packt

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Chatbot_data

Chatbot_data_for_Korean

License:MITStargazers:0Issues:0Issues:0
Stargazers:0Issues:1Issues:0

course

The Hugging Face course

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Stargazers:0Issues:0Issues:0
Language:Jupyter NotebookStargazers:0Issues:0Issues:0

fairseq

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

FiD

Fusion-in-Decoder

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

kolmev

Evaluation for korean language models (e.g. bert, roberta, bart, t5, gpt2...)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

lassl

Easy Language Model Pretraining leveraging Huggingface's Transformers and Datasets

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llmss

LLM simple serving (tensor model parallel, pubsub, grpc)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of autoregressive language models.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

marcopolo

Marco Emilio Polo (/ˈmɑːrkoʊ ˈpoʊloʊ/ (listen), Venetian: [ˈmaɾko ˈpolo], Italian: [ˈmarko ˈpɔːlo] (listen); c. 1254 – January 8, 1324)[1] was a Venetian merchant,[2][3] explorer, and writer from the Republic of Venice who travelled through Asia along the Silk Road between 1271 and 1295.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0
Language:Jupyter NotebookStargazers:0Issues:1Issues:0

pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

Language:C++License:NOASSERTIONStargazers:0Issues:0Issues:0
Stargazers:0Issues:1Issues:0
Language:PythonStargazers:0Issues:1Issues:0

tppys

Text processing by pyspark (just sample project)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

tpu-starter

Everything you want to know about Google Cloud TPU

Language:PythonLicense:CC0-1.0Stargazers:0Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

YaLM-100B

Pretrained language model with 100B parameters

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0