Takuya Makino (tma15)

tma15

Geek Repo

Location:Kanagawa, Japan

Home Page:https://tma15.github.io

Github PK Tool:Github PK Tool

Takuya Makino's starred repositories

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

License:NOASSERTIONStargazers:26139Issues:0Issues:0

self_supervised

A Pytorch-Lightning implementation of self-supervised algorithms

Language:PythonLicense:MITStargazers:527Issues:0Issues:0

olm-training

Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.

Language:PythonLicense:Apache-2.0Stargazers:91Issues:0Issues:0

data-preparation

Code used for sourcing and cleaning the BigScience ROOTS corpus

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:294Issues:0Issues:0

olm-datasets

Pipeline for pulling and processing online language model pretraining data from the web

Language:PythonLicense:Apache-2.0Stargazers:170Issues:0Issues:0

cramming

Cramming the training of a (BERT-type) language model into limited compute.

Language:PythonLicense:MITStargazers:1278Issues:0Issues:0

PaLM-rlhf-pytorch

Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM

Language:PythonLicense:MITStargazers:7663Issues:0Issues:0

tiktoken

tiktoken is a fast BPE tokeniser for use with OpenAI's models.

Language:PythonLicense:MITStargazers:11485Issues:0Issues:0

pytorch_scatter

PyTorch Extension Library of Optimized Scatter Operations

Language:PythonLicense:MITStargazers:1524Issues:0Issues:0

pyserini

Pyserini is a Python toolkit for reproducible information retrieval research with sparse and dense representations.

Language:PythonLicense:Apache-2.0Stargazers:1590Issues:0Issues:0

TRIME

[EMNLP 2022] Training Language Models with Memory Augmentation https://arxiv.org/abs/2205.12674

Language:PythonStargazers:189Issues:0Issues:0
Language:Jupyter NotebookLicense:NOASSERTIONStargazers:16Issues:0Issues:0

EMAT

Efficient Memory-Augmented Transformers

Language:PythonStargazers:34Issues:0Issues:0

FasterTransformer

Transformer related optimization, including BERT, GPT

Language:C++License:Apache-2.0Stargazers:5715Issues:0Issues:0
Language:PythonLicense:MITStargazers:69Issues:0Issues:0
Language:PythonStargazers:31Issues:0Issues:0

torchprofile

A general and accurate MACs / FLOPs profiler for PyTorch models

Language:PythonLicense:MITStargazers:550Issues:0Issues:0

Transkimmer

Code for ACL2022 publication Transkimmer: Transformer Learns to Layer-wise Skim

Language:PythonStargazers:21Issues:0Issues:0

intel-extension-for-transformers

⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡

Language:PythonLicense:Apache-2.0Stargazers:2089Issues:0Issues:0

dataloader

The merlin dataloader lets you rapidly load tabular data for training deep leaning models with TensorFlow, PyTorch or JAX

Language:PythonLicense:Apache-2.0Stargazers:401Issues:0Issues:0

length-adaptive-transformer

Official Pytorch Implementation of Length-Adaptive Transformer (ACL 2021)

Language:PythonLicense:Apache-2.0Stargazers:100Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:44Issues:0Issues:0

SentAugment

SentAugment is a data augmentation technique for NLP that retrieves similar sentences from a large bank of sentences. It can be used in combination with self-training and knowledge-distillation, or for retrieving paraphrases.

Language:PythonLicense:NOASSERTIONStargazers:363Issues:0Issues:0

voltaML

⚡VoltaML is a lightweight library to convert and run your ML/DL deep learning models in high performance inference runtimes like TensorRT, TorchScript, ONNX and TVM.

Language:PythonLicense:Apache-2.0Stargazers:1194Issues:0Issues:0

kernl

Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1506Issues:0Issues:0

model-soups

Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time

Language:PythonLicense:MITStargazers:399Issues:0Issues:0

transformers-bloom-inference

Fast Inference Solutions for BLOOM

Language:PythonLicense:Apache-2.0Stargazers:557Issues:0Issues:0

DeepSpeed-MII

MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.

Language:PythonLicense:Apache-2.0Stargazers:1817Issues:0Issues:0

AITemplate

AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.

Language:PythonLicense:Apache-2.0Stargazers:4505Issues:0Issues:0

pytorch-lightning

Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.

Language:PythonLicense:Apache-2.0Stargazers:27746Issues:0Issues:0