Taishi Nakamura's repositories
llama-recipes
Examples and recipes for Llama 2 model
multimodal
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
alignment-handbook
Robust recipes to align language models with human and AI preferences
bigcode-evaluation-harness
A framework for the evaluation of autoregressive code generation language models.
FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
FIN-bench
Evaluation of Finnish generative models
Megatron-LLM
distributed trainer for LLMs
Megatron-LM
Ongoing research training transformer models at scale
LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
llm-leaderboard
Project of llm evaluation to Japanese tasks
long-context
YaRN: Efficient Context Window Extension of Large Language Models
Megatron-LM-LUMI
Ongoing research training transformer models at scale
mlmm-evaluation
Multilingual Large Language Models Evaluation Benchmark
moe-recipes
Mixtre of Experts Library forked from kotoba-recipes
nccl-tests
NCCL Tests
rome
Locating and editing factual associations in GPT (NeurIPS 2022)
SEED
Empowers LLMs with the ability to see and draw.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.