okoge-kaz / llm-jp-dpo

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LLM-jp DPO (Direct Preference Optimization)

This repository contains the code for DPO of LLM-jp models.

Requirements

See pyproject.toml for the required packages.

Installation

poetry install
poetry shell

Training

Here is the command to train a model using 8 GPUs.

CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 accelerate launch --config_file accelerate_configs/zero2.yaml train.py

About

License:Apache License 2.0


Languages

Language:Python 100.0%