Vance0124 / Token-level-Direct-Preference-Optimization

Reference implementation for Token-level Direct Preference Optimization(TDPO)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Vance0124/Token-level-Direct-Preference-Optimization Stargazers