DachengLi1 / LongChat

Official repository for LongChat and LongEval

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

How to prepare the training data

ycsun1972 opened this issue · comments

Hi,
"We fine-tune the 7B and 13B models with 80k and 18k conversations, respectively."
Could you provide more details about the training data? How the 80k data are prepared? Are they all with length of 16k?

Is the data used for training longchat-v1.5 the same as previous version?

Same question about longchat-v1.5. Cannot find any details about the longchat-v1.5.

@Mooler0410 oh it is the same, we just use the same data, but based on llama2