openai / gpt-2

Code for the paper "Language Models are Unsupervised Multitask Learners"

Home Page:https://openai.com/blog/better-language-models/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Dose the pre-training data also use this prompt structure related to downstream tasks?

Aurora-slz opened this issue · comments

I read the gpt2 paper, but not sure whether the pre-training data from WebText will add format information.
For example, we konw data format will be english sentence = french sentencein the translation task. So during pre-training time, will we add similar promt to the training data?

Thanks!

interested about this