wuch15 / MT-BERT

One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

MT-BERT

One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers

About

One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers