AristotelisPap / Question-Answering-with-BERT-and-Knowledge-Distillation

Fine-tuned BERT on SQuAd 2.0 Dataset. Applied Knowledge Distillation (KD) and fine-tuned DistilBERT (student) using BERT as the teacher model. Reduced the size of the original BERT by 40%.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

AristotelisPap/Question-Answering-with-BERT-and-Knowledge-Distillation Stargazers