agiagoulas / qa_predictions

Question Answering Predictions from Finetuned Google BERT Models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Question Answering Predictions

This repository contains Question Answering Predictions of BERT Base Multilingual Models finetuned with different datasets in two files.

The MLQA Predictions file contains the predictions of BERT Modells finetuned with the named Datasets on the MLQA Dev Dataset.

The SQuAD Predictions file contains the predicitons of one BERT Modell finetuned with the SQuADv1.1 Train Dataset on the SQuADv1.1 Dev Dataset.

Predictions

MLQA Predictions

MLQA Predictions

Name in File Used Dataset for Finetuning Used Dataset for Prediction
original_squad SQuADv1.1 Train MLQA dev-context-de-question-de
translated_squad Translated SQuAD in German MLQA dev-context-de-question-de
german_mlqa MLQA train-context-de-question-de MLQA dev-context-de-question-de
zero_shot_english - MLQA dev-context-en-question-en
zero_shot_german - MLQA dev-context-de-question-de

SQuAD Prediction

SQuAD Predictions

Name in File Used Dataset for Finetuning Used Dataset for Prediction
original_squad SQuADv1.1 Train SQuADv1.1 Dev

Evalutation Results

The evaluation of the predictions with the MLQA Evaluation Script produces the following results in the Exact Match and F1 Metric.

Dataset Exact Match F1
Original SQuAD 81,901608 88,947487
Original SQuAD with MLQA 66,898954 80,266104
German SQuAD with MLQA 44,140625 61,369219
MLQA-de 39,0625 54,079560
Zero-Shot in English 0,0871080 5,5916380
Zero-Shot in German 0,1953125 5,9733396

Parameters

About

Question Answering Predictions from Finetuned Google BERT Models

License:Creative Commons Attribution Share Alike 4.0 International