ishana7 / ST2022

SIGTYP 2022 Shared Task

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

SIGTYP 2022 Shared Task: Prediction of Cognate Reflexes

In 2022, SIGTYP is hosting a shared task on predicting words and morphemes from cognate words in related languages (Cognate Reflex Prediction Task). We provide 20 cognate-coded multilingual wordlists (10 for development, and 10 withheld for testing) from which we have withheld varying amounts of cognate sets from which the systems provided by participants should later predict individual words and morphemes. The code we provide allows to transparently access the data from its standardized format, compute sample predictions with a baseline method, and to evaluate results with three evaluation metrics.

Important Dates

Important Links

1 The Reflex Prediction Task

In historical-comparative linguistics, scholars typically assemble words from related languages into cognate sets. In contrast to the notion of cognates in didactics and synchronic NLP applications, cognate words -- the members of a cognate set -- are commonly assumed to share a common origin regardless of their meaning. In addition, cognate sets should not contain borrowed words. Cognate words typically show so-called regular sound correspondences. This means that one can define a mapping across the individual phoneme systems of the individual languages. Thus, English t typically corresponds to a German ts (compare ten vs. zehn), and English d corresponds to German t (compare dove vs. Taube). The mappings often depend on certain contextual conditions and may differ, depending on the position in which they occur in a words. Due to regular sound correspondences, linguists can often predict fairly well how the cognate counterpart of a word in one language might sound in another language. However, prediction by linguists rarely takes only one language pair into account. The more reflexes (counterparts) a cognate set has in different languages, the easier it is to predict reflexes in individual languages.

In its simplest form, the data we need for the task of reflex prediction is a table in which each column represents a different language and each row a different cognate set. Whenever a reflex in a specific language is missing, this reflex can in theory be predicted with the help of the remaining reflexes. As an example, consider the following table showing reflexes of cognate sets in German, English, and Dutch.

Cognate Set German English Dutch
ASH a ʃ ə æ ʃ ɑ s
BITE b ai s ə n b ai t b ɛi t ə
BELLY b au x - b œi k

Since the reflex for the BELLY cognate sets is missing in English, we could try and predict it from known correspondences to German and Dutch. In fact, some English dialects even seem to keep the form bouk, which would be the correct prediction for this missing reflex.

When provided with more data of this kind, one can provide a model that would be able to predict an English form given a German and a Dutch form, as well as a German form, given a Dutch and an Englisch form, and so on. Note that not all cognate sets in real-life data will have reflexes for all words. Thus, we know about English bouk from dialect records, but without dialects or written sources from Middle English, we could only rely on prediction itself in order to guess how the word would sound if it would have been retained in the respective language.

Since predictions for words that have been completely lost cannot be evaluated directly, we will base our task on the prediction of artificially excluded word forms. Thus, we first take a dataset with cognates in a couple of related languages, and then artificially delete some of the words in the datasets, using varying proportions. When training a model to predict the missing word forms, we can just compare the predicted words directly with the words we have deleted automatically, as was done in one early study on word prediction (List 2019).

The task of supervised phonological reconstruction can be seen as a special case of the reflex prediction task. While we predict reflexes in any language in the generic reflex prediction task, we predict one specific reflex of a cognate set, the form in the ancestral language in automated phonological reconstruction.

2 Previous Work on Reflex Prediction

There are quite a few studies on supervised phonological reconstruction now. Unfortunately, only a few of them provide their source code and original data. The following list is not meant to be exhaustive, and we appreciate it if you point us to studies missing here by opening an issue.

  • Liviu Dinu and Alina Maria Ciobanu. 2014. Building a dataset of multilingual cognates for the Romanian lexicon. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), pages 1038–1043, Reykjavik, Iceland. European Language Resources Association (ELRA).
  • Timotheus Adrianus Bodt and Johann-Mattis List. 2021. Reflex prediction. a case study of Western Kho-Bwa. Diachronica:1–38.
  • Johann-Mattis List. 2019a. Automatic inference of sound correspondence patterns across multiple languages. Computational Linguistics, 45(1):137–161.
  • Carlo Meloni, Shauli Ravfogel, and Yoav Goldberg. 2021. Ab antiquo: Neural proto-language recon- struction. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4460–4473, Online. Association for Computational Linguistics.
  • Lisa Beinborn, Torsten Zesch, and Iryna Gurevych. 2013. Cognate production using character-based machine translation. In Proceedings of the Sixth International Joint Conference on Natural Language Processing, pages 883–891.
  • Peter Dekker and Willem Zuidema. 2021. Word prediction in computational historical linguistics. Journal of Language Modelling, 8(2):295–336.
  • Clémentine Fourrier, Rachel Bawden, and Benoît Sagot. 2021. Can cognate prediction be modelled as a low-resource machine translation task? In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 847–861, Online. Association for Computational Linguistics.

3 Datasets and Data Formats

Our data is taken from the Lexibank repository which offers wordlists from 100 standardized datasets (List et al. 2021. In the repository, a larger collection of datasets come with cognate sets provided by experts and with phonetic transcriptions which were standardized by the Lexibank team. An even larger part comes without cognate sets but with standardized phonetic transcriptions. Since cognate detection methods work well by now, allowing us to assess the cognates specifically for shallower language families with quite some confidence, this enables us to assemble a larger amount of datasets from different language families and either use cognate sets provided by experts or inferring cognates ourselves, using state-of-the-art methods for automated cognate detection which have been implemented in the LingPy software library.

This means that we will have quite a few datasets that we can offer for development and testing. The data is originally coded in CLDF formats (Forkel et al. 2018). In order to ease the access to the data in simple tabular form (which we consider most suitable for the training of supervised reflex prediction workflows), we provide small wrappers that allow to access individual datasets from within Python scripts.

The following code shows, for example, how one can load a file with training data.

from sigtypst2022 import load_cognate_file
languages, sounds, data = load_cognate_file("data/allenbai/training-0.20.tsv")

In this example, languages is a list of language names. The variable sounds is a dictionary with unique sounds as keys, pointing to a dictionary with language names as values which themselves link to a list of cognate set identifiers and positions in the word. The variable data is a dictionary of dictionaries with cognate set identifiers as key, and as value a dictionary with language names as key and segmented words as value.

All wordlist data are represented in this way, also the test data, and the solutions for the test data.

When testing the methods on the unseen datasets (the surprise data), we provide the input data in the form of a CSV file in which those words which should be predicted are marked by an ?, while missing entries (resulting from real gaps in the cognate sets) are simply left blank.

We provide different proportions of missing data, ranging from 10% up to 50% i steps of 10%. This will allow users to check how robust their systems are with respect to differing degrees of missing data.

4 Evaluation

4.1 Overview

The expected prediction result for a given reflex is a list of phonetic transcription symbols (we segment all words in our CLDF datasets into sound units). This prediction can be directly compared against the attested form, which was removed from the data when training the model. A common metric by which we can compare the predicted form with the attested form is the edit distance. Formally, the edit distance is identical with the Hamming distance between the alignment of two strings. Working with alignments should be preferred to working with an algorithm computing the edit distance alone, since alignments are very useful for error analysis. Furthermore, when working with algorithms that produce fuzzy predictions which might predict more than one probable sound in a given sound slot, or alternative word forms, it is easier to score these predictions based on alignments, as shown in the study of Bodt and List (2021).

An additional evaluation measure based on alignments was proposed by List (2019b). This measure takes into account that an algorithm might in theory commit systematic errors, which might be overly penalized by the edit distance. This method, which computes the B-Cubed F-Scores between the aligned predictions and attested forms, has been ignored in most approaches to supervised reflex prediction, but we consider it nevertheless useful to include it into the evaluation scores to be reported, since it comes theoretically much closer to the idea of regular sound correspondences in classical historical linguistics.

After some discussion, we decided to add the BLEU scores (Papineni et al. 2002), although there were mixed opinions in our team, to which degree they are useful for this task in the end. But BLEU scores were now included with LingRex, version 1.2 and are therefore included in our little library for the shared task.

Our dedicated Python package for the shared task allows to compute all evaluation measures mentioned above (edit distance in raw and normalized form, B-Cubed F-Scores, BLEU Scores) from TSV files which can be passed to the command line as input. For the purpose of development, scholars can also use these metrics directly from within their Python scripts.

As a baseline, we provide the method by List (2019) in a new version by List et al. (forthcoming) for which we made a new release of the LingRex Python package (Version 1.2).

4.2 Results

The following results can be plotted with the st2022 package (version 1.4) by typing:

$ st2022 --meta-evaluation --datapath=data-surprise --datasets=datasets-surprise.json --format=pipe

They contain all ranks for the various cuts of the data, for three major methods (we exclude the non-normalized edit distance here).

System NED 1 BCFS 1 BLEU 1 NED 2 BCFS 2 BLEU 2 NED 3 BCFS 3 BLEU 3 NED 4 BCFS 4 BLEU 4 NED 5 BCFS 5 BLEU 5 NED B-Cubed FS BLEU Aggregate Aggregate (STD)
Mockingbird-I1 1 1 1 1 1 1 1 1 1 1 1 1 1 2 1 1 1.2 1 1.06667 0.258199
Mockingbird-N1-A 2 3 2 2 3 2 3 3 3 3 3 3 3 3 3 2.6 3 2.6 2.73333 0.457738
Mockingbird-N1-B 3 4 3 3 4 3 2 4 2 2 4 2 2 4 2 2.4 4 2.4 2.93333 0.883715
Baseline-Baseline-SVM 4 2 4 4 2 4 5 2 5 5 7 5 8 7 7 5.2 4 5 4.73333 1.90738
Mockingbird-N1-C 6 6 6 5 8 5 4 7 4 4 6 4 4 6 4 4.6 6.6 4.6 5.26667 1.27988
CEoT-Extalign-RF 5 5 5 6 6 6 6 8 6 6 8 6 7 8 8 6 7 6.2 6.4 1.12122
CrossLingference-Julia 9 7 9 8 5 8 8 5 8 8 2 8 5 1 5 7.6 4 7.6 6.4 2.47271
Baseline-Baseline 7 8 7 7 7 7 7 6 7 7 5 7 6 5 6 6.8 6.2 6.8 6.6 0.828079
Leipzig-Transformer 8 9 8 9 9 9 9 9 9 9 9 9 9 9 9 8.8 9 8.8 8.86667 0.351866

5 Data for Development and Data for the Final Evaluation (Surprise Data)

Our development data, which users should use to test and design their models, consists of 10 CLDF datasets of varying size, language families, and time depths. Some datasets may be extremely small, giving algorithms a hard time to predict words correctly, others are rather large, but languages may be distantly related. Data for the development phase may contain datasets which have been used in previous studies.

For the actual testing of different systems, we have selected 10 CLDF datasets which are similar to the data in the training set, with respect to the language families in question and the size of the datasets, and have so far only rarely been used in previous studies.

6 Participation

In order to participate in the shared task, we ask participants to write their code in a transparent way that can be directly tested. Testable code requires that all dependencies of the code package are listed and that the code has been tested in a virtual environment. We also ask for detailed instructions that help in installing the packages and explain how they should be used.

7 Concluding Remarks

In case of questions, we kindly ask to raise issues, which we will try to answer as quickly as possible. In order to guarantee that users use the same datasets and the same code basis to access the data, we will provide releases of this code base. If problems should be detected at a later stage, we can make a new release and ask all users to switch to the most recent version.

About

SIGTYP 2022 Shared Task

License:MIT License


Languages

Language:Python 96.3%Language:Makefile 2.6%Language:TeX 1.1%