yangkky / UniLanguage

Homology reduced UniProt, train-/valid-/testsets for language modeling

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

What is UniLanguage?

[PAPER]    [CITATION]    [CODE]    [DATASET DOWNLOAD]

Abstract

Motivation

Language modelling (LM) on biological sequences is an emergent topic in the field of bioinformatics. Current research has shown that language modelling of proteins can create context-dependent representations that can be applied to improve performance on different protein prediction tasks. However, little effort has been directed towards analyzing the properties of the datasets used to train language models. Additionally, only the performance of cherry-picked downstream tasks are used to assess the capacity of LMs.

Results

We analyze the entire UniProt database and investigate the different properties that can bias or hinder the performance of LMs such as homology, domain of origin, quality of the data, and completeness of the sequence. We evaluate n-gram and Recurrent Neural Network (RNN) LMs to assess the impact of these properties on performance. To our knowledge, this is the first protein dataset with an emphasis on language modelling. Our inclusion of properties specific to proteins gives a detailed analysis of how well natural language processing methods work on biological sequences. We find that organism domain and quality of data have an impact on the performance, while the completeness of the proteins has little influence. %How does it perform? The RNN based LM can learn to model Bacteria, Eukarya, and Archaea; but struggles with Viruses. By using the LM we can also generate novel proteins that are shown to be similar to real proteins.

Process of creating UniLanguage

We scrape all of UniProt - 250M proteins as of 2019-10-01. The proteins with experimental evidence (high quality, 1% of UniProt) are homology partitioned, with 20% homology, into train-, validation-, and test set with 60% train, 10% validation and 30% test for each domain on both full proteins and fragments.

Predicted proteins (low quality, 99% of UniProt) are homology compared against the validation and testset with 20% homology. If predicted proteins do not overlap they are kept as a predicted training set, split into domains and fragments. This removes about 50% of the predicted proteins (approx. 85 million!), which truly indicates the homogenousness in UniProt proteins.

Dataset and results

We provide a script to obtain train, validation and test sets for all domains, qualities and protein completion.

Script usage

python get_data.py --domain [domain] --complete [complete] --quality [quality]

Parameters:

  • Domain: Eukarya=euk, Bacteria=bac, Archaea=arc, Virus=vir
  • Complete: Complete proteins=full, Fragmented proteins=frag
  • Quality: Experimental evidence=exp, Predicted=pred

Example 1: get all the data!

python get_data.py --all

Example 2: fragments for archaea, predicted quality

python get_data.py --domain arc --complete frag --quality pred

Examples 3: complete proteins for Eukarya, full and experimental quality

python get_data.py --domain euk --complete full --quality exp

Citation

[PAPER]    [CITATION]    [CODE]    [DATASET DOWNLOAD]

PUT IN CITATION

About

Homology reduced UniProt, train-/valid-/testsets for language modeling


Languages

Language:Python 100.0%