rdk / p2rank

P2Rank: Protein-ligand binding site prediction tool based on machine learning. Stand-alone command line program / Java library for predicting ligand binding pockets from protein structure.

Home Page:https://rdk.github.io/p2rank/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

What is a huge dataset?

skodapetr opened this issue · comments

In the training-tutorial.md there is a note '''Turn off for huge datasets that won't fit to memory.'''. It would really help to give an estimate of how much memory does the p2rank use. As one may consider a hundreds of proteins a small dataset that should fit into main memory. For example if I have 2500 proteins and 32G RAM I would expect that it is fine, but is it?

commented

This doesn't really have a simple answer. It depends on many factors including size of proteins, density of SAS points and number of RF trees you want to train in parallell. Yes, 2500 proteins of average size (say a 2500 atoms) will definitely fit into 32GB of RAM, but it might not be enough RAM to train RF in parallel on 16 threads.

I have reformulated the tutorials and added section Raquired memory and memory/time trade-offs to the training-tutorial.md. Check if you are still missing any information there.

I don't have particular memory usage estimates at hand, I will try to measure them and include them at some point - or will welcome your contribution.
One recent estimate though: training on MG dataset of ~2200 proteins with 12 threads, default density and no subsampling needed ~55GB RAM.