Poodle
Projecting Observations On a Deep Learned Embedding
Background
Clustering techniques that use deep learned embeddings often outperform conventional clustering techniques such as k-means [1]. However, when it comes to projecting new samples onto the learned embedding there is a lack of guidelines & tools. We built POODLE to facilitate the projection of new samples onto this product space. Samples are clustered one-by-one according to their orientation in the latent space.
Deep learning technique
We used the autoencoder architecture of MAUI as an example. However, one could also adopt a different deep learning architecture or even a factor analysis technique (like MOFA). Currently, this github repo does not provide examples for other techniques.
Robust to difference in dimensionality
Poodle is flexible for situations where certain data is absent in the clinic, as one may build a shared product space and only project patients on the variables present in both sets. However, ensure that the key features are still included. The more you diverge from the initial set of features, the more you'll loose the cluster essence.
Installation
Once you have downloaded the github repo you can install the required packages by running:
$ pip install -r requirements.txt
How does poodle work?
Arguably, the best way to get familiar with the capabilities of poodle, is to start a notebook session on your device and open the following example: Start here
For those that prefer to start right away, we have listed the essential functions down below:
Functions in poodle
Specify the structure of the data
You need to specify the columns for each modality, in case you want to use different modalities. If your variables are all of the same type, one list will suffice.
d_input = {'cat' : list(CATEGORIC_VARIABLES), 'num' : list(NUMERIC_VARIABLES)}
Disclaimer: Columns need to be present in both the original & new set. If there happens to be a discrepancy between the sets you need to learn a shared product space (see this notebook for more info).
Prepare patient projection
# Before projecting a new patient, you need to update metadata with the new information.
from poodle import utils as pup
# create metadata
df_meta = pup.getMetaDataPatient(df_cluster, list(df_cluster['pseudoId'].values), new_pat)
Labeling new patients
# Now you can project the patient onto the learned embedding. You need to supply the following:
# the model (i.e. maui), metadata, original latent space, modality information and sample data.
# project a new patient and calculate similarity with each cluster
z = pup.getOrientation(maui_model, df_meta, z_existent, d_input, sample, sim_matrix, cluster_label='PhenoGraph_clusters')
# Collect coordinates of newly added patients
y_pred = pup.classifyPatient(z, path='../example_data/model/labeler/')
Output:
z
: the orientation of a patient relative to each clusters (represented by a few scalars as shown in figure 1)y_pred
: the predicted probabilities for a new patient
Visualization in poodle
Check quality of replicate clusters vs shared product space
from poodle import visualization as viz
import pandas as pd
# Import clustering probabilities of all new patients
df_projection = pd.read_csv('../example_data/results/ClusterAssignment_NewSamples.csv', sep=',')
# Plot both original & replicate distribution
viz.plotQualityControl(df_cluster[['Cluster', 'pseudoId']], df_projection, z_existent, pd.DataFrame(z_new))
Show differences in spatial variance
viz.plotSpatialVariation(l_new, l_old)
Map a specific patient
viz.plotClusterMapping(df_meta, z, new_pat)
Show top 10 closest neighbours
df_neighbours = pup.find_neighbours(df_meta, z, new_pat)
viz.plot_neighbours(df_neighbours, new_pat)
Identify batch effects with tSNE
pup.plot_tsne(fit1, group_id=metadata_original['IS_REPLICATE'], path='figures/original/tsne_original.png')
pup.plot_tsne(fit2, group_id=metadata_w_replicates['IS_REPLICATE'], path='figures/replication/tsne_replication_projected.png')
Product space (2D representation) | Newly projected patients on product space |
---|---|
Methodology
Assigning new samples to clusters
Poodle assigns a label based on the orientation of the (new) sample in the learned embedding. It does so by creating two distributions for each cluster:
- The sample vs cluster similarity (How similar is the sample to the cluster?)
- The within cluster similarity (How stable is the cluster?)
The characteristics of these distributions are captured and used to predict the cluster that a sample belongs to. We adopt an ML-technique to carefully weight each of these predictors untill the Labeler reaches a near perfect performance (see this notebook for confusion matrix).
WIP
Be aware that this github repo is still a work in progress. We will update the readme as we make new additions to the tool. For example: we aim to add baseline comparison and batch correction in the near future. Furthermore, we want to improve the functionality for users that work with uniform data (w/ single modality).