Diana21170648 / graphicalLasso

Graphical lasso for Sparse Precision Estimation(SPE)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Graphical Lasso

This is a series of realizations of graphical lasso , which is an idea initially from Sparse inverse covariance estimation with the graphical lasso by Jerome Friedman , Trevor Hastie , and Robert Tibshirani. Graphical Lasso maximizes likelihood of precision matrix:

The objective can be formulated as,

Before that, Estimation of Precision is based on neighborhood regression, which failed to consider overall information,

Here is a (partial) historical overview of sparse precision estimation.

the question is often formulated like this: Given N observations , with dimension p , covariance , how to model the relationship of the p features ? or in more graphcial model terms , how to pin down the edges among the p vertices?

This problem is very practically significant. Say if the price of cement increases, should the price of bricks really increase too, even if there covariance is positive? Probably not, but both might be linked through a series of ETFs like the wages of the labor. The covariance matrix is a victim of causality! Cement and bricks prices have positive covariance because they are both dependent on another set of indices, but have (close to) zero conditional correlation. So conditional independece often matters , which is decided by inverse covariance matrix.

if the (i,j) entry of the precision matrix C equals to 0 , that is = 0 , then the ith and the jth feature are conditional independet given the remain features.

The goal of graphical lasso is to estimate sparse graph, which is often controlled by the inverse coveriance matrix or so-called precision matrix given adequate data and is superiorly fast.

Laplacian and Precision

A common question is often asked: What is the difference between Graph Laplacian and Precission? The answer is that Laplacian are built on the kernel function and distance, which means that Laplaican considers local manifolds and there is an positive edge if 2 points are close. However, Precision considers overall structures, which means significantly distant points are also considered in the graph. Say we have 20 points belonging 2 classes, and 10% edges are retained, the graphs are like that,

The blue lines in the pictures represent positive edges while the gray ones represent negatives. So if precision is used as regularization terms, it has 2 effects: it will make points between one class close and make points in different class diastant.

Edited by plzhai in 2019/3/24

About

Graphical lasso for Sparse Precision Estimation(SPE)


Languages

Language:Python 100.0%