pathakdivya / resources

A collaborative list of resources for Computational Neuroscience

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

List of Resources

A collaborative list of resources for Computational Neuroscience

Interesting Papers/ Articles/ Blog Posts:

Information Theory

  • Foundational paper in the field of information theory by Claude Shannon in 1948 A Mathematical Theory of Communication. Might be helpful to watch this video by Kan Academy describing the work (from Markov Chain perspective) before diving into the paper.

    Details!

    This work developed the concepts of information entropy and redundancy, and introduced the term bit (which Shannon credited to John Tukey) as a unit of information. It was also in this paper that the Shannon–Fano coding technique was proposed – a technique developed in conjunction with Robert Fano.
    Shannon's article laid out the basic elements of communication:

    • An information source that produces a message
    • A transmitter that operates on the message to create a signal which can be sent through a channel
    • A channel, which is the medium over which the signal, carrying the information that composes the message, is sent
    • A receiver, which transforms the signal back into the message intended for delivery
    • A destination, which can be a person or a machine, for whom or which the message is intended

    More on Shannon and his contributions to the world of Computer sci, entropy, info theory, signal detection etc.

  • Ian Goodfellow's (developed GANs) Book Chapter on Information Theory from a Deep Learning Perspective

    Details!

    Goodfellow is best known for inventing generative adversarial networks (GANs). He is also the lead author of the textbook Deep Learning. At Google, he developed a system enabling Google Maps to automatically transcribe addresses from photos taken by Street View cars and demonstrated security vulnerabilities of machine learning systems.

Entropy

Noise

Brain Oscillations

Dimensionality

General Dimnesionality

  • Towards the neural population doctrine

    Details!

    We detail four areas of the field where the joint analysis of neural populations has significantly furthered our understanding of computation in the brain: correlated variability, decoding, neural dynamics, and artificial neural networks.

  • SVD and PCA explained. Handout walking through the math behind both and a few other topics (regression,covariance etc.).

    Details!

    This handout is a review of some basic concepts in linear algebra. For a detailed introduction, consult a linear algebra text. Linear Algebra and its Applications by Gilbert Strang (Harcourt, Brace, Jovanovich, 1988) is excellent.

Non-Linear Dimensionality Reduction

  • Using t-SNE. An interactive guide on how to use t-SNE effectively

    Details! Although extremely useful for visualizing high-dimensional data, t-SNE plots can sometimes be mysterious or misleading. By exploring how it behaves in simple cases, we can learn to use it more effectively.
  • Perform non-linear dimensionality reduction with Isomap and LLE in Python from scratch

  • Isomap tutorial in Python

  • Looking at different non-linear dimensionality reductions methods: Iterative Non-linear Dimensionality Reduction with Manifold Sculpting.

    Details!

    Many algorithms have been recently developed for reducing dimensionality by projecting data onto an intrinsic non-linear manifold. Unfortunately, existing algorithms often lose significant precision in this transformation. Manifold Sculpting is a new algorithm that iteratively reduces dimensionality by simulating surface tension in local neighborhoods. We present several experiments that show Manifold Sculpting yields more accurate results than existing algorithms with both generated and natural data-sets. Manifold Sculpting is also able to benefit from both prior dimensionality reduction efforts.

  • Using manifolds/ dimensionality reduction on sleep data. The intrinsic attractor manifold and population dynamics of a canonical cognitive circuit across waking and sleep

    Details!

    We characterize and directly visualize manifold structure in the mammalian head direction circuit, revealing that the states form a topologically nontrivial one-dimensional ring. The ring exhibits isometry and is invariant across waking and rapid eye movement sleep. This result directly demonstrates that there are continuous attractor dynamics and enables powerful inference about mechanism.

  • A Global Geometric Framework for Nonlinear Dimensionality Reduction

    Details!

    Here we describe an approach to solving dimensionality reduction problems that uses easily measured local metric information to learn the underlying global geometry of a data set. Unlike classical techniques such as principal component analysis (PCA) and multidimensional scaling (MDS), our approach is capable of discovering the nonlinear degrees of freedom that underlie complex natural observations, such as human handwriting or images of a face under different viewing conditions. In contrast to previous algorithms for nonlinear dimensionality reduction, ours efficiently computes a globally optimal solution, and, for an important class of data manifolds, is guaranteed to converge asymptotically to the true structure.

Modeling

General Modeling

Optimization for Modeling

Machine Learning

General Machine Learning

Autoencoders

  • Variational autoencoders used with dimensionality reduction. VAE-SNE: a deep generative model for simultaneous dimensionality reduction and clustering

    Details!

    Description: We introduce a method for both dimension reduction and clustering called VAE-SNE (variational autoencoder stochastic neighbor embedding). Our model combines elements from deep learning, probabilistic inference, and manifold learning to produce interpretable compressed representations while also readily scaling to tens-of-millions of observations. Unlike existing methods, VAE-SNE simultaneously compresses high-dimensional data and automatically learns a distribution of clusters within the data --- without the need to manually select the number of clusters. This naturally creates a multi-scale representation, which makes it straightforward to generate coarse-grained descriptions for large subsets of related observations and select specific regions of interest for further analysis.

Books

Datasets

  • A list of open datasets that span EEG, MEG, ECoG, and LFP.

  • A large list of BCI resources including datasets, tutorials, papers, books etc.

  • The TUH EEG Corpus, a list of several EEG dataset with several resources. Requies filling out form to download the data.

  • Project Tycho named after Tycho Brache. The project aims to share reliable massive neural and behavioral data for understanding brain mechanism.

    Details!

    Tycho Brahe was a Danish nobleman, astronomer, and writer known for his accurate and comprehensive astronomical observations. He was born in the then Danish peninsula of Scania. Tycho was well known in his lifetime as an astronomer, astrologer, and alchemist.

Videos

  • Gradients of Brain Organization Workshop.

    Details!

    Description: Recent years have seen a rise of new methods and applications to study smooth spatial transitions — or gradients — of brain organization. Identification and analysis of cortical gradients provides a framework to study brain organization across species, to examine changes in brain development and aging, and to more generally study the interrelation between brain structure, function and cognition. We will bring together outstanding junior and senior scientists to discuss the challenges and opportunities afforded by this emerging perspective.

Memes

About

A collaborative list of resources for Computational Neuroscience

License:MIT License