John St. John (jstjohn)

jstjohn

Geek Repo

Company:Nvidia

Location:Santa Clara, CA

Home Page:https://www.linkedin.com/in/johnstjohn/

Github PK Tool:Github PK Tool

John St. John's starred repositories

ml-4m

4M: Massively Multimodal Masked Modeling

Language:PythonLicense:Apache-2.0Stargazers:1435Issues:0Issues:0

LLM101n

LLM101n: Let's build a Storyteller

Stargazers:25538Issues:0Issues:0

ChannelViT

Channel Vision Transformers: An Image Is Worth C x 16 x 16 Words

Language:PythonLicense:NOASSERTIONStargazers:45Issues:0Issues:0

mup

maximal update parametrization (µP)

Language:Jupyter NotebookLicense:MITStargazers:1255Issues:0Issues:0

mutransformers

some common Huggingface transformers in maximal update parametrization (µP)

Language:Jupyter NotebookLicense:MITStargazers:75Issues:0Issues:0

mamba

Mamba SSM architecture

Language:PythonLicense:Apache-2.0Stargazers:11901Issues:0Issues:0

Tranception

Official repository for the paper "Tranception: Protein Fitness Prediction with Autoregressive Transformers and Inference-time Retrieval"

Language:PythonLicense:MITStargazers:129Issues:0Issues:0

the-algorithm-ml

Source code for Twitter's Recommendation Algorithm

Language:PythonLicense:AGPL-3.0Stargazers:9979Issues:0Issues:0

OpenAttack

An Open-Source Package for Textual Adversarial Attack.

Language:PythonLicense:MITStargazers:663Issues:0Issues:0

mm-cot

Official implementation for "Multimodal Chain-of-Thought Reasoning in Language Models" (stay tuned and more will be updated)

Language:PythonLicense:Apache-2.0Stargazers:3721Issues:0Issues:0

Open-Assistant

OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.

Language:PythonLicense:Apache-2.0Stargazers:36864Issues:0Issues:0

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

License:NOASSERTIONStargazers:25958Issues:0Issues:0

KendallWeightedDistance

Code repository for computation of Kendall's weighted distance

Language:RStargazers:2Issues:0Issues:0
Language:PythonStargazers:98Issues:0Issues:0

dowhy

DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.

Language:PythonLicense:MITStargazers:6922Issues:0Issues:0

insertion-deletion-ddpm

Implementation of Insertion-deletion Denoising Diffusion Probabilistic Models

License:MITStargazers:30Issues:0Issues:0

jaxopt

Hardware accelerated, batchable and differentiable optimizers in JAX.

Language:PythonLicense:Apache-2.0Stargazers:908Issues:0Issues:0

set_transformer

Pytorch implementation of set transformer

Language:Jupyter NotebookLicense:MITStargazers:529Issues:0Issues:0

theseus

A library for differentiable nonlinear optimization

Language:PythonLicense:MITStargazers:1680Issues:0Issues:0

d2l-en

Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 500 universities from 70 countries including Stanford, MIT, Harvard, and Cambridge.

Language:PythonLicense:NOASSERTIONStargazers:22648Issues:0Issues:0

openfold

Trainable, memory-efficient, and GPU-friendly PyTorch reproduction of AlphaFold 2

Language:PythonLicense:Apache-2.0Stargazers:2645Issues:0Issues:0

paxml

Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.

Language:PythonLicense:Apache-2.0Stargazers:438Issues:0Issues:0

imagen-pytorch

Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch

Language:PythonLicense:MITStargazers:7899Issues:0Issues:0

attention_with_linear_biases

Code for the ALiBi method for transformer language models (ICLR 2022)

Language:PythonLicense:MITStargazers:493Issues:0Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

Language:PythonLicense:Apache-2.0Stargazers:6729Issues:0Issues:0

enformer-tensorflow-sonnet-training-script

The full training script for Enformer - Tensorflow Sonnet

Language:PythonLicense:Apache-2.0Stargazers:17Issues:0Issues:0

latent-diffusion

High-Resolution Image Synthesis with Latent Diffusion Models

Language:Jupyter NotebookLicense:MITStargazers:11188Issues:0Issues:0

DALLE2-pytorch

Implementation of DALL-E 2, OpenAI's updated text-to-image synthesis neural network, in Pytorch

Language:PythonLicense:MITStargazers:10977Issues:0Issues:0

enformer-pytorch

Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch

Language:PythonLicense:MITStargazers:408Issues:0Issues:0

fairseq

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Language:PythonLicense:MITStargazers:29880Issues:0Issues:0