Shashank Gupta (shatu)

shatu

Geek Repo

Company:Allen Institute for AI (AI2)

Location:Seattle, Washington

Home Page:https://shashankgupta.info/

Twitter:@shashank_bits

Github PK Tool:Github PK Tool


Organizations
CogComp

Shashank Gupta's repositories

adapter-transformers

Huggingface Transformers + Adapters = ❤️

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

alpaca-lora

Instruct-tune LLaMA on consumer hardware

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DialoGPT

Large-scale pretraining for dialogue

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
Language:DockerfileStargazers:0Issues:1Issues:0

awesome-system-design-resources

This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems

License:GPL-3.0Stargazers:0Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

dopamine

Dopamine is a research framework for fast prototyping of reinforcement learning algorithms.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:2Issues:0

FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Generating_Text_Summary_With_GPT2

A simple approach to use GPT2-medium (345M) for generating high quality text summaries with minimal training.

Language:Jupyter NotebookStargazers:0Issues:1Issues:0

gorilla

Gorilla: An API store for LLMs

License:Apache-2.0Stargazers:0Issues:0Issues:0

langchain

⚡ Building applications with LLMs through composability ⚡

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

NeuralDialog-CVAE

Tensorflow Implementation of Knowledge-Guided CVAE for dialog generation ACL 2017. It is released by Tiancheng Zhao (Tony) from Dialog Research Center, LTI, CMU

Language:OpenEdge ABLLicense:Apache-2.0Stargazers:0Issues:1Issues:0

nlp_tasks

Natural Language Processing Tasks and References

License:Apache-2.0Stargazers:0Issues:2Issues:0

OpenNMT-tf

Neural machine translation and sequence learning using TensorFlow

Language:PythonLicense:MITStargazers:0Issues:2Issues:0

ParlAI

A framework for training and evaluating AI models on a variety of openly available dialogue datasets.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

prm800k

800,000 step-level correctness labels on LLM solutions to MATH problems

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

PyMarlin

Lightweight Deep Learning Model Training library based on PyTorch

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

pytorch-pretrained-BERT

📖The Big-&-Extending-Repository-of-Transformers: Pretrained PyTorch models for Google's BERT, OpenAI GPT & GPT-2, Google/CMU Transformer-XL.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0

shatu.github.io

Code for the personal website

Language:HTMLStargazers:0Issues:1Issues:0

SimCSE

EMNLP'2021: SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

SpaceFusion

An implementation for the SpaceFusion model, https://arxiv.org/abs/1902.11205

Language:PythonStargazers:0Issues:1Issues:0

spaCy

💫 Industrial-strength Natural Language Processing (NLP) with Python and Cython

Language:PythonLicense:MITStargazers:0Issues:2Issues:0

stanford_alpaca

Code and documentation to train Stanford's Alpaca models, and generate the data.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

TheoremQA

The dataset and code for paper: TheoremQA: A Theorem-driven Question Answering dataset

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

ThoughtSource

A central, open resource for data and tools related to chain-of-thought reasoning in large language models. Developed @ Samwald research group: https://samwald.info/

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

tree-of-thought-llm

Tree of Thoughts: Deliberate Problem Solving with Large Language Models

Stargazers:0Issues:0Issues:0

unify-parameter-efficient-tuning

Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0