Yiheng Xu (ranpox)

ranpox

Geek Repo

Location:C-137

Home Page:https://yihengxu.com

Twitter:@yihengxu_

Github PK Tool:Github PK Tool


Organizations
doc-analysis
hitlug
OpenLemur

Yiheng Xu's starred repositories

metaseq

Repo for external large-scale work

Language:PythonLicense:MITStargazers:6386Issues:109Issues:292

BLIP

PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation

Language:Jupyter NotebookLicense:BSD-3-ClauseStargazers:4256Issues:34Issues:187

OpenAgents

OpenAgents: An Open Platform for Language Agents in the Wild

Language:PythonLicense:Apache-2.0Stargazers:3399Issues:39Issues:91

torchscale

Foundation Architecture for (M)LLMs

Language:PythonLicense:MITStargazers:2918Issues:45Issues:71

awesome-phd-advice

Collection of advice for prospective and current PhD students

License:MITStargazers:1375Issues:31Issues:0

OpenDelta

A plug-and-play library for parameter-efficient-tuning (Delta Tuning)

Language:PythonLicense:Apache-2.0Stargazers:938Issues:18Issues:61

bigscience

Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.

Language:ShellLicense:NOASSERTIONStargazers:937Issues:36Issues:19

natural-instructions

Expanding natural instructions

Language:PythonLicense:Apache-2.0Stargazers:897Issues:21Issues:161

how-do-vits-work

(ICLR 2022 Spotlight) Official PyTorch implementation of "How Do Vision Transformers Work?"

Language:PythonLicense:Apache-2.0Stargazers:787Issues:7Issues:41
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:783Issues:22Issues:28

aclpubcheck

Tools for checking ACL paper submissions

Language:PythonLicense:MITStargazers:530Issues:4Issues:45

Lemur

[ICLR 2024]Lemur: Open Foundation Models for Language Agents

Language:PythonLicense:Apache-2.0Stargazers:516Issues:9Issues:6

magma

MAGMA - a GPT-style multimodal model that can understand any combination of images and language. NOTE: The freely available model from this repo is only a demo. For the latest multimodal and multilingual models from Aleph Alpha check out our website https://app.aleph-alpha.com

Language:PythonLicense:MITStargazers:459Issues:11Issues:34

gpt_paper_assistant

GPT4 based personalized ArXiv paper assistant bot

Language:PythonLicense:Apache-2.0Stargazers:417Issues:6Issues:10

CLIP-ViL

[ICLR 2022] code for "How Much Can CLIP Benefit Vision-and-Language Tasks?" https://arxiv.org/abs/2107.06383

Language:PythonLicense:MITStargazers:381Issues:9Issues:33

METER

METER: A Multimodal End-to-end TransformER Framework

Language:PythonLicense:MITStargazers:350Issues:6Issues:36

OSWorld

OSWorld: A real computer environment for multimodal agents to evaluate open-ended computer tasks

Language:PythonLicense:Apache-2.0Stargazers:347Issues:9Issues:3

xlang-paper-reading

Paper collection on building and evaluating language model agents via executable language grounding

SEAL

Search Engines with Autoregressive Language models

Language:PythonLicense:NOASSERTIONStargazers:269Issues:6Issues:13

MetaICL

An original implementation of "MetaICL Learning to Learn In Context" by Sewon Min, Mike Lewis, Luke Zettlemoyer and Hannaneh Hajishirzi

Language:PythonLicense:NOASSERTIONStargazers:238Issues:9Issues:21

RetrivalLMPapers

Paper collections of retrieval-based (augmented) language model.

EditEval

An instruction-based benchmark for text improvements.

Language:PythonLicense:CC0-1.0Stargazers:138Issues:8Issues:2

DallEval

DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generation Models (ICCV 2023)

Language:Jupyter NotebookLicense:MITStargazers:133Issues:7Issues:2

iclr2024-openreview-submissions

ICLR 2024 OpenReivew Submission Data

WorldModelPapers

Paper collections of the continuous effort start from World Models.

License:MITStargazers:99Issues:8Issues:0

icl-selective-annotation

[ICLR 2023] Code for our paper "Selective Annotation Makes Language Models Better Few-Shot Learners"

ZeroGen

[EMNLP 2022] Code for our paper “ZeroGen: Efficient Zero-shot Learning via Dataset Generation”.

Sparkles

Sparkles: Unlocking Chats Across Multiple Images for Multimodal Instruction-Following Models

Language:PythonLicense:BSD-3-ClauseStargazers:35Issues:1Issues:3

GPT-4V-API

Self-hosted GPT-4V api

Language:JavaScriptLicense:MITStargazers:30Issues:1Issues:1

multilingual-conala

[EACL'23] MCoNaLa: A Benchmark for Code Generation from Multiple Natural Languages

Language:PythonLicense:CC-BY-SA-4.0Stargazers:21Issues:2Issues:1