yangchao (emigmo)

emigmo

Geek Repo

Company:Tsinghua University

Location:Beijing

Github PK Tool:Github PK Tool


Organizations
tsinghua-rll

yangchao's starred repositories

LLMs-from-scratch

Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:28116Issues:303Issues:91

MinerU

A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。

Language:PythonLicense:AGPL-3.0Stargazers:11868Issues:68Issues:411

sglang

SGLang is a fast serving framework for large language models and vision language models.

Language:PythonLicense:Apache-2.0Stargazers:5400Issues:55Issues:541
Language:PythonLicense:Apache-2.0Stargazers:4099Issues:52Issues:119

Monkey

【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models

Language:PythonLicense:MITStargazers:1782Issues:22Issues:131

Memary

The Open Source Memory Layer For Autonomous Agents

Language:Jupyter NotebookLicense:MITStargazers:1404Issues:14Issues:28
Language:PythonLicense:Apache-2.0Stargazers:1214Issues:14Issues:112
Language:PythonLicense:NOASSERTIONStargazers:1170Issues:21Issues:76

openvla

OpenVLA: An open-source vision-language-action model for robotic manipulation.

Language:PythonLicense:MITStargazers:1087Issues:17Issues:99

ADAS

Automated Design of Agentic Systems

Language:PythonLicense:Apache-2.0Stargazers:917Issues:18Issues:15

VITA

✨✨VITA: Towards Open-Source Interactive Omni Multimodal LLM

Language:PythonLicense:NOASSERTIONStargazers:807Issues:38Issues:45
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:795Issues:18Issues:73

Awesome-Robotics-3D

A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites

SPPO

The official implementation of Self-Play Preference Optimization (SPPO)

Language:PythonLicense:Apache-2.0Stargazers:477Issues:29Issues:18
Language:PythonLicense:MITStargazers:337Issues:6Issues:6

3D-VLA

[ICML 2024] 3D-VLA: A 3D Vision-Language-Action Generative World Model

OPERA

[CVPR 2024 Highlight] OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation

Language:PythonLicense:MITStargazers:266Issues:2Issues:45

VIINA

VIINA: Violent Incident Information from News Articles on the 2022 Russian Invasion of Ukraine

OmniCorpus

OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text

embodied-agents

Seamlessly integrate state-of-the-art transformer models into robotics stacks

Language:PythonLicense:Apache-2.0Stargazers:160Issues:5Issues:14

MCTS-DPO

This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:121Issues:1Issues:3

droid

Distributed Robot Interaction Dataset.

Language:Jupyter NotebookStargazers:109Issues:4Issues:24

llm_debate

Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"

Language:PythonLicense:MITStargazers:75Issues:4Issues:2

PAI

[ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs

Language:PythonLicense:MITStargazers:49Issues:2Issues:4

Video-of-Thought

Codes for ICML 2024 paper: "Video-of-Thought: Step-by-Step Video Reasoning from Perception to Cognition"

Source-Bias

Code for "Neural Retrievers are Biased Towards LLM-Generated Content"

Language:PythonStargazers:12Issues:3Issues:0

nesca-pytorch

PyTorch Implementation for the paper "Let Me Help You! Neuro-Symbolic Short-Context Action Anticipation" accepted to RA-L'24.

Language:PythonLicense:MITStargazers:9Issues:0Issues:0

Dr_DPO

Towards Robust Alignment of Language Models: Distributionally Robustifying Direct Preference Optimization

Language:PythonStargazers:7Issues:1Issues:0