uobinxiao's starred repositories

auto-code-rover

A project structure aware autonomous software engineer aiming for autonomous program improvement. Resolved 30.67% tasks (pass@1) in SWE-bench lite with each task costs less than $0.7.

Language:PythonLicense:NOASSERTIONStargazers:2355Issues:0Issues:0

Awesome-Video-Object-Segmentation

A curated list of video object segmentation (vos) papers, datasets, and projects.

Stargazers:144Issues:0Issues:0
Language:PythonStargazers:4Issues:0Issues:0

llama.cpp

LLM inference in C/C++

Language:C++License:MITStargazers:61096Issues:0Issues:0

Awesome-TimeSeries-SpatioTemporal-LM-LLM

A professional list on Large (Language) Models and Foundation Models (LLM, LM, FM) for Time Series, Spatiotemporal, and Event Data.

Stargazers:820Issues:0Issues:0

Transformers-for-NLP-2nd-Edition

Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL-E including jump starting GPT-4, speech-to-text, text-to-speech, text to image generation with DALL-E, Google Cloud AI,HuggingGPT, and more

Language:Jupyter NotebookLicense:MITStargazers:738Issues:0Issues:0

VisCPM

[ICLR'24 spotlight] Chinese and English Multimodal Large Model Series (Chat and Paint) | 基于CPM基础模型的中英双语多模态大模型系列

Language:PythonStargazers:1034Issues:0Issues:0

MiniCPM-V

MiniCPM-Llama3-V 2.5: A GPT-4V Level Multimodal LLM on Your Phone

Language:PythonLicense:Apache-2.0Stargazers:7776Issues:0Issues:0

Time-LLM

[ICLR 2024] Official implementation of " 🦙 Time-LLM: Time Series Forecasting by Reprogramming Large Language Models"

Language:PythonLicense:Apache-2.0Stargazers:993Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:21307Issues:0Issues:0

wesam

[CVPR 2024] Code for "Improving the Generalization of Segmentation Foundation Model under Distribution Shift via Weakly Supervised Adaptation"

Language:PythonLicense:MITStargazers:104Issues:0Issues:0

Semantic-SAM

Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"

Language:PythonStargazers:2044Issues:0Issues:0

MGM

Official repo for "Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models"

Language:PythonLicense:Apache-2.0Stargazers:3072Issues:0Issues:0

mPLUG-DocOwl

mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding

Language:PythonLicense:Apache-2.0Stargazers:1108Issues:0Issues:0

DeepSeek-Coder

DeepSeek Coder: Let the Code Write Itself

Language:PythonLicense:MITStargazers:6004Issues:0Issues:0

evolutionary-model-merge

Official repository of Evolutionary Optimization of Model Merging Recipes

Language:PythonLicense:Apache-2.0Stargazers:1076Issues:0Issues:0

FuseAI

FuseAI Project

Language:PythonStargazers:367Issues:0Issues:0

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:49137Issues:0Issues:0

NaViT

My implementation of "Patch n’ Pack: NaViT, a Vision Transformer for any Aspect Ratio and Resolution"

Language:PythonLicense:MITStargazers:143Issues:0Issues:0

Xwin-LM

Xwin-LM: Powerful, Stable, and Reproducible LLM Alignment

Language:PythonStargazers:1004Issues:0Issues:0

MultiHiertt

Data and code for ACL 2022 paper "MultiHiertt: Numerical Reasoning over Multi Hierarchical Tabular and Textual Data"

Language:PythonLicense:MITStargazers:40Issues:0Issues:0

ReAcTable

The code base for paper: "ReAcTable: Enhancing ReAct for Table Question Answering"

Language:HTMLStargazers:7Issues:0Issues:0

TRICE

[NAACL 2024] Making Language Models Better Tool Learners with Execution Feedback

Language:PythonLicense:MITStargazers:32Issues:0Issues:0

Neural-Network-Parameter-Diffusion

We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters

Language:PythonStargazers:785Issues:0Issues:0

ToolQA

ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels (easy/hard) across eight real-life scenarios.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:216Issues:0Issues:0

toolformer-pytorch

Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI

Language:PythonLicense:MITStargazers:1920Issues:0Issues:0

self-rag

This includes the original implementation of SELF-RAG: Learning to Retrieve, Generate and Critique through self-reflection by Akari Asai, Zeqiu Wu, Yizhong Wang, Avirup Sil, and Hannaneh Hajishirzi.

Language:PythonLicense:MITStargazers:1595Issues:0Issues:0

Awesome-Mamba-Papers

Awesome Papers related to Mamba.

Stargazers:937Issues:0Issues:0

VILA

VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)

Language:PythonLicense:Apache-2.0Stargazers:874Issues:0Issues:0

TinyLLaVA_Factory

A Framework of Small-scale Large Multimodal Models

Language:PythonLicense:Apache-2.0Stargazers:473Issues:0Issues:0