Linhui Xiao (linhuixiao)

linhuixiao

Geek Repo

Company:UCAS, Chinese Academy of Sciences

Location:Beijing, China

Home Page:https://scholar.google.com/citations?hl=zh-CN&user=4rTE4ogAAAAJ

Twitter:@xiao_linhui

Github PK Tool:Github PK Tool

Linhui Xiao's starred repositories

DynRefer

DynRefer: Delving into Region-level Multi-modality Tasks via Dynamic Resolution

Language:PythonLicense:Apache-2.0Stargazers:31Issues:0Issues:0

Books

My book list

License:MITStargazers:449Issues:0Issues:0
Language:MATLABLicense:GPL-3.0Stargazers:10277Issues:0Issues:0

Awesome-Visual-Dialog

A curated publication list on visual dialog

Stargazers:11Issues:0Issues:0

CVPR2022-FTCL

[CVPR 2022] Fine-grained Temporal Contrastive Learning for Weakly-supervised Temporal Action Localization

Language:PythonLicense:MITStargazers:46Issues:0Issues:0

CVPR2023-OWTAL

[CVPR 2023] Cascade Evidential Learning for Open-world Weakly-supervised Temporal Action Localization

Language:PythonLicense:MITStargazers:8Issues:0Issues:0

ECCV2022-DELU

[ECCV 2022] Dual-Evidential Learning for Weakly-supervised Temporal Action Localization

Language:PythonLicense:MITStargazers:38Issues:0Issues:0

CVPR2023-CMPAE

[CVPR 2023] Collecting Cross-Modal Presence-Absence Evidence for Weakly-Supervised Audio-Visual Event Perception

Language:PythonLicense:MITStargazers:33Issues:0Issues:0

ICLR2024-REDL

[ICLR 2024 Spotlight] R-EDL: Relaxing Nonessential Settings of Evidential Deep Learning

Language:PythonLicense:MITStargazers:30Issues:0Issues:0

Awesome-Visual-Grounding

A Survey on Open Visual Grounding

License:Apache-2.0Stargazers:2Issues:0Issues:0

awesome-described-object-detection

A curated list of papers and resources related to Described Object Detection, Open-Vocabulary/Open-World Object Detection and Referring Expression Comprehension. Updated frequently and pull requests welcomed.

Stargazers:160Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:54752Issues:0Issues:0

Libra

Simple PyTorch implementation of "Libra: Building Decoupled Vision System on Large Language Models" (accepted by ICML 2024)

Language:PythonLicense:Apache-2.0Stargazers:38Issues:0Issues:0

HiVG

Hierarchical Multimodal Fine-grained Modulation for Visual Grounding.

Language:PythonLicense:Apache-2.0Stargazers:26Issues:0Issues:0

COMM

Pytorch code for paper From CLIP to DINO: Visual Encoders Shout in Multi-modal Large Language Models

License:MITStargazers:180Issues:0Issues:0

MAE-pytorch

Unofficial PyTorch implementation of Masked Autoencoders Are Scalable Vision Learners

Language:PythonStargazers:2571Issues:0Issues:0

FlagAI

FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.

Language:PythonLicense:Apache-2.0Stargazers:3814Issues:0Issues:0

s4

Structured state space sequence models

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2289Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:12661Issues:0Issues:0

mamba

Mamba SSM architecture

Language:PythonLicense:Apache-2.0Stargazers:11963Issues:0Issues:0

Awesome-Mamba-Papers

Awesome Papers related to Mamba.

Stargazers:1005Issues:0Issues:0

MultiModalMamba

A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.

Language:PythonLicense:MITStargazers:422Issues:0Issues:0

Vim

[ICML 2024] Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model

Language:PythonLicense:Apache-2.0Stargazers:2673Issues:0Issues:0

r-mae

PyTorch implementation of R-MAE https//arxiv.org/abs/2306.05411

Language:PythonLicense:NOASSERTIONStargazers:106Issues:0Issues:0

DataOptim

A collection of visual instruction tuning datasets.

Language:PythonLicense:MITStargazers:73Issues:0Issues:0
Stargazers:35Issues:0Issues:0

NExT-Chat

The code of the paper "NExT-Chat: An LMM for Chat, Detection and Segmentation".

Language:PythonLicense:Apache-2.0Stargazers:191Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:74Issues:0Issues:0

SCLIP

Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference

Language:PythonStargazers:102Issues:0Issues:0

SceneGraphParser

A python toolkit for parsing captions (in natural language) into scene graphs (as symbolic representations).

Language:PythonLicense:MITStargazers:523Issues:0Issues:0