Qichao Ying (yingqichao)

yingqichao

Geek Repo

Company:NVIDIA | Fudan University

Location:Shanghai

Home Page:https://scholar.google.com/citations?user=3sletjoAAAAJ&hl=zh-CN

Twitter:@shinydotcom1631

Github PK Tool:Github PK Tool

Qichao Ying's starred repositories

segment-anything

The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:47055Issues:305Issues:663
Language:PythonLicense:NOASSERTIONStargazers:34523Issues:300Issues:353

Grounded-Segment-Anything

Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:14912Issues:113Issues:386

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Advances on Multimodal Large Language Models

pulse

PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models

FastSAM

Fast Segment Anything

Language:PythonLicense:AGPL-3.0Stargazers:7407Issues:56Issues:204

Segment-Everything-Everywhere-All-At-Once

[NeurIPS 2023] Official implementation of the paper "Segment Everything Everywhere All at Once"

Language:PythonLicense:Apache-2.0Stargazers:4336Issues:59Issues:147

sam-hq

Segment Anything in High Quality [NeurIPS 2023]

Language:PythonLicense:Apache-2.0Stargazers:3662Issues:77Issues:138

EditAnything

Edit anything in images powered by segment-anything, ControlNet, StableDiffusion, etc. (ACM MM)

Language:PythonLicense:Apache-2.0Stargazers:3295Issues:39Issues:57

SparseConvNet

Submanifold sparse convolutional networks

Language:C++License:NOASSERTIONStargazers:2031Issues:44Issues:224

bolei_awesome_posters

CVPR and NeurIPS poster examples and templates. May we have in-person poster session soon!

SparK

[ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling"

Language:PythonLicense:MITStargazers:1429Issues:26Issues:83

Video-ChatGPT

[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.

Language:PythonLicense:CC-BY-4.0Stargazers:1176Issues:14Issues:119

muse-maskgit-pytorch

Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch

Language:PythonLicense:MITStargazers:860Issues:34Issues:36

phenaki-pytorch

Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 minutes in length, in Pytorch

Language:PythonLicense:MITStargazers:747Issues:38Issues:32

FocalNet

[NeurIPS 2022] Official code for "Focal Modulation Networks"

Language:PythonLicense:MITStargazers:685Issues:15Issues:53

DynamicViT

[NeurIPS 2021] [T-PAMI] DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification

Language:Jupyter NotebookLicense:MITStargazers:564Issues:10Issues:44

DISTS

IQA: Deep Image Structure and Texture Similarity Metric

Language:PythonLicense:MITStargazers:372Issues:4Issues:12

dreamsim

DreamSim: Learning New Dimensions of Human Visual Similarity using Synthetic Data (NeurIPS 2023 Spotlight)

Language:PythonLicense:MITStargazers:354Issues:11Issues:18

segment-anything-with-clip

Segment Anything combined with CLIP

Language:PythonLicense:Apache-2.0Stargazers:329Issues:1Issues:4

davit

[ECCV 2022]Code for paper "DaViT: Dual Attention Vision Transformer"

Language:PythonLicense:MITStargazers:325Issues:5Issues:15

CLIP-IQA

[AAAI 2023] Exploring CLIP for Assessing the Look and Feel of Images

Language:PythonLicense:NOASSERTIONStargazers:318Issues:4Issues:32

Dynamic-Vision-Transformer

Accelerating T2t-ViT by 1.6-3.6x.

multiple-attention

The code of multi-attention deepfake detection

A-ViT

Official PyTorch implementation of A-ViT: Adaptive Tokens for Efficient Vision Transformer (CVPR 2022)

Language:PythonLicense:Apache-2.0Stargazers:146Issues:4Issues:14

ffhqr-dataset

FFHQR -- the first large-scale retouching dataset for computer vision research.

HQ-50K

HQ-50K: A Large-scale, High-quality Dataset for Image Restoration

CRHD-3K

The first high-definition cloth retouching dataset CRHD-3K.

Language:PythonLicense:NOASSERTIONStargazers:28Issues:1Issues:1