Ray Yang's repositories
ScalableViT
This is code of paper "ScalableViT: Rethinking the Context-oriented Generalization of Vision Transformer"
RetinaNet-BCB
This is the repository for paper RetinaNet-BCB
VLM_survey
Vision-Language Models for Vision Tasks: A Survey
alpaca-lora
Instruct-tune LLaMA on consumer hardware
Awesome-Anything
General AI methods for Anything: AnyObject, AnyGeneration, AnyModel, AnyTask, AnyX
awesome-papers-fewshot
Collection for Few-shot Learning
Awesome_Prompting_Papers_in_Computer_Vision
A curated list of prompt-based paper in computer vision and vision-language learning.
cocoapi
COCO API - Dataset @ http://cocodataset.org/
intellij-idea-tutorial
🌻 This is a tutorial of IntelliJ IDEA, you can know how to use IntelliJ IDEA better and better.
mmdetection
OpenMMLab Detection Toolbox and Benchmark
TOOD
TOOD: Task-aligned One-stage Object Detection, ICCV2021 Oral
UniFormer
[ICLR2022] official implementation of UniFormer
UniInst
UniInst
Awesome-Multimodal-Research
A curated list of Multimodal Related Research.
Awesome-Referring-Image-Segmentation
:books: A collection of papers about Referring Image Segmentation.
GPT4Tools-yr
GPT4Tools is an intelligent system that can automatically deciding, controlling, and utilizing different visual foundation models, allowing the user to interact with images during a conversation.
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
SUIM
Semantic Segmentation of Underwater Imagery: Dataset and Benchmark. #IROS2020
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.