linzhiqiu's repositories

cross_modal_adaptation

Cross-modal few-shot adaptation with CLIP

Language:PythonLicense:MITStargazers:341Issues:9Issues:22

t2v_metrics

Evaluating text-to-image/video/3D models with VQAScore

Language:PythonLicense:Apache-2.0Stargazers:280Issues:15Issues:15

visual_gpt_score

VisualGPTScore for visio-linguistic reasoning

CLIP-FlanT5

Training code for CLIP-FlanT5

Language:PythonLicense:Apache-2.0Stargazers:26Issues:2Issues:3
Language:Jupyter NotebookStargazers:14Issues:5Issues:0

leco

Learning with Ever-Changing Ontology

Language:Jupyter NotebookStargazers:3Issues:2Issues:0

vl_finetuning

Few-shot Finetuning of CLIP

Language:PythonLicense:MITStargazers:1Issues:3Issues:0

avalanche

Avalanche: an End-to-End Library for Continual Learning.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
License:MITStargazers:0Issues:2Issues:0
Language:HTMLStargazers:0Issues:1Issues:0
Language:PythonStargazers:0Issues:2Issues:0
Language:HTMLStargazers:0Issues:1Issues:0

debiased-pseudo-labeling

[CVPR 2022] Debiased Learning from Naturally Imbalanced Pseudo-Labels

Language:Jupyter NotebookLicense:MITStargazers:0Issues:1Issues:0

dino

PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

HRNet-Semantic-Segmentation

The OCR approach is rephrased as Segmentation Transformer: https://arxiv.org/abs/1909.11065. This is an official implementation of semantic segmentation for HRNet. https://arxiv.org/abs/1908.07919

Language:PythonLicense:NOASSERTIONStargazers:0Issues:1Issues:0

HTML4Vision

A simple HTML visualization tool for computer vision research :hammer_and_wrench:

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

linzhiqiu.github.io

Zhiqiu Lin's site

Language:JavaScriptLicense:MITStargazers:0Issues:1Issues:0

LLaVA

[NeurIPS 2023 Oral] Visual Instruction Tuning: LLaVA (Large Language-and-Vision Assistant) built towards GPT-4V level capabilities.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0
Language:JavaScriptStargazers:0Issues:0Issues:0

lmms-eval

Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.

License:NOASSERTIONStargazers:0Issues:0Issues:0

mmselfsup

OpenMMLab Self-Supervised Learning Toolbox and Benchmark

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

PerceptualSimilarity

LPIPS metric. pip install lpips

Language:PythonLicense:BSD-2-ClauseStargazers:0Issues:1Issues:0

pytorchvideo

A deep learning library for video understanding research.

License:Apache-2.0Stargazers:0Issues:0Issues:0

streamlit-feedback-video

Collect user feedback from within your Streamlit app

License:MITStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

video_annotation

Video Annotation Format

Language:PythonStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0

vision-language-models-are-bows

Experiments and data for the paper "When and why vision-language models behave like bags-of-words, and what to do about it?" Oral @ ICLR 2023

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

why-winoground-hard

Code for 'Why is Winoground Hard? Investigating Failures in Visuolinguistic Compositionality', EMNLP 2022

Language:PythonLicense:MITStargazers:0Issues:1Issues:0