Xiao Feng Zhang's starred repositories
DroneVehicle
Drone-based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning
LLaVA-PruMerge
LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models
Prompt-Highlighter
[CVPR 2024] Prompt Highlighter: Interactive Control for Multi-Modal LLMs
sam-clip-segmentation
Image Instance Segmentation - Zero Shot - OpenAI's CLIP + Meta's SAM
Diff-Plugin
[CVPR 2024] Official code release of our paper "Diff-Plugin: Revitalizing Details for Diffusion-based Low-level tasks"
ResUNetFormer
This Keras code is for the paper A. Jamali, S. K. Roy, J. Li and P. Ghamisi, "[Neighborhood Attention Makes the Encoder of ResUNet Stronger for Accurate Road Extraction]," in IEEE Geoscience and Remote Sensing Letters, doi: 10.1109/LGRS.2024.3354560 [https://ieeexplore.ieee.org/document/10400502].
Awesome-MLLM-Hallucination
Papers about Hallucination in Multi-Modal Large Language Models (MLLMs)
Transformer-Explainability
[CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize classifications by Transformer based networks.
InstructIR
InstructIR: High-Quality Image Restoration Following Human Instructions https://huggingface.co/spaces/marcosv/InstructIR
Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
label-words-are-anchors
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
MechanisticProbe
Towards a Mechanistic Interpretation of Multi-Step Reasoning Capabilities of Language Models
Awesome-LLM-Reasoning
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought, Instruction-Tuning and Multimodality.
awesome-multimodal-in-medical-imaging
A collection of resources on applications of multi-modal learning in medical imaging.
KwaiAgents
A generalized information-seeking agent system with Large Language Models (LLMs).