There are 3 repositories under visual-prompting topic.
streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VL
[CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts
Exploring Visual Prompts for Adapting Large-Scale Models
[TPAMI] Searching prompt modules for parameter-efficient transfer learning.
[NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"
Official implementation for CVPR'23 paper "BlackVIP: Black-Box Visual Prompting for Robust Transfer Learning"
đź‘€ Visual Instruction Inversion: Image Editing via Visual Prompting (NeurIPS 2023)
[CVPR 2023] VoP: Text-Video Co-operative Prompt Tuning for Cross-Modal Retrieval
[arXiv] "Uncovering the Hidden Cost of Model Compression" by Diganta Misra, Agam Goyal, Bharat Runwal, and Pin-Yu Chen
These notes and resources are compiled from the crash course Prompt Engineering for Vision Models offered by DeepLearning.AI.