[CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts
Exploring Visual Prompts for Adapting Large-Scale Models
#计算机科学#[TPAMI] Searching prompt modules for parameter-efficient transfer learning.
[ICCV 2023 & AAAI 2023] Binary Adapters & FacT, [Tech report] Convpass
[NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"
Official implementation for CVPR'23 paper "BlackVIP: Black-Box Visual Prompting for Robust Transfer Learning"
👀 Visual Instruction Inversion: Image Editing via Visual Prompting (NeurIPS 2023)
[ECCV 2024] API: Attention Prompting on Image for Large Vision-Language Models
[CVPR 2023] VoP: Text-Video Co-operative Prompt Tuning for Cross-Modal Retrieval
[ICLR24] AutoVP: An Automated Visual Prompting Framework and Benchmark
#计算机科学#[ICML 2024] Visual-Text Cross Alignment: Refining the Similarity Score in Vision-Language Models
A simple GUI for experimenting with visual prompting
[IEEE BigData'24] Code used in Paper "Benchmarking Human and Automated Prompting in the Segment Anything Model"
These notes and resources are compiled from the crash course Prompt Engineering for Vision Models offered by DeepLearning.AI.