Source code for TOMM 2024 paper "MMICT: Boosting Multi-Modal Fine-Tuning with In-Context Examples" [arXiv preprint].
The required environment is included in requirements.txt
.
We train and test our model on:
To train the model:
bash run.sh
We thank the developers of LAVIS, BLIP-2, CLIP, for their public code release.