This is a sub-repository of the HADA where we applied HADA on-the-top of LAVIS library.
Our paper has been accepted at ECIR'23
HADA is a framework that combines any pretrained SOTA models in image-text retrieval (ITR) to produce a better result. A special feature in HADA is that this framework only introduces a tiny number of additonal trainable parameters. Thus, it does not required multiple GPUs to train HADA or external large-scale dataset (although pretraining may further improve the performance).
In this sub-repo, we used HADA to combine 2 SOTA models including BLIP and CLIP from LAVIS library. The total recall was increase by 3.94% on the Flickr30k dataset and 1.49% on MSCOCO.
Please install LAVIS before running.
We used mlflow-ui to keep track the performance between configurations. Please modify or remove this related-part if you do not want to use.
We uploaded the pretrained models here.
Remember to update the path in the config files in Config folders. Then you can train or evaluate by the file main.py
# Train (2 steps, 1st step for initial learning and 2nd step for finetuning)
# Flickr
python main.py -cp Config/flickr30k/C1.yml -rm train
python main.py -cp Config/flickr30k/C1_cont.yml -rm train
# MSCOCO
python main.py -cp Config/mscoco/C1.yml -rm train
python main.py -cp Config/mscoco/C1_cont_Argment.yml -rm train
# Test (skip train if you download pretrained model)
# Flickr
python main.py -cp Config/flickr30k/C1_cont.yml -rm test
# MSCOCO
python main.py -cp Config/mscoco/C1_cont_Argment.yml -rm test
For any issue or comment, you can directly email me at manh.nguyen5@mail.dcu.ie