Haokun Wen1, Xian Zhang1, Xuemeng Song2*, Yinwei Wei3, Liqiang Nie1*
1 Harbin Institute of Technology (Shenzhen), Shenzhen, China
2 Shandong University, Qingdao, China
3 Monash University, Melbourne, Australia
* Corresponding authors
- Paper: ACM DL
- Checkpoints: Download Link
clip==0.2.0
matplotlib==3.5.1
numpy==1.22.3
pandas==1.4.2
Pillow==9.0.1 / 10.0.0
seaborn==0.12.2
torch==1.7.0
torchvision==0.8.0
tqdm==4.65.0
Spell-corrected data files for FashionIQ and Shoes are provided in correction_dict folder.
For CIRR, test results must be submitted through the official CIRR evaluation website. Submission files can be generated using:
python cirr_test_submission.pypython train.py --dataset 'fashioniq' --model_dir <output_dir> \
--mu_ 0.1 --nu_ 10 --kappa_ 0.5 --tau_ 0.1 --P 4 --Q 8python train.py --dataset 'shoes' --model_dir <output_dir> \
--mu_ 0.05 --nu_ 5 --kappa_ 0.5 --tau_ 0.1 --P 3 --Q 6python train.py --dataset 'cirr' --model_dir <output_dir> \
--mu_ 0.1 --nu_ 1 --kappa_ 0.1 --tau_ 0.05 --P 4 --Q 8If you find this work useful, please cite:
@inproceedings{wen2023target,
title={Target-Guided Composed Image Retrieval},
author={Wen, Haokun and Zhang, Xian and Song, Xuemeng and Wei, Yinwei and Nie, Liqiang},
booktitle={Proceedings of the ACM International Conference on Multimedia},
pages={915--923},
year={2023}
}This program is licensed under the GNU General Public License 3.0.
See https://www.gnu.org/licenses/gpl-3.0.html for details.
Any derivative work must also be licensed under the GNU GPL as published by the Free Software Foundation, either Version 3 or (at your option) any later version.