CLIP-ReID icon indicating copy to clipboard operation
CLIP-ReID copied to clipboard

Official implementation for "CLIP-ReID: Exploiting Vision-Language Model for Image Re-identification without Concrete Text Labels" (AAAI 2023)

Results 35 CLIP-ReID issues
Sort by recently updated
recently updated
newest added

纯小白,我在train的过程中为什么在我的out put dir中只生成了train_log.txt训练的模型本身去了哪里?

Hi @Syliz517 , Thanks for sharing your great work. I have downloaded some of pretrained model weights and evaluated them. On Market1501 and Duke, the results are the same as...

Hi @Syliz517 , First, thank for your great work. One point I just would like to ask you regarding the training data of `MSMT17`, it shows that both `train` and...

First of all, thank you very much for your contribution in the field of re-identification! I had some problems when using your model. When I read the vehicle model into...

![image](https://github.com/Syliz517/CLIP-ReID/assets/140691950/52258522-9b52-4fc9-9a96-8a23efdea7c3) I don't know where this 'your_trained_checkpoints_path/Vit-B-16_60.pth' is stored, can someones help me , thank you so much.

Hello Author, Thank you for your work on Clip-Reid. I'm facing some confusion regarding the position embedding adaptation of Vision Transformer (ViT) backbone when altering the resolution and the correct...

在代码中,第一阶段的训练中image encoder是冻结的,可学习的text tokens和和text encoder是可学习的。这和论文里描述的只有text tokens是可学习的,image encoder和text encoder是冻结的不匹配呀。

Why apply triplet loss to img_feature_last? here, img_feature_last is the output of the second-to-last module of the ViT model.

Hi there, is there a way to train with a .csv file that has image paths and text descriptions?