Issues
- 1
How to infer captions on my own images
#31 opened by victorup - 1
- 1
Where can I found annotations for SNLI-VE?
#32 opened by 1219521375 - 0
- 1
- 1
Checkpoint for GQA model
#12 opened by aurooj - 0
pretrained weigths for VQA
#34 opened by guanhdrmq - 2
Checkpoint for SNLI-VE
#26 opened by sramshetty - 1
Pythia Feature Extraction
#28 opened by shamanthak-hegde - 1
Data dir for mcan_clip_grid_feature.py
#24 opened by Fly2flies - 0
- 2
About precompute
#25 opened by StylesZhang - 3
- 1
- 1
- 11
Captioning model training script fails
#2 opened by j-min - 4
- 1
The clip_feature
#19 opened by Timon0327 - 1
CLIP-VIT-B-Transformer captioning results
#20 opened by YuanEZhou - 2
About the training time of Pythia
#15 opened by tingxueronghua - 1
Train with a single GPU
#16 opened by ruinianxu - 2
- 2
- 1
Missed Link
#11 opened by jdiazram - 1
configuration file for CLIP-Res50x4
#10 opened by itsyoavshalev - 1
evaluating vqa using pythia
#9 opened by itsyoavshalev - 1
Grad-CAM visualization code
#7 opened by yangbang18 - 0
Grad-CAM visualization code
#8 opened by yangbang18 - 1
Pretrained weights for image captioning
#6 opened by zhuang93 - 1
How to combine CLIP with Oscacr(or VinVL)?
#4 opened by 594422814 - 1
About clip feature extraction
#5 opened by LittleDonkey1203 - 0
- 1
MS COCO Caption scores with MLE objective
#1 opened by j-min