vision-language-transformer
There are 16 repositories under vision-language-transformer topic.
salesforce/LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
salesforce/BLIP
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
AlibabaResearch/AdvancedLiterateMachinery
A collection of original, innovative ideas and algorithms towards Advanced Literate Machinery. This project is maintained by the OCR Team in the Language Technology Lab, Tongyi Lab, Alibaba Group.
henghuiding/ReLA
[CVPR2023 Highlight] GRES: Generalized Referring Expression Segmentation
shenyunhang/APE
[CVPR 2024] Aligning and Prompting Everything All at Once for Universal Visual Perception
henghuiding/Vision-Language-Transformer
[ICCV2021 & TPAMI2023] Vision-Language Transformer and Query Generation for Referring Segmentation
sdc17/UPop
[ICML 2023] UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers.
haoliuhl/instructrl
Instruction Following Agents with Multimodal Transforemrs
sdc17/CrossGET
[ICML 2024] CrossGET: Cross-Guided Ensemble of Tokens for Accelerating Vision-Language Transformers.
sMamooler/CLIP_Explainability
code for studying OpenAI's CLIP explainability
yiren-jian/BLIText
[NeurIPS 2023] Bootstrapping Vision-Language Learning with Decoupled Language Pre-training
unitaryai/VTC
VTC: Improving Video-Text Retrieval with User Comments
marialymperaiou/knowledge-enhanced-multimodal-learning
A list of research papers on knowledge-enhanced multimodal learning
aurooj/VLM_SS
Mini-batch selective sampling for knowledge adaption of VLMs for mammography.
atharva-naik/MMML-TermProject-VizWiz-VQA-Challenge
VizWiz Challenge Term Project for Multi Modal Machine Learning @ CMU (11777)