gqa
There are 14 repositories under gqa topic.
bknyaz/sgg
Train Scene Graph Generation for Visual Genome and GQA in PyTorch >= 1.2 with improved zero and few-shot generalization.
Bruce-Lee-LY/decoding_attention
Decoding Attention is specially optimized for MHA, MQA, GQA and MLA using CUDA core for the decoding stage of LLM inference.
phiyodr/vqaloader
PyTorch DataLoader for many VQA datasets
DigitalPhonetics/Intrinsic-Subgraph-Generation-for-VQA
Predicting a subgraph alongside the answer in a graph based VQA model
haukzero/from-mha-to-mla
MHA, MQA, GQA, MLA 相关原理及简要实现
leaderj1001/Vision-Language
Vision-Language, Solve GQA(Visual Reasoning in the Real World) dataset.
ycchen218/VisionQA-Llama2-OWLViT
This is a multimodal model design for the Vision Question Answering (VQA) task. It integrates the Llama2 13B, OWL-ViT, and YOLOv8 models.
ExplainableML/ZS-A2T
[GCPR 2023] Zero-shot Translation of Attention Patterns in VQA Models to Natural Language
sahasourav17/IntelliAnswer
A RAG-based question-answering system that processes user queries using local documents. It extracts relevant information to answer questions, falling back to a large language model when local sources are insufficient, ensuring accurate and contextual responses.
alexmirrington/honours-thesis
LaTeX files for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
alexmirrington/gat-vqa
Source code for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
eltoto1219/vltk
A toolkit for vision-language processing to support the increasing popularity of mulit-modal transformer-based models
NMPoole/CS5014-MLVisualAttributes
Case study of multi-layer perceptron and random forest techniques as applied to a subset of the GQA dataset.
sushantkumar23/baby-llama
Simple Llama architecture LLM in pytorch