Pinned Repositories
Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
bertviz
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
lmms-eval
Accelerating the development of large multimodal models (LMMs) with lmms-eval
Monkey
Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models
segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Transformer-Explainability
[CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize classifications by Transformer based networks.
VL-InterpreT
Visual Language Transformer Interpreter - An interactive visualization tool for interpreting vision-language transformers
VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 30+ HF models, 15+ benchmarks
LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
echo840's Repositories
echo840/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
echo840/bertviz
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
echo840/lmms-eval
Accelerating the development of large multimodal models (LMMs) with lmms-eval
echo840/Monkey
Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models
echo840/segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
echo840/Transformer-Explainability
[CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize classifications by Transformer based networks.
echo840/VL-InterpreT
Visual Language Transformer Interpreter - An interactive visualization tool for interpreting vision-language transformers
echo840/VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 30+ HF models, 15+ benchmarks