visual-instruction-tuning
There are 12 repositories under visual-instruction-tuning topic.
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
CircleRadon/Osprey
[CVPR2024] The code for "Osprey: Pixel Understanding with Visual Instruction Tuning"
ictnlp/LLaVA-Mini
LLaVA-Mini is a unified large multimodal model (LMM) that can support the understanding of images, high-resolution images, and videos in an efficient manner.
zjysteven/lmms-finetune
A minimal codebase for finetuning large multimodal models, supporting llava-1.5/1.6, llava-interleave, llava-next-video, llava-onevision, llama-3.2-vision, qwen-vl, qwen2-vl, phi3-v etc.
BAAI-DCAI/DataOptim
A collection of visual instruction tuning datasets.
ChenDelong1999/polite-flamingo
𦩠Visual Instruction Tuning with Polite Flamingo - training multi-modal LLMs to be both clever and polite! (AAAI-24 Oral)
fraction-ai/GAP
Gamified Adversarial Prompting (GAP): Crowdsourcing AI-weakness-targeting data through gamification. Boost model performance with community-driven, strategic data collection
bigai-nlco/VideoTGB
[EMNLP 2024] A Video Chat Agent with Temporal Prior
hllj/Vistral-V
Vistral-V: Visual Instruction Tuning for Vistral - Vietnamese Large Vision-Language Model.
zjr2000/REVERIE
[ECCV2024] Reflective Instruction Tuning: Mitigating Hallucinations in Large Vision-Language Models
jingyi0000/Awesome-Visual-Instruction-Tuning
Visual Instruction Tuning towards General-Purpose Multimodal Model: A Survey
yueying-teng/generate-language-image-instruction-following-data
Mistral assisted visual instruction data generation by following LLaVA