rujiawang329's Stars
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
mistralai/mistral-inference
Official inference library for Mistral models
QwenLM/Qwen-Agent
Agent framework and applications built upon Qwen>=2.0, featuring Function Calling, Code Interpreter, RAG, and Chrome extension.
allenai/OLMo
Modeling, training, eval, and inference code for OLMo
langchain-ai/open-canvas
📃 A better UX for chat, writing content, and coding with LLMs.
LLaVA-VL/LLaVA-NeXT
jingyi0000/VLM_survey
Collection of AWESOME vision-language models for vision tasks
real-stanford/diffusion_policy
[RSS 2023] Diffusion Policy Visuomotor Policy Learning via Action Diffusion
InternLM/Tutorial
LLM&VLM Tutorial
AgibotTech/agibot_x1_infer
The inference module for AgiBot X1.
merveenoyan/smol-vision
Recipes for shrinking, optimizing, customizing cutting edge vision models. 💜
allenai/dolma
Data and tools for generating and inspecting OLMo pre-training data.
OpenDriveLab/DriveLM
[ECCV 2024 Oral] DriveLM: Driving with Graph Visual Question Answering
mbzuai-oryx/LLaVA-pp
🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)
AgibotTech/agibot_x1_hardware
The hardware design for AgiBot X1.
zubair-irshad/Awesome-Robotics-3D
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
gokayfem/awesome-vlm-architectures
Famous Vision Language Models and Their Architectures
allenai/OLMoE
OLMoE: Open Mixture-of-Experts Language Models
zhengli97/Awesome-Prompt-Adapter-Learning-for-VLMs
A curated list of awesome prompt/adapter learning methods for vision-language models like CLIP.
123penny123/Awesome-LLM-RL
A comprehensive list of PAPERS, CODEBASES, and, DATASETS on Decision Making using Foundation Models including LLMs and VLMs.
allenai/OLMo-Eval
Evaluation suite for LLMs
hustvl/Senna
Bridging Large Vision-Language Models and End-to-End Autonomous Driving
IrohXu/Awesome-Multimodal-LLM-Autonomous-Driving
[WACV 2024 Survey Paper] Multimodal Large Language Models for Autonomous Driving
ge25nab/Awesome-VLM-AD-ITS
This repository collects research papers of large Vision Language Models in Autonomous driving and Intelligent Transportation System. The repository will be continuously updated to track the latest update.
MCZhi/DTPP
[ICRA 2024] Differentiable Joint Conditional Prediction and Cost Evaluation for Tree Policy Planning
AIR-THU/DAIR-V2X-Seq
UT-Austin-RPL/Coopernaut
Coopernaut: End-to-End Driving with Cooperative Perception for Networked Vehicles
DLUT-LYZ/CODA-LM
Official PyTorch implementation of CODA-LM(https://arxiv.org/abs/2404.10595)
mistralai/mistral-evals
AIR-THU/V2X-Graph