Wykay's Stars
meta-llama/llama3
The official Meta Llama 3 GitHub site
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
state-spaces/mamba
Mamba SSM architecture
facebookresearch/segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
datawhalechina/self-llm
《开源大模型食用指南》基于Linux环境快速部署开源大模型,更适合中国宝宝的部署教程
01-ai/Yi
A series of large language models trained from scratch by developers @01-ai
huggingface/lerobot
🤗 LeRobot: Making AI for Robotics more accessible with end-to-end learning
QwenLM/Qwen-VL
The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
modelscope/ms-swift
Use PEFT or Full-parameter to finetune 400+ LLMs or 100+ MLLMs. (LLM: Qwen2.5, Llama3.2, GLM4, Internlm2.5, Yi1.5, Mistral, Baichuan2, DeepSeek, Gemma2, ...; MLLM: Qwen2-VL, Qwen2-Audio, Llama3.2-Vision, Llava, InternVL2, MiniCPM-V-2.6, GLM4v, Xcomposer2.5, Yi-VL, DeepSeek-VL, Phi3.5-Vision, ...)
xinyu1205/recognize-anything
Open-source and strong foundation image recognition models.
refuel-ai/autolabel
Label, clean and enrich text datasets with LLMs.
dvlab-research/LISA
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
AnswerDotAI/fsdp_qlora
Training LLMs with QLoRA + FSDP
Farama-Foundation/D4RL
A collection of reference environments for offline reinforcement learning
openvla/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
mlfoundations/dclm
DataComp for Language Models
google-deepmind/open_x_embodiment
penghao-wu/vstar
PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"
peteanderson80/Matterport3DSimulator
AI Research Platform for Reinforcement Learning from Real Panoramic Images.
mlfoundations/open_lm
A repository for research on medium sized language models.
TRI-ML/prismatic-vlms
A flexible and efficient codebase for training visually-conditioned language models (VLMs)
wayveai/Driving-with-LLMs
PyTorch implementation for the paper "Driving with LLMs: Fusing Object-Level Vector Modality for Explainable Autonomous Driving"
pacman100/LLM-Workshop
LLM Workshop by Sourab Mangrulkar
NVlabs/OmniDrive
LostXine/LLaRA
LLaRA: Large Language and Robotics Assistant
OpenDriveLab/ELM
[ECCV 2024] Embodied Understanding of Driving Scenarios
UT-Austin-RPL/VIOLA
Official implementation for VIOLA
opendilab/OpenPaL
Building open-ended embodied agent in battle royale FPS game
sreeramsa/DriveSim
Simulator designed to generate diverse driving scenarios.
NMS05/DinoV2-SigLIP-Phi3-LoRA-VLM