h0jicha's Stars
sindresorhus/awesome
😎 Awesome lists about all kinds of interesting topics
microsoft/autogen
A programming framework for agentic AI 🤖
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
SakanaAI/AI-Scientist
The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑🔬
LiheYoung/Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
promptfoo/promptfoo
Test your prompts, agents, and RAGs. Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.
openai/human-eval
Code for the paper "Evaluating Large Language Models Trained on Code"
prs-eth/Marigold
[CVPR 2024 - Oral, Best Paper Award Candidate] Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation
google/style-aligned
Official code for "Style Aligned Image Generation via Shared Attention"
siyuanliii/masa
Official Implementation of CVPR24 highligt paper: Matching Anything by Segmenting Anything
Thinklab-SJTU/Awesome-LLM4AD
A curated list of awesome LLM for Autonomous Driving resources (continually updated)
DirtyHarryLYL/LLM-in-Vision
Recent LLM-based CV and related works. Welcome to comment/contribute!
liliu-avril/Awesome-Segment-Anything
This repository is for the first comprehensive survey on Meta AI's Segment Anything Model (SAM).
jy0205/LaVIT
LaVIT: Empower the Large Language Model to Understand and Generate Visual Content
shenyunhang/APE
[CVPR 2024] Aligning and Prompting Everything All at Once for Universal Visual Perception
LMD0311/Awesome-World-Model
Collect some World Models for Autonomous Driving papers.
allenai/reward-bench
RewardBench: the first evaluation tool for reward models.
WooooDyy/AgentGym
Code and implementations for the paper "AgentGym: Evolving Large Language Model-based Agents across Diverse Environments" by Zhiheng Xi et al.
SHI-Labs/VCoder
VCoder: Versatile Vision Encoders for Multimodal Large Language Models, arXiv 2023 / CVPR 2024
NVlabs/OmniDrive
ali-vilab/Ranni
MILVLG/imp
a family of highly capabale yet efficient large multimodal models
jamesjg/FoodSAM
FoodSAM: Any Food Segmentation
JiuTian-VL/JiuTian-LION
[CVPR 2024] LION: Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge
THUDM/VisualAgentBench
Towards Large Multimodal Models as Visual Foundation Agents
reddy-lab-code-research/PPOCoder
Code for the TMLR 2023 paper "PPOCoder: Execution-based Code Generation using Deep Reinforcement Learning"
algomatic-inc/awesome-ai-agents-guide
🤖 A collection of AI agents includes research papers, blogs, and products focused on developing autonomous systems.
samschulter/omnilabeltools
A Python toolkit for the OmniLabel benchmark providing code for evaluation and visualization
YuehaoYin/FoodLMM
nejumi/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.