ZhenhuaZJim's Stars
Significant-Gravitas/AutoGPT
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
openai/whisper
Robust Speech Recognition via Large-Scale Weak Supervision
OpenInterpreter/open-interpreter
A natural language interface for computers
dair-ai/Prompt-Engineering-Guide
🐙 Guides, papers, lecture, notebooks and resources for prompt engineering
facebookresearch/segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
geekan/MetaGPT
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
karpathy/nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
microsoft/autogen
A programming framework for agentic AI 🤖 (PyPi: autogen-agentchat)
ultralytics/ultralytics
Ultralytics YOLO11 🚀
GitHubDaily/GitHubDaily
坚持分享 GitHub 上高质量、有趣实用的开源技术教程、开发者工具、编程网站、技术资讯。A list cool, interesting projects of GitHub.
meta-llama/llama3
The official Meta Llama 3 GitHub site
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
THU-MIG/yolov10
YOLOv10: Real-Time End-to-End Object Detection [NeurIPS 2024]
mistralai/mistral-inference
Official inference library for Mistral models
huggingface/lerobot
🤗 LeRobot: Making AI for Robotics more accessible with end-to-end learning
LiheYoung/Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
TimDettmers/bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
OpenDriveLab/UniAD
[CVPR 2023 Best Paper Award] Planning-oriented Autonomous Driving
xinyu1205/recognize-anything
Open-source and strong foundation image recognition models.
facebookresearch/ijepa
Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."
fudan-zvg/Semantic-Segment-Anything
Automated dense category annotation engine that serves as the initial semantic labeling for the Segment Anything dataset (SA-1B).
qianqianwang68/omnimotion
rmokady/CLIP_prefix_caption
Simple image captioning model
siyuanliii/masa
Official Implementation of CVPR24 highligt paper: Matching Anything by Segmenting Anything
dog-qiuqiu/FastestDet
:zap: A newly designed ultra lightweight anchor free target detection algorithm, weight only 250K parameters, reduces the time consumption by 10% compared with yolo-fastest, and the post-processing is simpler
lapisrocks/LanguageAgentTreeSearch
[ICML 2024] Official repository for "Language Agent Tree Search Unifies Reasoning Acting and Planning in Language Models"
NVlabs/genvs
facebookresearch/searchformer
Official codebase for the paper "Beyond A* Better Planning with Transformers via Search Dynamics Bootstrapping".
OpenMask3D/openmask3d