nuonepeaceyy's Stars
TheAlgorithms/Python
All Algorithms implemented in Python
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
OpenBMB/MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
mrdbourke/pytorch-deep-learning
Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.
Visualize-ML/Book4_Power-of-Matrix
Book_4_《矩阵力量》 | 鸢尾花书:从加减乘除到机器学习;上架!
open-mmlab/mmdetection3d
OpenMMLab's next-generation platform for general 3D object detection.
OpenDriveLab/UniAD
[CVPR 2023 Best Paper Award] Planning-oriented Autonomous Driving
invictus717/MetaTransformer
Meta-Transformer for Unified Multimodal Learning
ActiveVisionLab/Awesome-LLM-3D
Awesome-LLM-3D: a curated list of Multi-modal Large Language Model in 3D world Resources
UMass-Foundation-Model/3D-LLM
Code for 3D-LLM: Injecting the 3D World into Large Language Models
megvii-research/PETR
[ECCV2022] PETR: Position Embedding Transformation for Multi-View 3D Object Detection & [ICCV2023] PETRv2: A Unified Framework for 3D Perception from Multi-Camera Images
youquanl/Segment-Any-Point-Cloud
[NeurIPS'23 Spotlight] Segment Any Point Cloud Sequences by Distilling Vision Foundation Models
vasgaowei/BEV-Perception
Bird's Eye View Perception
hailanyi/3D-Detection-Tracking-Viewer
3D detection and tracking viewer (visualization) for kitti & waymo dataset
Haiyang-W/UniTR
[ICCV2023] Official Implementation of "UniTR: A Unified and Efficient Multi-Modal Transformer for Bird’s-Eye-View Representation"
IrohXu/Awesome-Multimodal-LLM-Autonomous-Driving
[WACV 2024 Survey Paper] Multimodal Large Language Models for Autonomous Driving
xk-huang/segment-caption-anything
[CVPR 24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloading the trained model checkpoints, and example notebooks / gradio demo that show how to use the model.
runnanchen/CLIP2Scene
wudongming97/Prompt4Driving
[AAAI2025] Language Prompt for Autonomous Driving
jxbbb/TOD3Cap
[ECCV 2024] TOD3Cap: Towards 3D Dense Captioning in Outdoor Scenes
yunlu-chen/PointMixup
Implementation for paper "PointMixup: Augmentation for Point Cloud". Accepted to ECCV 2020 as spotlight presentation
YurongYou/MODEST
Code release for "Learning to Detect Mobile Objects from LiDAR Scans Without Labels" [CVPR 2022]
hailanyi/CPD
Commonsense Prototype for Outdoor Unsupervised 3D Object Detection (CVPR 2024)
xmed-lab/NuInstruct
xmuqimingxia/CoIn
CoIn: Contrastive Instance Feature Mining for Outdoor 3D Object Detection with Very Limited Annotations(ICCV2023)
xmuqimingxia/3D-HANet
3D HANet: A Flexible 3D Heatmap Auxiliary Network for Object Detection(TGRS)
IcebergKnight/NMS-Loss
xmuqimingxia/HINTED
HINTED: Hard Instance Enhanced Detector with Mixed-Density Feature Fusion for Sparsely-Supervised 3D Object Detection(CVPR 2024)
zhouqiu/SOGDet
sogdet