Pinned Repositories
LLaMA2-Accessory
An Open-source Toolkit for LLM Development
LLaVA-NeXT
LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
I2P-MAE
[CVPR 2023] Learning 3D Representations from 2D Pre-trained Models via Image-to-Point Masked Autoencoders
MathVerse
[ECCV 2024] Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?
MonoDETR
[ICCV 2023] The first DETR model for monocular 3D object detection with depth-guided transformer
Personalize-SAM
Personalize Segment Anything Model (SAM) with 1 shot in 10 seconds
Point-M2AE
[NeurIPS 2022] Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training
Point-NN
[CVPR 2023] Parameter is Not All You Need: Starting from Non-Parametric Networks for 3D Point Cloud Analysis
PointCLIP
[CVPR 2022] PointCLIP: Point Cloud Understanding by CLIP
ZrrSkywalker's Repositories
ZrrSkywalker/Personalize-SAM
Personalize Segment Anything Model (SAM) with 1 shot in 10 seconds
ZrrSkywalker/Point-NN
[CVPR 2023] Parameter is Not All You Need: Starting from Non-Parametric Networks for 3D Point Cloud Analysis
ZrrSkywalker/MonoDETR
[ICCV 2023] The first DETR model for monocular 3D object detection with depth-guided transformer
ZrrSkywalker/PointCLIP
[CVPR 2022] PointCLIP: Point Cloud Understanding by CLIP
ZrrSkywalker/I2P-MAE
[CVPR 2023] Learning 3D Representations from 2D Pre-trained Models via Image-to-Point Masked Autoencoders
ZrrSkywalker/Point-M2AE
[NeurIPS 2022] Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training
ZrrSkywalker/MathVerse
[ECCV 2024] Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?
ZrrSkywalker/MAVIS
Mathematical Visual Instruction Tuning for Multi-modal Large Language Models
ZrrSkywalker/LLaMA-Adapter
Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
ZrrSkywalker/CaFo
[CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners
ZrrSkywalker/MonoDETR-MV
The multi-view version of MonoDETR on nuScenes dataset
ZrrSkywalker/ZrrSkywalker
ZrrSkywalker/CALIP
Enhancing Zero-shot CLIP with Cross-Modality Attention
ZrrSkywalker/CLIP-Adapter
ZrrSkywalker/Tip-Adapter
ZrrSkywalker/MathVista
MathVista: data, code, and evaluation for Mathematical Reasoning in Visual Contexts
ZrrSkywalker/ZrrSkywalker.github.io