WingkeungM
Remote Sensing,Computer Vision Postdoc Fellow @ Tsinghua University
WHU -> UCAS -> THUBeijing, China
WingkeungM's Stars
NanmiCoder/MediaCrawler
小红书笔记 | 评论爬虫、抖音视频 | 评论爬虫、快手视频 | 评论爬虫、B 站视频 | 评论爬虫、微博帖子 | 评论爬虫、百度贴吧帖子 | 百度贴吧评论回复爬虫 | 知乎问答文章|评论爬虫
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
HumanAIGC/EMO
Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
lllyasviel/Omost
Your image is almost there!
LiheYoung/Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
FoundationVision/VAR
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
pytorch-labs/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
dvlab-research/MGM
Official repo for "Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models"
MzeroMiko/VMamba
VMamba: Visual State Space Models,code is based on mamba
yuweihao/MambaOut
MambaOut: Do We Really Need Mamba for Vision?
Ucas-HaoranWei/Vary
[ECCV 2024] Official code implementation of Vary: Scaling Up the Vision Vocabulary of Large Vision Language Models.
yyyujintang/Awesome-Mamba-Papers
Awesome Papers related to Mamba.
mini-sora/minisora
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
Jack-bo1220/Awesome-Remote-Sensing-Foundation-Models
NAOSI-DLUT/Campus2024
2024届互联网校招信息汇总
Ucas-HaoranWei/Vary-toy
Official code implementation of Vary-toy (Small Language Model Meets with Reinforced Vision Vocabulary)
youquanl/Segment-Any-Point-Cloud
[NeurIPS'23 Spotlight] Segment Any Point Cloud Sequences by Distilling Vision Foundation Models
magic-research/Dataset_Quantization
[ICCV2023] Dataset Quantization
rsdler/Remote-Sensing-in-CVPR2024
Papers related to remote sensing in CVPR 2024
aim-uofa/SegPrompt
Official Implementation of ICCV 2023 Paper - SegPrompt: Boosting Open-World Segmentation via Category-level Prompt Learning
Ucas-HaoranWei/Vary-tiny-600k
Vary-tiny codebase upon LAVIS (for training from scratch)and a PDF image-text pairs data (about 600k including English/Chinese)
T1aNS1R/Evil-Geniuses
Ucas-HaoranWei/Vary-family
fullcyxuc/B-Seg
Code for our SIGGRAPH'2023 paper: "UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building Instance Segmentation"
canoe-Z/PETDet
[TGRS2023] Official implement for PETDet.
MAVREC/mavrec-code
This code is provided for reproducibility of results in the paper: Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve Aerial Visual Perception?
chunbolang/RARE
Official PyTorch Implementation of Retain and Recover: Delving into Information Loss for Few-Shot Segmentation (TIP'23).
StarBurstStream0/SIRS
Official code for "SIRS: Multi-task Joint Learning for Remote Sensing Foreground-entity Image-text Retrieval" TGRS2024
Na-Z/LIOND
[AAAI 2024] Robust Visual Recognition with Class-Imbalanced Open-World Noisy Data