nightlight9013's Stars
ytongbai/LVM
lucidrains/vector-quantize-pytorch
Vector (and Scalar) Quantization, in Pytorch
Qiyuan-Ge/PaintMind
Fast and controllable text-to-image model.
yuhangzang/ContextDET
Contextual Object Detection with Multimodal Large Language Models
flyywh/VCM_resources
MarkMoHR/Awesome-Referring-Image-Segmentation
:books: A collection of papers about Referring Image Segmentation.
MzeroMiko/VMamba
VMamba: Visual State Space Models,code is based on mamba
dome272/VQGAN-pytorch
Pytorch implementation of VQGAN (Taming Transformers for High-Resolution Image Synthesis) (https://arxiv.org/pdf/2012.09841.pdf)
vvvm23/vqvae-2
PyTorch implementation of VQ-VAE-2 from "Generating Diverse High-Fidelity Images with VQ-VAE-2"
thuanz123/enhancing-transformers
An unofficial implementation of both ViT-VQGAN and RQ-VAE in Pytorch
lucidrains/magvit2-pytorch
Implementation of MagViT2 Tokenizer in Pytorch
HengLan/VastTrack
VastTrack: Vast Category Visual Object Tracking
Event-AHU/Mamba_State_Space_Model_Paper_List
[Mamba-Survey-2024] Paper list for State-Space-Model/Mamba and it's Applications
Sygil-Dev/muse-maskgit-pytorch
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
huggingface/open-muse
Open reproduction of MUSE for fast text2image generation.
bo-miao/SgMg
[ICCV 2023] Spectrum-guided Multi-granularity Referring Video Object Segmentation.
pengzhiliang/G2SD
ihaeyong/WSN
Winning SubNetwork (WSN)
luanyunteng/pytorch-be-your-own-teacher
A pytorch implementation of paper 'Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation', https://arxiv.org/abs/1905.08094
Ashespt/AdvBCT
The official implementation of AdvBCT
google-research/google-research
Google Research
HobbitLong/SupContrast
PyTorch implementation of "Supervised Contrastive Learning" (and SimCLR incidentally)
pengsida/learning_research
本人的科研经验
microsoft/VideoX
VideoX: a collection of video cross-modal models
bhpfelix/segment-anything-finetuner
Simple Finetuning Starter Code for Segment Anything
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
Vision-Intelligence-and-Robots-Group/count-anything
an empirical study on few-shot counting using segment anything (SAM)
czczup/ViT-Adapter
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions
gaomingqi/Track-Anything
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
jhgan00/image-retrieval-transformers
(Unofficial) PyTorch implementation of Training Vision Transformers for Image Retrieval(El-Nouby, Alaaeldin, et al. 2021).