JMcarrot's Stars
ai-dawang/PlugNPlay-Modules
mit-han-lab/efficientvit
EfficientViT is a new family of vision models for efficient high-resolution vision.
PKINet/PKINet
Official implementation of CVPR2024 Paper "Poly Kernel Inception Network for Remote Sensing Detection".
hzwer/WritingAIPaper
Writing AI Conference Papers: A Handbook for Beginners
Ruixxxx/Awesome-Vision-Mamba-Models
[Official Repo] A Survey on Vision Mamba: Models, Applications and Challenges
datawhalechina/tiny-universe
《大模型白盒子构建指南》:一个全手搓的Tiny-Universe
henghuiding/Vision-Language-Transformer
[ICCV2021 & TPAMI2023] Vision-Language Transformer and Query Generation for Referring Segmentation
awaisrauf/Awesome-CV-Foundational-Models
SkalskiP/awesome-foundation-and-multimodal-models
👁️ + 💬 + 🎧 = 🤖 Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]
xiuqhou/Relation-DETR
[ECCV2024 Oral] Official implementation of the paper "Relation DETR: Exploring Explicit Position Relation Prior for Object Detection"
ZhanYang-nwpu/Awesome-Remote-Sensing-Multimodal-Large-Language-Model
Multimodal Large Language Models for Remote Sensing (RS-MLLMs): A Survey
zytx121/Awesome-VLGFM
A Survey on Vision-Language Geo-Foundation Models (VLGFMs)
xg416/DATUM
Official repo for the Deep Atmospheric TUrbulence Mitigation network
vlislab22/AIAA-5027
linhuixiao/CLIP-VG
[TMM 2023] Self-paced Curriculum Adapting of CLIP for Visual Grounding.
impiga/Plain-DETR
[ICCV2023] DETR Doesn’t Need Multi-Scale or Locality Design
yuyongcan/Benchmark-TTA
kkakkkka/ETRIS
[ICCV-2023] The official code of Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
dongzhang89/FPT
Implementation for paper: Feature Pyramid Transformer
QY1994-0919/CFPNet
Centralized Feature Pyramid for Object Detection
duzw9311/CFPT
Cross-Layer Feature Pyramid Transformer for Small Object Detection in Aerial Images
wjiazheng/SwinPA-Net
SwinPA-Net: Swin Transformer-Based Multiscale Feature Pyramid Aggregation Network for Medical Image Segmentation
amazon-science/polygon-transformer
cmhungsteve/Awesome-Transformer-Attention
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
MenghaoGuo/Awesome-Vision-Attentions
Summary of related papers on visual attention. Related code will be released based on Jittor gradually.
TheShadow29/awesome-grounding
awesome grounding: A curated list of research papers in visual grounding
linhuixiao/Awesome-Visual-Grounding
A Survey on Visual Grounding
open-mmlab/mmcv
OpenMMLab Computer Vision Foundation
52CV/CVPR-2024-Papers
jingyi0000/VLM_survey
Collection of AWESOME vision-language models for vision tasks