yuanpf123
My name is Pengfei Yuan,a postgraduate in the School of Computer Science and Information Engineering, Hefei University of Technology.
No. 485, Danxia Road, Shushan District, Hefei (Feicuihu Campus of Hefei University of Technology)
yuanpf123's Stars
lzw-lzw/awesome-remote-sensing-vision-language-models
Awesome-Remote-Sensing-Vision-Language-Models
jaychempan/PIR-CLIP
📖 Official Code for “PIR-CLIP: Remote Sensing Image-text Retrieval with Prior Instruction Representation Learning”
gaopengcuhk/CLIP-Adapter
jeonsworld/ViT-pytorch
Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)
KMnP/vpt
❄️🔥 Visual Prompt Tuning [ECCV 2022] https://arxiv.org/abs/2203.12119
muzairkhattak/multimodal-prompt-learning
[CVPR 2023] Official repository of paper titled "MaPLe: Multi-modal Prompt Learning".
wangzhecheng/SkyScript
Official repo for "SkyScript: A Large and Semantically Diverse Vision-Language Dataset for Remote Sensing"
microsoft/Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
DmitryRyumin/AAAI-2024-Papers
AAAI 2024 Papers: Explore a comprehensive collection of innovative research papers presented at one of the premier artificial intelligence conferences. Seamlessly integrate code implementations for better understanding. ⭐ experience the forefront of progress in artificial intelligence with this repository!
shuanglinyan/CFine
CLIP-Driven Fine-grained Text-Image Person Re-identification
gaopengcuhk/Tip-Adapter
strawberrypie/bert_adapter
Implementation of the paper Parameter-Efficient Transfer Learning for NLP, Houlsby [Google], 2019. Published in ICML 2019.
KaiyangZhou/Dassl.pytorch
A PyTorch toolbox for domain generalization, domain adaptation and semi-supervised learning.
xiaoyuan1996/GaLR
Source code of paper "Remote Sensing Cross-Modal Image-Text Retrieval Based on Global and Local Information"
OpenGVLab/CaFo
[CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners
Sha-Lab/FEAT
The code repository for "Few-Shot Learning via Embedding Adaptation with Set-to-Set Functions"
adapter-hub/adapters
A Unified Library for Parameter-Efficient and Modular Transfer Learning
ZiyuGuo99/CALIP
[AAAI 2023] Zero-Shot Enhancement of CLIP with Parameter-free Attention
aneeshan95/Sketch_LVM
Project page for the paper 'CLIP for All Things Zero-Shot Sketch-Based Image Retrieval, Fine-Grained or Not'
songrise/CLIP-Count
[ACM MM23] CLIP-Count: Towards Text-Guided Zero-Shot Object Counting
om-ai-lab/RS5M
RS5M: a large-scale vision language dataset for remote sensing [TGRS]
om-ai-lab/awesome-RSVLM
Collection of Remote Sensing Vision-Language Models
salesforce/BLIP
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
tianyizhou/DIHCL
Curriculum Learning by Dynamic Instance Hardness (NeurIPS 2020)
geoaigroup/awesome-vision-language-models-for-earth-observation
A curated list of awesome vision and language resources for earth observation.
GuyHacohen/curriculum_learning
Code implementing the experiments described in the paper "On The Power of Curriculum Learning in Training Deep Networks" by Hacohen & Weinshall (ICML 2019)
LeapLabTHU/EfficientTrain
1.5−3.0× lossless training or pre-training speedup. An off-the-shelf, easy-to-implement algorithm for the efficient training of foundation visual backbones.
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
nazmul-karim170/C-SFDA_Source-Free-Domain-Adaptation
[CVPR 2023] Official Implementation of "C-SFDA: A Curriculum Learning Aided Self-Training Framework for Efficient Source Free Domain Adaptation""
microsoft/Oscar
Oscar and VinVL