22109095's Stars
HCPLab-SYSU/Embodied_AI_Paper_List
[Embodied-AI-Survey-2024] Paper list and projects for Embodied AI
JiazuoYu/Awesome-Prompt-Adapter-Learning-for-Vision-Language-Models
A curated list of prompt/adapter learning methods for vision-language models.
JiazuoYu/PathWeave
Code for paper "LLMs Can Evolve Continually on Modality for X-Modal Reasoning" NeurIPS2024
DAMO-NLP-SG/Video-LLaMA
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
yuweihao/MambaOut
MambaOut: Do We Really Need Mamba for Vision?
Paranioar/SGRAF
[AAAI2021] The code of “Similarity Reasoning and Filtration for Image-Text Matching”
Paranioar/RCAR
[TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”
Paranioar/UniPT
[CVPR2024] The code of "UniPT: Universal Parallel Tuning for Transfer Learning with Efficient Parameter and Memory"
Paranioar/Awesome_Image_Text_Retrieval_Benchmark
The Unified Code of Image-Text Retrieval for Further Exploration.
Paranioar/DBL
[TIP2024] The code of “Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching”
Rongtao-Xu/Awesome-LLM-EN
Lionelsy/Conference-Accepted-Paper-List
Some Conferences' accepted paper lists (including AI, ML, Robotic)
YicongHong/Thinking-VLN
Ideas and thoughts about the fascinating Vision-and-Language Navigation
Paranioar/Awesome_Matching_Pretraining_Transfering
The Paper List of Large Multi-Modality Model (Perception, Generation, Unification), Parameter-Efficient Finetuning, Vision-Language Pretraining, Conventional Image-Text Matching for Preliminary Insight.
JiazuoYu/MoE-Adapters4CL
Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024
pengsida/learning_research
本人的科研经验
cmhungsteve/Awesome-Transformer-Attention
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
micronDLA/MobileViTv3
22109095/SimOWT
This repository is an official implementation of the paper A Simple Baseline for Open-World Tracking via Self-training.
hkchengrex/Tracking-Anything-with-DEVA
[ICCV 2023] Tracking Anything with Decoupled Video Segmentation
statusrank/XCurve
XCurve is an end-to-end PyTorch library for X-Curve metrics optimizations in machine learning.
jianghaojun/Awesome-Parameter-Efficient-Transfer-Learning
A collection of parameter-efficient transfer learning papers focusing on computer vision and multimodal domains.
longzw1997/Open-GroundingDino
This is the third party implementation of the paper Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection.
jiawen-zhu/ViPT
[CVPR23] Visual Prompt Multi-Modal Tracking
MasterBin-IIAU/UNINEXT
[CVPR'23] Universal Instance Perception as Object Discovery and Retrieval
MasterBin-IIAU/Unicorn
[ECCV'22 Oral] Towards Grand Unification of Object Tracking
YangLiu14/detectron2-OWT
Detectron2 is FAIR's next-generation platform for object detection and segmentation.
YangLiu14/Open-World-Tracking
Official code for "Opening up Open World Tracking" (CVPR 2022)