jingchenchen's Stars
aim-uofa/Matcher
[ICLR'24] Matcher: Segment Anything with One Shot Using All-Purpose Feature Matching
mlfoundations/open_flamingo
An open-source framework for training large multimodal models.
haoosz/ade-czsl
[CVPR 2023] Learning Attention as Disentangler for Compositional Zero-shot Learning
BatsResearch/csp
Learning to compose soft prompts for compositional zero-shot learning.
junkunyuan/Awesome-Domain-Generalization
Awesome things about domain generalization, including papers, code, etc.
PengtaoJiang/Segment-Anything-CLIP
Connecting segment-anything's output masks with the CLIP model; Awesome-Segment-Anything-Works
OptimalScale/LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Wangt-CN/EqBen
[ICCV'23 Oral] The introduction and toolkit for EqBen Benchmark
RenzeLou/awesome-instruction-learning
Papers and Datasets on Instruction Tuning and Following. ✨✨✨
aioz-ai/CFR_VQA
Coarse-to-Fine Reasoning for Visual Question Answering (CVPRW'22)
cvlab-columbia/viper
Code for the paper "ViperGPT: Visual Inference via Python Execution for Reasoning"
wqshmzh/CANet-CZSL
Official pytorch implementation of CVPR2023 paper "Learning Conditional Attributes for Compositional Zero-Shot Learning"
Jingkang50/OpenOOD
Benchmarking Generalized Out-of-Distribution Detection
zhoudw-zdw/Awesome-Few-Shot-Class-Incremental-Learning
Awesome Few-Shot Class-Incremental Learning
xialeiliu/Awesome-Incremental-Learning
Awesome Incremental Learning
XPixelGroup/HAT
CVPR2023 - Activating More Pixels in Image Super-Resolution Transformer Arxiv - HAT: Hybrid Attention Transformer for Image Restoration
muliyangm/DeCa
Code for DeCa (IEEE Trans. Multimedia 2022)
PeixianChen/MEDet
jingchenchen/ReasoningConsistency-VQA
virajprabhu/premise-emnlp17
Code for The Promise of Premise: Harnessing Question Premises in Visual Question Answering (EMNLP 2017)
arijitray1993/VQARelevance
Models and Codes for the paper Question Relevance in VQA: Identifying Non-Visual And False-Premise Questions
iQua/M-DGT
The source code of the CVPR22 paper titled "Multi-Modal Dynamic Graph Transformer for Visual Grounding".
WANGXinyiLinda/CMLE
This is the repository for the paper: Counterfactual Maximum Likelihood Estimation for Training Deep Networks.
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
bharathgs/Awesome-pytorch-list
A comprehensive list of pytorch related content on github,such as different models,implementations,helper libraries,tutorials etc.
jokieleung/awesome-visual-question-answering
A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.
zhoubolei/introRL
Intro to Reinforcement Learning (强化学习纲要)
aamini/evidential-deep-learning
Learn fast, scalable, and calibrated measures of uncertainty using neural networks!
eric-ai-lab/awesome-vision-language-navigation
A curated list for vision-and-language navigation. ACL 2022 paper "Vision-and-Language Navigation: A Survey of Tasks, Methods, and Future Directions"
JakobCode/UncertaintyInNeuralNetworks_Resources