Pinned Repositories
Accepted-Paper-List
Update continuously the accepted papers from the top conferences in AI (2023)
alpaca-lora
Instruct-tune LLaMA on consumer hardware
awesome-egocentric-vision
A curated list of egocentric (first-person) vision and related area resources
BiHGH
ACMMM 2022 Oral
EgoVideoQA
[ICML 2024] Official Implementation for Multi-Factor Adaptive Vision Selection for Egocentric Video Question Answering
HCQA
[CVPR 2024 Workshop] The champion solution for Ego4D EgoSchema Challenge in CVPR 2024
Hyu-Zhang.github.io
ISR
multimodal_dialog_summary
Summary of papers and projects for visual dialog, video dialog, and multimodal dialog (Updating)
VQA
A Baseline for Toloka Visual Question Answering Challenge at WSDM Cup 2023
Hyu-Zhang's Repositories
Hyu-Zhang/BiHGH
ACMMM 2022 Oral
Hyu-Zhang/EgoVideoQA
[ICML 2024] Official Implementation for Multi-Factor Adaptive Vision Selection for Egocentric Video Question Answering
Hyu-Zhang/HCQA
[CVPR 2024 Workshop] The champion solution for Ego4D EgoSchema Challenge in CVPR 2024
Hyu-Zhang/ISR
Hyu-Zhang/VQA
A Baseline for Toloka Visual Question Answering Challenge at WSDM Cup 2023
Hyu-Zhang/Accepted-Paper-List
Update continuously the accepted papers from the top conferences in AI (2023)
Hyu-Zhang/Hyu-Zhang.github.io
Hyu-Zhang/multimodal_dialog_summary
Summary of papers and projects for visual dialog, video dialog, and multimodal dialog (Updating)
Hyu-Zhang/alpaca-lora
Instruct-tune LLaMA on consumer hardware
Hyu-Zhang/awesome-egocentric-vision
A curated list of egocentric (first-person) vision and related area resources
Hyu-Zhang/DeepLearning.github.io
Hyu-Zhang/AliceMind
ALIbaba's Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab
Hyu-Zhang/Awesome-Mamba-Papers
Awesome Papers related to Mamba.
Hyu-Zhang/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
Hyu-Zhang/awesome-open-gpt
Collection of Open Source Projects Related to GPT,GPT相关开源项目合集🚀、精选🔥🔥
Hyu-Zhang/Inpaint-Anything
Inpaint anything using Segment Anything and inpainting models.
Hyu-Zhang/LARCH
A novel contextuaL imAge seaRch sCHeme (LARCH)
Hyu-Zhang/Multi-Modal-Transformer
The repository collects many various multi-modal transformer architectures, including image transformer, video transformer, image-language transformer, video-language transformer and self-supervised learning models. Additionally, it also collects many useful tutorials and tools in these related domains.
Hyu-Zhang/OFA
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
Hyu-Zhang/papers
Summaries of machine learning papers
Hyu-Zhang/UMD
Official implementation of "User Attention-guided Multimodal Dialog Systems"
Hyu-Zhang/ViT-Adapter
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions