HenryHZY
Interested in multimodal learning (vision-and-language) and parameter-efficient learning.
LaVi Lab led by Prof. Liwei Wang @ CSE, CUHKHong Kong
Pinned Repositories
annotated_deep_learning_paper_implementations
🧑🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
Ask-Anything
[CVPR2024][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
Awesome-Multimodal-LLM
Research Trends in LLM-guided Multimodal Learning.
cheatsheets-ai
Essential Cheat Sheets for deep learning and machine learning researchers https://medium.com/@kailashahirwar/essential-cheat-sheets-for-machine-learning-and-deep-learning-researchers-efb6a8ebd2e5
CLEVA
[EMNLP 2023 Demo] CLEVA: Chinese Language Models EVAluation Platform
Visual-Table
[EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"
VL-PET
[ICCV2023] Official code for "VL-PET: Vision-and-Language Parameter-Efficient Tuning via Granularity Control"
TG-Vid
[EMNLP 2024] Official code for "Enhancing Temporal Modeling of Video LLMs via Time Gating"
helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image models in HEIM (https://arxiv.org/abs/2311.04287) and vision-language models in VHELM (https://arxiv.org/abs/2410.07112).
HenryHZY's Repositories
HenryHZY/Awesome-Multimodal-LLM
Research Trends in LLM-guided Multimodal Learning.
HenryHZY/VL-PET
[ICCV2023] Official code for "VL-PET: Vision-and-Language Parameter-Efficient Tuning via Granularity Control"
HenryHZY/glados_auto_checkin
HenryHZY/GlaDOS-auto-checkin
GLADOS自动签到,多账号签到
HenryHZY/annotated_deep_learning_paper_implementations
🧑🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
HenryHZY/Ask-Anything
[CVPR2024][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
HenryHZY/cheatsheets-ai
Essential Cheat Sheets for deep learning and machine learning researchers https://medium.com/@kailashahirwar/essential-cheat-sheets-for-machine-learning-and-deep-learning-researchers-efb6a8ebd2e5
HenryHZY/CLEVA
[EMNLP 2023 Demo] CLEVA: Chinese Language Models EVAluation Platform
HenryHZY/Conference-Acceptance-Rate
Acceptance rates for the major AI conferences
HenryHZY/Visual-Table
[EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"
HenryHZY/latex_paper_writing_tips
Tips for Writing a Research Paper using LaTeX
HenryHZY/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning: LLaVA (Large Language-and-Vision Assistant) built towards GPT-4V level capabilities.
HenryHZY/MyArxiv
HenryHZY/ST-LLM
Official implementation of the paper "ST-LLM: Large Language Models Are Effective Temporal Learners"