jainie-max's Stars
lucidrains/mixture-of-experts
A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models
TUDB-Labs/MoE-PEFT
An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT
THUDM/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
VITA-Group/Diffusion4D
"Diffusion4D: Fast Spatial-temporal Consistent 4D Generation via Video Diffusion Models", Hanwen Liang*, Yuyang Yin*, Dejia Xu, Hanxue Liang, Zhangyang Wang, Konstantinos N. Plataniotis, Yao Zhao, Yunchao Wei
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
lloongx/DIKI
[ECCV 2024] Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models
google-research/l2p
Learning to Prompt (L2P) for Continual Learning @ CVPR22 and DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning @ ECCV22
JingyangQiao/prompt-gradient-projection
liangyanshuo/InfLoRA
The official implementation of the CVPR'2024 work Interference-Free Low-Rank Adaptation for Continual Learning
QwenLM/Qwen2-VL
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
LLaVA-VL/LLaVA-NeXT
e2b-dev/awesome-ai-agents
A list of AI autonomous agents
jun0wanan/awesome-large-multimodal-agents
WooooDyy/LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
langgenius/dify
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
donydchen/mvsplat
🌊 [ECCV'24 Oral] MVSplat: Efficient 3D Gaussian Splatting from Sparse Multi-View Images
Ekko-zn/AIGCDetectBenchmark
Daisy-Zhang/Awesome-AIGC-Detection
A collection list of AIGC detection related papers.
NeeluMadan/ViFM_Survey
Foundation Models for Video Understanding: A Survey
yunlong10/Awesome-LLMs-for-Video-Understanding
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
THUDM/GLM-4
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
ShuvenduRoy/CoPrompt
[ICLR'24] Consistency-guided Prompt Learning for Vision-Language Models
SCLBD/DeepfakeBench
A comprehensive benchmark of deepfake detection
flyingby/Awesome-Deepfake-Generation-and-Detection
A Survey on Deepfake Generation and Detection
Daisy-Zhang/Awesome-Deepfakes-Detection
A list of tools, papers and code related to Deepfake Detection.
DepthAnything/Depth-Anything-V2
Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
VITA-Group/4DGen
"4DGen: Grounded 4D Content Generation with Spatial-temporal Consistency", Yuyang Yin*, Dejia Xu*, Zhangyang Wang, Yao Zhao, Yunchao Wei
ZjjConan/Multi-Modal-Adapter
The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".
miccunifi/KDPL
[ECCV 2024] - Improving Zero-shot Generalization of Learned Prompts via Unsupervised Knowledge Distillation
black-forest-labs/flux
Official inference repo for FLUX.1 models