mashijie1028
Ph.D. student @ Institute of Automation, Chinese Academy of Sciences (CASIA). Previously B.E. @ Tsinghua University.
CASIABeijing
mashijie1028's Stars
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
coder/code-server
VS Code in the browser
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
gradio-app/gradio
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
Delgan/loguru
Python logging made (stupidly) simple
black-forest-labs/flux
Official inference repo for FLUX.1 models
mlfoundations/open_clip
An open source implementation of CLIP.
kohya-ss/sd-scripts
ostris/ai-toolkit
Various AI scripts. Mostly Stable Diffusion stuff.
jingyi0000/VLM_survey
Collection of AWESOME vision-language models for vision tasks
baaivision/Emu3
Next-Token Prediction is All You Need
XLabs-AI/x-flux
Stability-AI/sd3.5
TencentARC/Open-MAGVIT2
Open-MAGVIT2: Democratizing Autoregressive Visual Generation
LAION-AI/CLIP_benchmark
CLIP-like model evaluation
TencentARC/LLaMA-Pro
[ACL 2024] Progressive LLaMA with Block Expansion.
AILab-CVC/SEED-X
Multimodal Models in Real World
AILab-CVC/SEED-Bench
(CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.
Atomic-man007/Awesome_Multimodel_LLM
Awesome_Multimodel is a curated GitHub repository that provides a comprehensive collection of resources for Multimodal Large Language Models (MLLM). It covers datasets, tuning techniques, in-context learning, visual reasoning, foundational models, and more. Stay updated with the latest advancement.
baaivision/DIVA
Diffusion Feedback Helps CLIP See Better
SHI-Labs/CuMo
CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts
Hsu1023/DuQuant
[NeurIPS 2024 Oral🔥] DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs.
Understanding-Visual-Datasets/VisDiff
Official implementation of "Describing Differences in Image Sets with Natural Language" (CVPR 2024 Oral)
TencentARC/mllm-npu
mllm-npu: training multimodal large language models on Ascend NPUs
dongzhuoyao/Diffusion-Representation-Learning-Survey-Taxonomy
Baijiong-Lin/LoRA-Torch
PyTorch Reimplementation of LoRA (featuring with supporting nn.MultiheadAttention)
mashijie1028/Happy-CGCD
Official code for NeurIPS 2024 paper "Happy: A Debiased Learning Framework for Continual Generalized Category Discovery"
KyanChen/MakeMultiHeadNaive
Use naive MultiheadAttention implement to replace nn.MultiheadAttention in pytorch
mashijie1028/TrustDD
Code for our paper "Towards Trustworthy Dataset Distillation" (Pattern Recognition 2025)
mashijie1028/UCAS-CASIA-Beamer-Theme
Beamer template for UCAS and CASIA.