xuejunzhang2002's Stars
lucidrains/vit-pytorch
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
kyegomez/Vit-RGTS
Open source implementation of "Vision Transformers Need Registers"
OpenBMB/MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
OpenGVLab/Multi-Modality-Arena
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP-2, and many more!
zjunlp/KnowledgeCircuits
[NeurIPS 2024] Knowledge Circuits in Pretrained Transformers
lancopku/label-words-are-anchors
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
nickjiang2378/vl-interp
Official Pytorch implementation of "Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations"
kmeng01/rome
Locating and editing factual associations in GPT (NeurIPS 2022)
shuyhere/Awesome-Sparse-Autoencoder
Collection of Reverse Engineering in Large Model
UVA-Computer-Vision-Lab/rivanna_resource
zjunlp/Kformer
[NLPCC 2022] Kformer: Knowledge Injection in Transformer Feed-Forward Layers
zepingyu0512/neuron-attribution
code for EMNLP 2024 paper: Neuron-Level Knowledge Attribution in Large Language Models
google-deepmind/svo_probes
The SVO-Probes Dataset for Verb Understanding
om-ai-lab/VL-CheckList
Evaluating Vision & Language Pretraining Models with Objects, Attributes and Relations. [EMNLP 2022]
jedyang97/awesome-cs-phd-application-advice
A curated list of awesome advice for computer science Ph.D. applicants.
sled-group/moh
Official Repository of Multi-Object Hallucination in Vision-Language Models (NeurIPS 2024)
ZHZisZZ/modpo
[ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization
codefuse-ai/MFTCoder
High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.
SihanXU/TransferWiki
A Wiki for students who want to transfer to another university
JacobPfau/fillerTokens
allenai/OLMoE
OLMoE: Open Mixture-of-Experts Language Models
bbycroft/llm-viz
3D Visualization of an GPT-style LLM
NielsRogge/Transformers-Tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
Yangyi-Chen/SOLO
[TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"
epfLLM/Megatron-LLM
distributed trainer for LLMs
hkproj/pytorch-paligemma
Coding a Multimodal (Vision) Language Model from scratch in PyTorch with full explanation: https://www.youtube.com/watch?v=vAmKB7iPkWw
LucasAlegre/morl-baselines
Multi-Objective Reinforcement Learning algorithms implementations.
sustcsonglin/flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
All-Hands-AI/OpenHands
🙌 OpenHands: Code Less, Make More