THUDM
ChatGLM, CogVLM, CodeGeeX, WebGLM, GLM-130B, CogView, CogVideo | CogDL, GNNs, AMiner | Knowledge Engineering Group (KEG) & Data Mining at Tsinghua University
FIT Building, Tsinghua University
Pinned Repositories
AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
AutoWebGLM
ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
ChatGLM3
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
CodeGeeX
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
CodeGeeX2
CodeGeeX2: A More Powerful Multilingual Code Generation Model
CogVLM
a state-of-the-art-level open visual language model | 多模态预训练模型
CogVLM2
GPT4V-level open-source multi-modal model based on Llama3-8B
GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
THUDM's Repositories
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
THUDM/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
THUDM/ChatGLM3
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
THUDM/CodeGeeX
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
THUDM/CodeGeeX2
CodeGeeX2: A More Powerful Multilingual Code Generation Model
THUDM/CogVLM
a state-of-the-art-level open visual language model | 多模态预训练模型
THUDM/AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
THUDM/CogDL
CogDL: A Comprehensive Library for Graph Deep Learning (WWW 2023)
THUDM/CogVLM2
GPT4V-level open-source multi-modal model based on Llama3-8B
THUDM/SwissArmyTransformer
SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.
THUDM/LongBench
LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding
THUDM/AutoWebGLM
THUDM/RelayDiffusion
The official implementation of "Relay Diffusion: Unifying diffusion process across resolutions for image synthesis" [ICLR 2024 Spotlight]
THUDM/Inf-DiT
Official implementation of Inf-DiT: Upsampling Any-Resolution Image with Memory-Efficient Diffusion Transformer
THUDM/CogCoM
THUDM/LongAlign
LongAlign: A Recipe for Long Context Alignment Encompassing Data, Training, and Evaluation
THUDM/kgTransformer
kgTransformer: pre-training for reasoning over complex KG queries (KDD 22)
THUDM/ScenarioMeta
Source code and dataset for KDD 2019 paper "Sequential Scenario-Specific Meta Learner for Online Recommendation"
THUDM/ChatGLM-Math
THUDM/SciGLM
SciGLM: Training Scientific Language Models with Self-Reflective Instruction Annotation and Tuning
THUDM/WhoIsWho
KDD'23 Web-Scale Academic Name Disambiguation: the WhoIsWho Benchmark, Leaderboard, and Toolkit
THUDM/NaturalCodeBench
THUDM/paper-source-trace
THUDM/Self-Contrast
Extensive Self-Contrast Enables Feedback-Free Language Model Alignment
THUDM/whoiswho-top-solutions
THUDM/RecDCL
RecDCL: Dual Contrastive Learning for Recommendation (WWW'24, Oral)
THUDM/OAG-AQA
THUDM/Reviewer-Rec
THUDM/tot-prediction
THUDM/Megatron-LM
Ongoing research training transformer models at scale