CompVis - Computer Vision and Learning LMU Munich
Computer Vision and Learning research group at Ludwig Maximilian University of Munich (formerly Computer Vision Group at Heidelberg University)
Germany
Pinned Repositories
adaptive-style-transfer
source code for the ECCV18 paper A Style-Aware Content Loss for Real-time HD Style Transfer
depth-fm
[AAAI 2025, Oral] DepthFM: Fast Monocular Depth Estimation with Flow Matching
EDGS
A PyTorch implementation of the paper "EDGS: Eliminating Densification for Efficient Convergence of 3DGS"
geometry-free-view-synthesis
Is a geometric model required to synthesize novel views from a single image?
latent-diffusion
High-Resolution Image Synthesis with Latent Diffusion Models
metric-learning-divide-and-conquer
Source code for the paper "Divide and Conquer the Embedding Space for Metric Learning", CVPR 2019
stable-diffusion
A latent text-to-image diffusion model
taming-transformers
Taming Transformers for High-Resolution Image Synthesis
vunet
A generative model conditioned on shape and appearance.
zigma
A PyTorch implementation of the paper "ZigMa: A DiT-Style Mamba-based Diffusion Model" (ECCV 2024)
CompVis - Computer Vision and Learning LMU Munich's Repositories
CompVis/stable-diffusion
A latent text-to-image diffusion model
CompVis/latent-diffusion
High-Resolution Image Synthesis with Latent Diffusion Models
CompVis/taming-transformers
Taming Transformers for High-Resolution Image Synthesis
CompVis/depth-fm
[AAAI 2025, Oral] DepthFM: Fast Monocular Depth Estimation with Flow Matching
CompVis/EDGS
A PyTorch implementation of the paper "EDGS: Eliminating Densification for Efficient Convergence of 3DGS"
CompVis/geometry-free-view-synthesis
Is a geometric model required to synthesize novel views from a single image?
CompVis/zigma
A PyTorch implementation of the paper "ZigMa: A DiT-Style Mamba-based Diffusion Model" (ECCV 2024)
CompVis/fm-boosting
[ECCV 2024, Oral] FMBoost: Boosting Latent Diffusion with Flow Matching
CompVis/net2net
Network-to-Network Translation with Conditional Invertible Neural Networks
CompVis/tread
CompVis/retrieval-augmented-diffusion-models
Official codebase for the Paper “Retrieval-Augmented Diffusion Models”
CompVis/attribute-control
Fine-Grained Subject-Specific Attribute Expression Control in T2I Models
CompVis/discrete-interpolants
The official implementation of "[MASK] is All You Need"
CompVis/LoRAdapter
CompVis/cleandift
CompVis/diff2flow
[CVPR 2025] Diff2Flow: Training Flow Matching Models via Diffusion Model Alignment
CompVis/ipoke
iPOKE: Poking a Still Image for Controlled Stochastic Video Synthesis
CompVis/SCFlow
[ICCV 2025] SCFlow: Implicitly Learning Style and Content Disentanglement with Flow Models
CompVis/distilldift
[WACV 2025] DistillDIFT: Distillation of Diffusion Features for Semantic Correspondence
CompVis/instant-lora-composition
CompVis/maskflow
MaskFlow: Discrete Flows For Flexible and Efficient Long Video Generation
CompVis/DisCLIP
[AAAI 2025] Does VLM Classification Benefit from LLM Description Semantics?
CompVis/RepTok
Adapting Self-Supervised Representations as a Latent Space for Efficient Generation
CompVis/flow-poke-transformer
CompVis/cuneiform-sign-detection-dataset
Dataset provided with the article "Deep learning for cuneiform sign detection with weak supervision using transliteration alignment". It comprises image references, transliterations and sign annotations of clay tablets from the Neo-Assyrian epoch.
CompVis/cuneiform-sign-detection-code
Code for the article "Deep learning of cuneiform sign detection with weak supervision using transliteration alignment"
CompVis/Art-fm
[ICCV 2025] Stochastic Interpolants for Revealing Stylistic Flows across the History of Art
CompVis/wast3d
Official project page for the paper "WaSt-3D: Wasserstein-2 Distance for Scene-to-Scene Stylization on 3D Gaussians"
CompVis/zigma2
CompVis/mask
The official implementation of "[MASK] is All You Need"