/sketch2img

Sketch-to-Image Generation without retraining diffusion models

Primary LanguagePython

sketch2img

Sketch-to-Image Generation without retraining diffusion models. (WIP)

Currently supported method:

  • Sketch-Guided Text-to-Image Diffusion (google)
  • inject with additional pretrained self-attn layer or cross-attn layer (gligen?)

Note: Paint-With-Words already moved to https://github.com/Mikubill/sd-paint-with-words

Sketch-Guided Text-to-Image Diffusion

Paper | Demo

intro

Sketch-Guided Text-to-Image is a method proposed by researchers in Google Research to guide the inference process of a pretrained text-to-image diffusion model with an edge predictor that operates on the internal activations of the core network of the diffusion model, encouraging the edge of the synthesized image to follow a reference sketch.

Pretrained LGP Weights: https://huggingface.co/nyanko7/sketch2img-edge-predictor-train/blob/main/edge_predictor.pt