/MuLan

MuLan: Adapting Multilingual Diffusion Models for 110+ Languages (无需额外训练为任意扩散模型支持多语言能力)

Primary LanguagePython

🌻 MuLan

PyPI - Downloads

# pip install mulankit
from diffusers import StableDiffusionPipeline
+ import mulankit

pipe = StableDiffusionPipeline.from_pretrained('Lykon/dreamshaper-8')
+ pipe = mulankit.transform(pipe, 'mulanai/mulan-lang-adapter::sd15_aesthetic.pth')
image = pipe('一只蓝色的🐶 in the 바다').images[0]
一只蓝色的 🐶 in the 바다 (Dreamshaper-8) レゴシュワルツェネッガー (SDXL-lightning) 一只可爱的猫头鹰 (MVDream) 海浪风景 (AnimateDiff)
dreamshaper8 一只戴着帽子的 rabbit レゴアーノルド・シュワルツェネッガー 海浪

What is it ?

We present MuLan, a versatile framework to equip any diffusion model with multilingual generation abilities natively by up to 110+ languages around the world. With properly trained text encoder from noisy data, we demonstrate that MuLan could be trained on English only data and support other languages zero-shot. Additionally, we introduce Language Adapter. A language adapter with less than 20M parameters, trained against a frozen denoiser and a text encoder, can be readily combined with any homologous community models/tools, such as LoRA, LCM, ControlNet, and IP-Adapter, without any finetuning.

无需额外训练,MuLan(木兰)可以为任何扩散模型提供原生的多语言能力。MuLan可以仅在英语数据上进行训练,即可泛化到其他多达110多种语言上。通过引入了语言适配器,我们可以将 MuLan 的多语言能力无缝地插入到任何同类社区模型/工具(如LoRA、LCM、ControlNet和IP-Adapter)中并且无需任何微调。

demo.mp4

News

  • optimize memory usage.
  • release technical report.
  • 2024-5-14: release code and models.

How to use

We have hosted a gradio demo here.

MuLan supports

  • Base models: Stable Diffusion 1.5, 2.1, XL, Pixart-Alpha/Sigma.
  • Downstream models: ControlNet, LCM, LoRA, finetuned models and etc.
  • Video models: AnimateDiff.
  • 3D models: MVDream.

Please refer to the USAGE.md and examples for more details.

Model Release

Model Description Link
MuLan-Language-Adapter Adapters for SDXL, SD1.5/2.1, Pixart hf-model
MuLan-Pixart Full finetuned model hf-model

See more at our Huggingface 🌻 Homepage.

Citation

If you find this repo helpful, please considering citing us.

@article{lai2024mulan,
  title={MuLan: Adapting Multilingual Diffusion Models for 110 + Languages},
  year={2024}
}

Visitors

Acknowledgement

Our work is made possible by the open-source of these great works.

Stable Diffusion · Pixart-Alpha · InternVL

If you want to join our WeChat group, please scan the following QR Code to add our assistant as a Wechat friend:

image