/PosterLLaVA

Primary LanguagePythonApache License 2.0Apache-2.0

PosterLLaVa: Constructing a Unified Multi-modal Layout Generator with LLM

This repository is for the paper "PosterLLaVa: Constructing a Unified Multi-modal Layout Generator with LLM" (under review).

🎊 News: Automatic Text-to-poster pipeline PosterGen

Now we are heading from layout toward the real sense of graphic poster design. A text-to-poster pipeline, PosterGen, will soon be available online to support the real-world application of PosterLLaVA. There are some previewing examples:

🀝 Usage Acknowledgement

Notice we only authorize using the proposed dataset for scientific research. One should NOT use it for commercial purposes without our authorization.

πŸ—“οΈ Schedule

[2024.03.26] Release online demo and pre-trained model on hugging faceπŸ€—.

[2024.06.05] Release arXiv paperπŸ“.

[2024.07.04] Release QB-Poster datasetπŸ“Š. (raw files contain original poster images and JSON annotations, inpainting and saliency detection techniques are needed for obtaining background images and saliency maps. Our paper used lama for inpainting and basenet for saliency detection.)

[2024.07.04] Release User-Constrained datasetπŸ“Š. (only include user-constraint annotation files. please refer to the CGL-dataset and PosterLayout dataset to get the poster images and bounding box annotations.)

[2024.07.04] Release data pre-processing, training, and inferencing code.

[2024.08.29] An automatic text-to-poster system PosterGen (with PosterLLaVA as the backbone) will soon be open-sourced to supplement this work.

[Coming Soon] Release evaluation code.

Environment

Run the following code to build the environment.

pip install --upgrade pip  # enable PEP 660 support
pip install -e .
pip install -e ".[train]"
pip install flash-attn --no-build-isolation

Data Processing

Download the dataset files and arrange them as follows (QB-Poster as an example). Run the saliency detection method to get 'saliency_map' and the inpainting method to get 'inpainted_1x' and 'inpainted_1d5x' (used for inference and training respectively; notice we randomly inpainted 0.5x more regions besides the ground-truth bounding box area to avoid overfitting.)

β”œβ”€β”€ data
β”‚  β”œβ”€β”€ prompt_template.txt
β”‚  └── qbposter <--
β”‚      β”œβ”€β”€ get_prompt.py
|      └── raw
β”‚          β”œβ”€β”€ original_poster
β”‚          β”œβ”€β”€ saliency_map
β”‚          β”œβ”€β”€ inpainted_1x
β”‚          β”œβ”€β”€ inpainted_1d5x
β”‚          └── annotation.json
...
└── README.md

Run the data preprocessing script.

python data/qbposter/get_prompt.py

Ultimately you will get two processed JSON files (each containing instruction-answer pairs) like this.

β”œβ”€β”€ data
β”‚  β”œβ”€β”€ prompt_template.txt
β”‚  └── qbposter
β”‚        β”œβ”€β”€ get_prompt.py
β”‚        β”œβ”€β”€ qbposter_train_instruct.json <--
β”‚        └── qbposter_val_instruct.json   <--
...
└── README.md

Training

Please download LLaVa-v1.5 pre-trained checkpoint and CLIP vision encoder first and put it in the 'huggingface' subfolder.

β”œβ”€β”€ data
β”œβ”€β”€ huggingface <--
|      β”œβ”€β”€ llava-v1.5-7b
|      └── clip-vit-large-patch14-336
β”œβ”€β”€ scripts
|      └── qbposter
|            β”œβ”€β”€ finetune.sh <--
|            └── inference.sh
...
└── README.md

Then run the following script.

qbposter/finetune.sh

Inference

Please download the pre-trained PosterLLaVa_v0 checkpoint, which is initialized with LLaVa-v1.5 checkpoint and fine-tuned on the following combined datasets.

  • 7k banner layouts from Ad Banner dataset.
  • 60k commercial poster layouts from CGL-dataset and PosterLayout with text constraints.
  • 4k social media poster layouts from QB-Poster dataset.

Put it in the 'pretrained_model' subfolder.

β”œβ”€β”€ data
β”œβ”€β”€ huggingface
β”œβ”€β”€ pretrained_model <--
|      └── posterllava_v0
β”œβ”€β”€ scripts
|      └── qbposter
|            β”œβ”€β”€ finetune.sh
|            └── inference.sh <--
...
└── README.md

Then run the following script to generate JSON format layout.

qbposter/inference.sh

Evaluation

Coming Soon...

Citation

If you find this project/paper useful, please give us a star/citation.

@misc{yang2024posterllava,
      title={PosterLLaVa: Constructing a Unified Multi-modal Layout Generator with LLM}, 
      author={Tao Yang and Yingmin Luo and Zhongang Qi and Yang Wu and Ying Shan and Chang Wen Chen},
      year={2024},
      eprint={2406.02884},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2406.02884}, 
}