Auto-Encoding Morph-Tokens for Multimodal LLM

Kaihang Pan1, Siliang Tang1, Juncheng Li1,2†, Zhaoyu Fan1, Wei Chow1, Shuicheng Yan3, Tat-Seng Chua2, Yueting Zhuang1, Hanwang Zhang3,4

1Zhejiang University, 2National University of Singapore, 3Skywork AI, 4Nanyang Technological University

Corresponding Author

Overview

We introduce Morph-Tokens to resolve the conflicting objectives of visual comprehension and generation. The term ''morph'' implies a transformation where the pre-MLLM visual-tokens are not necessarily equal to the post-MLLM ones. Specifically, the pre-MLLM tokens are abstract semantics, serving as visual prompts for comprehension tasks. In contrast, the post-MLLM tokens are visually complete tokens for image generation, thanks to the powerful comprehension ability of MLLM that recovers the lost visual features due to abstraction. The framework of our morph-token-based MLLM is shown in the following figure:

On this basis, we propose a 3-stage training strategy as shown in the following figure. After training, it shows remarkable abilities, exceling at both multimodal comprehension and generation.

Acknowledgment

Thanks to the open source of the following projects:

  • LAVIS: A Library for Language-Vision Intelligence.
  • MiniGPT-4: Enhancing Vision-language Understanding with Advanced Large Language Models.
  • taming-transformers: Taming Transformers for High-Resolution Image Synthesis.
  • SEED: Making LLaMA SEE and Draw with SEED Tokenizer.

Citation

If you found this work useful, please consider citing our paper as follows:

@misc{pan2024autoencoding,
      title={Auto-Encoding Morph-Tokens for Multimodal LLM}, 
      author={Kaihang Pan and Siliang Tang and Juncheng Li and Zhaoyu Fan and Wei Chow and Shuicheng Yan and Tat-Seng Chua and Yueting Zhuang and Hanwang Zhang},
      year={2024},
      eprint={2405.01926},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}