DreamLLM: Synergistic Multimodal Comprehension and Creation
Abstract
This paper presents DreamLLM, a learning framework that first achieves versatile Multimodal Large Language Models (MLLMs) empowered with frequently overlooked synergy between multimodal comprehension and creation. DreamLLM operates on two fundamental principles. The first focuses on the generative modeling of both language and image posteriors by direct sampling in the raw multimodal space. This approach circumvents the limitations and information loss inherent to external feature extractors like CLIP, and a more thorough multimodal understanding is obtained. Second, DreamLLM fosters the generation of raw, interleaved documents, modeling both text and image contents, along with unstructured layouts. This allows DreamLLM to learn all conditional, marginal, and joint multimodal distributions effectively. As a result, DreamLLM is the first MLLM capable of generating free-form interleaved content. Comprehensive experiments highlight DreamLLM's superior performance as a zero-shot multimodal generalist, reaping from the enhanced learning synergy. Project page: https://dreamllm.github.io.
Cite
Text
Dong et al. "DreamLLM: Synergistic Multimodal Comprehension and Creation." International Conference on Learning Representations, 2024.Markdown
[Dong et al. "DreamLLM: Synergistic Multimodal Comprehension and Creation." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/dong2024iclr-dreamllm/)BibTeX
@inproceedings{dong2024iclr-dreamllm,
title = {{DreamLLM: Synergistic Multimodal Comprehension and Creation}},
author = {Dong, Runpei and Han, Chunrui and Peng, Yuang and Qi, Zekun and Ge, Zheng and Yang, Jinrong and Zhao, Liang and Sun, Jianjian and Zhou, Hongyu and Wei, Haoran and Kong, Xiangwen and Zhang, Xiangyu and Ma, Kaisheng and Yi, Li},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/dong2024iclr-dreamllm/}
}