43
5
v1v2v3 (latest)

MMAR: Towards Lossless Multi-Modal Auto-Regressive Probabilistic Modeling

Main:8 Pages
11 Figures
Bibliography:4 Pages
6 Tables
Appendix:6 Pages
Abstract

Recent advancements in multi-modal large language models have propelled the development of joint probabilistic models capable of both image understanding and generation. However, we have identified that recent methods suffer from loss of image information during understanding task, due to either image discretization or diffusion denoising steps. To address this issue, we propose a novel Multi-Modal Auto-Regressive (MMAR) probabilistic modeling framework. Unlike discretization line of method, MMAR takes in continuous-valued image tokens to avoid information loss in an efficient way. Differing from diffusion-based approaches, we disentangle the diffusion process from auto-regressive backbone model by employing a light-weight diffusion head on top each auto-regressed image patch embedding. In this way, when the model transits from image generation to understanding through text generation, the backbone model's hidden representation of the image is not limited to the last denoising step. To successfully train our method, we also propose a theoretically proven technique that addresses the numerical stability issue and a training strategy that balances the generation and understanding task goals. Extensive evaluations on 18 image understanding benchmarks show that MMAR significantly outperforms most of the existing joint multi-modal models, surpassing the method that employs pre-trained CLIP vision encoder. Meanwhile, MMAR is able to generate high quality images. We also show that our method is scalable with larger data and model size.

View on arXiv
@article{yang2025_2410.10798,
  title={ MMAR: Towards Lossless Multi-Modal Auto-Regressive Probabilistic Modeling },
  author={ Jian Yang and Dacheng Yin and Yizhou Zhou and Fengyun Rao and Wei Zhai and Yang Cao and Zheng-Jun Zha },
  journal={arXiv preprint arXiv:2410.10798},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.