TL;DR: We introduce LLaDA-V, a competitive diffusion-based vision-language model, outperforming other diffusion MLLMs.