I design diffusion language models, a new paradigm for parallel text generation.

Unlike traditional language models that generate one token at a time, diffusion models output tokens in parallel. My research explores novel diffusion language models and architectures to enhance their quality, generation speed, and training efficiency.

Selected Works

Arriola, M., Gokaslan, A., Chiu, J. T., Yang, Z., Qi, Z., Han, J., Sahoo, S. S., Kuleshov, V. Block Diffusion: Interpolating Between Autoregressive and Diffusion Language Models. ICLR 2025 (Oral, Top 1.77%). [Paper] [Blog] [Code]

Block Diffusion Paper Visualization

Arriola, M., Venkat, N., Granskog, J., Germanidis, A. Adapting Autoregressive Vision Language Models for Parallel Diffusion Decoding. Runway Research Blog. [Blog]

Arriola, M.*, Schiff, Y.*, Phung, H., Gokaslan, A., Kuleshov, V. Encoder-Decoder Block Diffusion Language Models for Efficient Training and Inference. NeurIPS 2025. [Paper: WIP] [Blog: WIP] [Code: WIP]

Sahoo, S., Arriola, M., Schiff, Y., Gokaslan, A., Marroquin, E., Chiu, J., Rush, A., Kuleshov, V. Simple and Effective Masked Diffusion Language Models. NeurIPS 2024. [Paper] [Blog] [Code]

Masked Diffusion Paper Visualization

News

  • Oct-25: Encoder-decoder diffusion LMs paper accepted to NeurIPS 2025. I was also recognized as a Top Reviewer!
  • Jun-25: Started a summer internship at Runway in NYC!
  • Apr‑25: Presenting Block Diffusion as an oral at ICLR 2025, main track.
  • Apr‑25: Invited talk at Amazon AGI on Block Diffusion.