Poster Session 6 · Friday, December 5, 2025 4:30 PM → 7:30 PM
#3500
Encoder-Decoder Diffusion Language Models for Efficient Training and Inference
Abstract
Discrete diffusion models enable parallel token sampling for faster inference than autoregressive approaches. However, prior diffusion models use a decoder-only architecture, which requires sampling algorithms that invoke the full network at every denoising step and incur high computational cost.
Our key insight is that discrete diffusion models perform two types of computation:
- representing clean tokens
- denoising corrupted tokens, which enables us to use separate modules for each task.
We introduce a framework for Efficient Encoder-Decoder Diffusion (E2D2), consisting of an architecture with specialized training and sampling algorithms, and we show that E2D2 achieves superior trade-offs between generation quality and inference throughput on summarization, translation, and mathematical reasoning tasks.