Full Professor, University of California, Los Angeles
3 papers at NeurIPS 2025
Substantially faster diffusion LLMs using a small auxiliary autoregressive model
We reduce training variance in equivariant generative models using a low-variance gradient estimator, improving stability and performance across molecular, crystal, and protein generation tasks.