Researcher, Google DeepMind
2 papers at NeurIPS 2025
We introduce TiledFlashLinearAttention a faster kernel algorithm for Linear RNNs and mLSTMs by improved Sequence Parallelism.
We provide a two-level factorization to learn structured latent states with interactions for object-centric RL.