Researcher, Google
2 papers at NeurIPS 2025
We introduce contextualized n-gram embeddings to extend input embedding layers, improving performance while maintaining fixed accelerator usage during inference.