PhD student, Beijing Institute of Technology
3 papers at NeurIPS 2025
We formalise quoting in conversation, release a training set and benchmark, and introduce a tiny adapter that lets LLMs exploit quoted spans with zero prompt overhead.
We propose DORA, a provably optimal resource allocation strategy for efficient test-time scaling.