Postdoc, Department of Computer Science, ETHZ - ETH Zurich
2 papers at NeurIPS 2025
By carefully coordinating off-the-shelf models with inference only, we show the DSP framework can achieve surprisingly good results in theorem proving, comparable to the frontier models with RL-based large-scale training.
We introduce the first theoretical framework for analyzing LLM reasoning errors, and bridge two typical sampling-based test-time scaling methods to achieve both low error and fast convergence.