Poster Session 4 · Thursday, December 4, 2025 4:30 PM → 7:30 PM
#5204
Scalable Best-of-N Selection for Large Language Models via Self-Certainty
Abstract
Best-of-N selection is a key technique for improving the reasoning performance of Large Language Models (LLMs) through increased test-time computation. Current state-of-the-art methods often employ computationally intensive reward models for response evaluation and selection. Reward-free alternatives, like self-consistency and universal self-consistency, are limited in their ability to handle open-ended generation tasks or scale effectively.
To address these limitations, we propose self-certainty, a novel and efficient metric that leverages the inherent probability distribution of LLM outputs to estimate response quality without requiring external reward models. We hypothesize that higher distributional self-certainty, aggregated across multiple samples, correlates with improved response accuracy, as it reflects greater confidence in the generated output.
Through extensive experiments on various reasoning tasks, we demonstrate that self-certainty:
- scales effectively with increasing sample size , akin to reward models but without the computational overhead;
- complements chain-of-thought, improving reasoning performance beyond greedy decoding; and
- generalizes to open-ended tasks where traditional self-consistency methods fall short.
Our findings establish self-certainty as a practical and efficient way for improving LLM reasoning capabilities. The code is available at
https://github.com/backprop07/Self-Certainty