Poster Session 1 · Wednesday, December 3, 2025 11:00 AM → 2:00 PM
#3002 Spotlight
On Traceability in Stochastic Convex Optimization
Abstract
In this paper, we investigate the necessity of traceability for accurate learning in stochastic convex optimization (SCO) under geometries. Informally, we say a learning algorithm is -traceable if, by analyzing its output, it is possible to identify at least of its training samples.
Our main results uncover a fundamental tradeoff between traceability and excess risk in SCO. For every , we establish the existence of an excess risk threshold below which every sample-efficient learner is traceable with the number of samples which is a constant fraction of its training sample. For , this threshold coincides with the best excess risk of differentially private (DP) algorithms, i.e., above this threshold, there exist algorithms that are not traceable, which corresponds to a sharp phase transition. For , this threshold instead gives novel lower bounds for DP learning, partially closing an open problem in this setup.
En route to establishing these results, we prove a sparse variant of the fingerprinting lemma, which is of independent interest to the community.