logo
today local_bar
Poster Session 4 · Thursday, December 4, 2025 4:30 PM → 7:30 PM
#3109 Spotlight

Is the acquisition worth the cost? Surrogate losses for Consistent Two-stage Classifiers

NeurIPS OpenReview

Abstract

Recent years have witnessed the emergence of a spectrum of foundation models, covering a broad range of capabilities and costs. Often, we effectively use foundation models as feature generators and train classifiers that use the outputs of these models to make decisions.
In this paper, we consider an increasingly relevant setting where we have two classifier stages. The first stage has access to features and has the option to make a classification decision or defer, while incurring a cost, to a second classifier that has access to features and . This is similar to the "learning to defer" setting, with the important difference that we train both classifiers jointly, and the second classifier has access to more information.
The natural loss for this setting is an loss, where a penalty is paidfor incorrect classification, as in , but an additional penalty is paidfor consulting the second classifier. The loss is unwieldyfor training. Our primary contribution in this paper is the derivationof a hinge-based surrogate loss that is much moreamenable to training but also satisfies the property that-consistency implies -consistency.