Associate Professor, Department of Computer Science, University of Toronto
4 papers at NeurIPS 2025
We scale the influence-function-based data valuation method to recent LLMs and their massive training datasets.
This paper introduces distributional training data attribution, a data attribution framework that accounts for stochasticity in deep learning training, enabling a mathematical justification for why influence functions work in this setting.
We apply the EKFAC-preconditioner on Neumann series iterations to arrive at an unbiased iHVP approximation for TDA that improves influence function and unrolled differentiation performance.