Bayesian Influence Functions for Hessian-Free Data Attribution

Philipp Alexander Kreer, mentored by Jesse Hoogland, researched Bayesian generalization of influence functions.

Summary

Classical influence functions face significant challenges when applied to deep neural networks, primarily due to non-invertible Hessians and high-dimensional parameter spaces. We propose the local Bayesian influence function (BIF), an extension of classical influence functions that replaces Hessian inversion with loss landscape statistics that can be estimated via stochastic-gradient MCMC sampling. This Hessian-free approach captures higher-order interactions among parameters and scales efficiently to neural networks with billions of parameters. We demonstrate state-of-the-art results on predicting retraining experiments.

Full paper
Previous
Previous

Influence Dynamics and Stagewise Data Attribution

Next
Next

Trivial Trojans: How Minimal MCP Servers Enable Cross-Tool Exfiltration of Sensitive Data