Bayesian Influence Functions for Hessian-Free Data Attribution

Philipp Alexander Kreer, mentored by Jesse Hoogland, researched Bayesian generalization of influence functions.

Summary

Classical influence functions face significant challenges when applied to deep neural networks, primarily due to non-invertible Hessians and high-dimensional parameter spaces. We propose the local Bayesian influence function (BIF), an extension of classical influence functions that replaces Hessian inversion with loss landscape statistics that can be estimated via stochastic-gradient MCMC sampling. This Hessian-free approach captures higher-order interactions among parameters and scales efficiently to neural networks with billions of parameters. We demonstrate state-of-the-art results on predicting retraining experiments.

Previous
Previous

Activation Probes Are Reliable With a Handful of Positive Examples

Next
Next

Trivial Trojans: How Minimal MCP Servers Enable Cross-Tool Exfiltration of Sensitive Data