Bruno Mlodozeniec
banner
brunokm.bsky.social
Bruno Mlodozeniec
@brunokm.bsky.social
PhD in Deep Learning at Cambridge. Previously Microsoft Research AI resident & researcher at Qualcomm. I want to find the key to generalisation.
If you want to learn more about how to apply influence functions to diffusion models, and the key take-aways for their use in this setting, check-out the paper! arxiv.org/abs/2410.13850
Influence Functions for Scalable Data Attribution in Diffusion Models
Diffusion models have led to significant advancements in generative modelling. Yet their widespread adoption poses challenges regarding data attribution and interpretability. In this paper, we aim to ...
arxiv.org
April 16, 2025 at 12:45 PM
For example: for even moderately sized datasets, the trained diffusion models' marginal probability distribution stays the same irrespective of which examples were removed from the training data, potentially making the influence functions task vacuous.
April 16, 2025 at 12:45 PM
We also point out several empirical challenges to the use of influence functions in diffusion models.
April 16, 2025 at 12:45 PM
In our paper, we empirically show that the choice of GGN and K-FAC approximation is crucial for the performance of influence functions, and that following our recommended design principles leads to the better performing approximations.
April 16, 2025 at 12:45 PM
Influence functions require the training loss Hessian matrix. Typically, a K-FAC approximation to a Generalised Gauss-Newton (GGN) matrix is used instead of the Hessian. However, it's not immediately obvious which GGN and K-FAC approximations to use in the diffusion
April 16, 2025 at 12:45 PM
Influence functions are already being used in deep learning, from classification and regression through to autoregressive LLMs. What's the challenge in adapting them to the diffusion setting?
April 16, 2025 at 12:45 PM
• Identifying and removing data responsible for undesirable behaviours (e.g. generating explicit content)
• Data valuation (how much did each training datapoint contribute towards generating the samples my users pay me for?)
April 16, 2025 at 12:45 PM
Answering how a model's behaviour changes upon removing training datapoints could help with:
• Quantifying impact of copyrighted data on a given sample (how much less likely is it that the model would generate this image if not for the works of a given artist?)
April 16, 2025 at 12:45 PM
Influence functions attempt to answer: how would the model's behaviour (e.g. probability of generating an image) change if the model was trained from scratch with some training datapoints removed.

They give an approximate answer, but without actually retraining the model.
April 16, 2025 at 12:45 PM
It’s an awesome piece of work, done on a surprisingly small budget compared to the performance
March 21, 2025 at 1:55 PM