Sam Power
@spmontecarlo.bsky.social
2.3K followers 1.9K following 1.1K posts
Lecturer in Maths & Stats at Bristol. Interested in probabilistic + numerical computation, statistical modelling + inference. (he / him). Homepage: https://sites.google.com/view/sp-monte-carlo Seminar: https://sites.google.com/view/monte-carlo-semina
Posts Media Videos Starter Packs
spmontecarlo.bsky.social
... since there is plenty of value in identifying subsets of the problem space for which better solutions are available, rather than complaining too much about the fact that these solutions aren't universal (which was anyways a big ask!).
spmontecarlo.bsky.social
I used to pull away from this a bit, on the basis that some of the historical motivation for ABC and co. involved models which are far worse-behaved than this, and so these new proposals wouldn't help much with these original problems. With the benefit of hindsight, this was a bit misguided ...
spmontecarlo.bsky.social
Hopefully it's clear that "culturally wrong" is not really a criticism here. Anyways, I've had similar reactions in the context of simulation-based inference (still ABC to me, but what can one do?), where you often see new methodology which assumes that the forward model is differentiable.
spmontecarlo.bsky.social
... and they are often quite loud about how they don't want to have to deal with derivatives through their simulation pipeline, and so on. So in this regard, the AD-compatibility of the EnKF is clearly true in an objective sense, but almost feels "culturally wrong" if this is your main context.
spmontecarlo.bsky.social
Not on this particular occasion, but certainly in the past!
spmontecarlo.bsky.social
Yeah, I'm used to a similar story (though I wonder if the story is the same for x-derivatives and θ-derivatives, which is not completely obvious). I roughly believe the story, but equally, there is such strong soft information as to just how useful gradients are, so it's hard to ignore indefinitely.
spmontecarlo.bsky.social
I had some chats lately with a couple of cosmologists, and it was really the same story at most links of the chain (gradients not quite available, supercomputers involved, emulator-curious), with the quasi-distinction that they were pretty sure that gradients would help if they were available.
spmontecarlo.bsky.social
... after how many years of "re-doing the numerics would be annoying, but clearly useful" do you (collectively) cave in, and make the leap? Anyways, as you say, there are other reasons why it's not quite so simple.
spmontecarlo.bsky.social
I very much agree with the spirit of this comment; most of my other thoughts relate to some other points which have come up here (e.g. memory as the real bottleneck in this and other applications). What I found myself reflecting on was that if the only bottleneck had been effort (roughly), then ...
spmontecarlo.bsky.social
At dinner, but in the mean time, I could maybe advise to look up Marc Bocquet (who maybe has some nice slides IIRC).
spmontecarlo.bsky.social
... and they are often quite loud about how they don't want to have to deal with derivatives through their simulation pipeline, and so on. So in this regard, the AD-compatibility of the EnKF is clearly true in an objective sense, but almost feels "culturally wrong" if this is your main context.
spmontecarlo.bsky.social
A side point, which came up through some of this reading: some works will make the remark that the Ensemble Kalman Filter is nicely compatible with automatic differentiation. This is true on its face, but always takes a minute for me, given that the EnKF is huge in numerical weather prediction ...
spmontecarlo.bsky.social
I got back onto this topic via arxiv.org/abs/2312.05910; an early paper is proceedings.mlr.press/v9/turner10a..., and a couple of intermediate works which I found useful are proceedings.neurips.cc/paper_files/... and proceedings.mlr.press/v97/ialongo1....
spmontecarlo.bsky.social
A glimpse behind the curtain. Yet more care is needed when looking at things like variational inference with inducing variable approaches (which also occupy an interesting space vis-à-vis conditioning, randomness, etc.).
spmontecarlo.bsky.social
I got back onto this topic via arxiv.org/abs/2312.05910; an early paper is proceedings.mlr.press/v9/turner10a..., and a couple of intermediate works which I found useful are proceedings.neurips.cc/paper_files/... and proceedings.mlr.press/v97/ialongo1....
spmontecarlo.bsky.social
There are also some nice subtleties around the difference between conditioning on a quantity and fixing a quantity, which seems to often be a slightly messy point when talking about random functions. Anyways, I find it all quite satisfying to resolve.
spmontecarlo.bsky.social
This weekend, I have been doing a bit of reading about Gaussian Process State Space Models (GP-SSMs). Aside from being interesting on the modelling and inference sides, they are a remarkably good exercise for what it really means to write down a joint distribution.
spmontecarlo.bsky.social
if the list of speakers looks unfamiliar to you, then I can recommend { Tessera, Salez, van Handel, Magee, Sahasrabudhe } as some who I can vouch for having given good talks, one way or another
spmontecarlo.bsky.social
very cool workshop last week - www.newton.ac.uk/event/oggw03/ - lots of talks from good speakers on interesting works. keen to watch my way through them when I get a moment!
spmontecarlo.bsky.social
Another one, drafted for the same workshop, but which didn't quite make it - but which I hope will see some use at a future event!
spmontecarlo.bsky.social
It was fun! Mostly tutorials by the physicists for the statisticians; quite a success from that point of view. Recordings here: youtube.com/playlist?lis...
CoSInES-Bayes4Health Masterclass on Computational Physics, April 2024 - YouTube
youtube.com
spmontecarlo.bsky.social
A drafted logo for a past workshop (we ended up going with something else in the end, but I still find this one cute in its own way).
Reposted by Sam Power
spmontecarlo.bsky.social
I'll order these so that I can tick them off a list on my computer; the numbers are not any sort of ranking. Summaries will be kept brief and hopefully stoke curiosity, rather than providing answers.
spmontecarlo.bsky.social
Have been rather quiet on here recently (for uninteresting and unconcerning reasons), but this shouldn't be taken as a lack of enthusiasm for papers which have been coming out recently - loads of really clever and creative works popping up. Will try to post about some of them this afternoon!
spmontecarlo.bsky.social
This is indeed in the works (after combing through some of my folders of notes), but in the interim, I can share a few things which I've put up directly as .pdf files on my website (sites.google.com/view/sp-mont...), rather than as blog posts per se. Notes 3-5 are 'new'.
spmontecarlo.bsky.social
Regrettable that this reads to me as cognate to "cash my gold".
spmontecarlo.bsky.social
I guess it's really "zero-variance when it should be" (which is already difficult!) rather than "zero-variance in general" (which would be a miracle). A really nice work though; resolves a problem which I'm sure has bothered plenty of people (myself included) for some time.