Leo Zhang
leoeleoleo1.bsky.social
Leo Zhang
@leoeleoleo1.bsky.social
PhD student at @oxfordstatistics.bsky.social working on generative models
For future work, we're excited about exploring the design space of SymDiff for improving molecular/protein diffusion models

Me and my co-authors will also be at ICLR presenting this work, so feel free to come chat with us to discuss more about SymDiff!
March 4, 2025 at 3:31 PM
In addition, as a consequence of the flexibility allowed by SymDiff, we report much lower computational costs compared to message-passing based EGNNs
(8/9)
March 4, 2025 at 3:31 PM
We note that the generality of our methodology should also allow us to apply SymDiff as a drop-in replacement to these other methods as well
March 4, 2025 at 3:31 PM
We show substantial improvements over EDM on QM9, GEOM-Drugs, and competitive performance with much more sophisticated recent baselines (which all use intrinsically equivariant architectures)
March 4, 2025 at 3:31 PM
To validate our methodology, we apply SymDiff to molecular generation with E(3)-equivariance

We take the basic molecule generation framework of EDM (Hoogeboom et al. 2022) and use SymDiff with Diffusion Transformers as a drop-in replacement for the EGNN they use in their reverse process
March 4, 2025 at 3:31 PM
✅ We show how to bypass previous requirements of a intrinsically equivariant sub-network with the Haar measure, for even more flexibility
✅ We also sketch how to extend SymDiff to score and flow-based models as well
March 4, 2025 at 3:31 PM
✅ We derive a novel objective to train our new model
✅ We overcome previous issues with symmetrisation concerning pathologies of cannonicalisation, and the computational cost and errors involved in evaluating frame averaging/probabilistic symmetrisation
March 4, 2025 at 3:31 PM
This provides us with the mathematical foundations to build SymDiff

Our key idea is to "symmetrise" the reverse kernels of a diffusion model to build equivariance with unconstrained architectures
March 4, 2025 at 3:31 PM
This is where recent work from Cornish (2024) (arxiv.org/abs/2406.11814) comes in. This generalises all previous work and extends it to the stochastic case using category-theoretic arguments, under the name of "stochastic symmetrisation"
March 4, 2025 at 3:31 PM
A key limitation of this line of work however is that they only consider deterministic functions - i.e. they cannot convert the stochastic kernels involved in generative modelling to be equivariant
March 4, 2025 at 3:31 PM
This decouples prediction from equivariance constraints, benefiting from the flexibility of the arbitrary base model, which doesn't suffer from the constraints of intrinsically equivariant architectures.
March 4, 2025 at 3:31 PM
As an alternative, people have proposed cannonicalisation/frame averaging/probabilistic symmetrisation which convert arbitrary networks to be equivariant through a learned group averaging
March 4, 2025 at 3:31 PM
One common criticism of (intrinsically) equivariant architectures is that due to the architecture constraints required to ensure equivariance, they suffer from worse expressivity, greater computational cost and implementation complexity
March 4, 2025 at 3:31 PM
There's been a lot of recent interest in whether equivariant models are really needed for data containing symmetries (think AlphaFold 3)
March 4, 2025 at 3:31 PM