Uthsav Chitra
uthsav.bsky.social
Uthsav Chitra
@uthsav.bsky.social
Assistant Professor at JHU CS developing statistical/ML methods for biological applications
GASTON algorithm: parametrize functions h,d with neural nets and learn from data!

Fun back-story: @braphael.bsky.social and I derived most of this model at the bar near an NCI workshop 🥂😅
January 24, 2025 at 9:27 PM
Model implicitly accounts for sparsity by “pooling” measurements across locations (x,y) with equal isodepth d(x,y).

These locations look like contours of equal height on an elevation map, hence the “topographic map” analogy.
January 24, 2025 at 9:27 PM
We handle sparsity w/ two assumptions:

(1) genes have *shared* gradient directions, i.e. each gradient ▽f_g(x,y) is proportional to shared vector field v(x,y)
(Equivalent to Jacobian of f being rank-1 everywhere)

(2) vector field v has no “curl”, so v=▽d is gradient of "spatial potential" d
January 24, 2025 at 9:27 PM
You can view ST data as samples of function f: R^2 → R^G, where f(x,y) is (high-dim) gene expression vector at location (x,y).

Spatial gradients are gradients ▽f_g of each component (gene)

Unfortunately, large data sparsity means naive estimation of gradient ▽f_g is very noisy 😱
January 24, 2025 at 9:27 PM