Meenakshi Khosla
meenakshikhosla.bsky.social
Meenakshi Khosla
@meenakshikhosla.bsky.social
Assistant Professor at UCSD Cognitive Science and CSE (affiliate) | Past: Postdoc @MIT, PhD @Cornell, B. Tech @IITKanpur | Interested in Biological and Artificial Intelligence
@andre-longon.bsky.social led/executed this project beautifully—he's applying to PhD programs this fall and would be an incredible addition to any lab!
October 8, 2025 at 8:55 PM
also thanks to @david-klindt.bsky.social
for an incredible collaboration.
October 8, 2025 at 8:55 PM
The takeaway: superposition isn’t just an interpretability issue—it warps alignment metrics too. Disentangling reveals the true representational overlap between models and between models and brains.
October 8, 2025 at 8:55 PM
Across toy models, ImageNet DNNs (ResNet, ViT), and even brain data (NSD), alignment scores jump once we replace base neurons with their disentangled SAE latents—showing that superposition can mask shared structure.
October 8, 2025 at 8:55 PM
We develop a theory showing how superposition arrangements deflate predictive-mapping metrics. Then we test it: disentangling with sparse autoencoders (SAEs) reveals hidden correspondences.
October 8, 2025 at 8:54 PM