banner
anaymehrotra.bsky.social
@anaymehrotra.bsky.social
PhD candidate @ Yale | Undergrad @ IITK | anaymehrotra.com

Learning Theory, Missing Data, Generation
We want language models that do not hallucinate

We want language models that have breadth (i.e., no mode-collapse)

Jon [email protected] asked: Can we get both?

Alkis Kalavasis, Grigoris Velegkas, and I show this is impossible: arxiv.org/abs/2411.09642

🧵(1/3)
November 25, 2024 at 7:29 PM