@kyunghyuncho.bsky.social
3.3K followers 330 following 370 posts
a mediocre combination of a mediocre AI scientist, a mediocre physicist, a mediocre chemist, a mediocre manager and a mediocre professor. see more at https://kyunghyuncho.me/
Posts Media Videos Starter Packs
Reposted
nicholaslourie.bsky.social
📄🔈✨ Deep learning is an empirical science, but we rely on basic empirical methods. What might a better foundation—a simple theory—for empirical work look like?

@kyunghyuncho.bsky.social, He He, and I move towards one in "Hyperparameter Loss Surfaces Are Simple Near their Optima" at #COLM2025!

🧵1/9
kyunghyuncho.bsky.social
i'll be at COLM on thursday (oct 9). DM me for a chat!
kyunghyuncho.bsky.social
the past few weekends here and there spread through a few months have been fun for me. i vibe-coded and along the way learned various services that enable one to rapidly deploy a trial version of a web service.

an awsome era for software development!
Reposted
mila-quebec.bsky.social
Exciting news! We're thrilled to announce the appointment of Professor Hugo Larochelle as Mila's new Scientific Director! A deep learning pioneer and former head of Google's AI lab in Montreal, Hugo's leadership will be pivotal in advancing AI for the benefit of all.

mila.quebec/en/news/hugo...
Hugo Larochelle becomes the new Scientific Director of Mila | Mila
Montreal (Quebec), September 2, 2025 – Mila, the Quebec Artificial Intelligence Institute, announces the appointment of Hugo Larochelle—Adjunct Professor at the Université de Montréal and former head ...
mila.quebec
kyunghyuncho.bsky.social
amen

those with savior and superiority complex, obsessed with sci fi, blinded by dollar signs, devoid of empathy, and severely gullible.
Reposted
furiosa.ai
Thank you to everyone in the AI community we met in NYC this summer and those who came to our happy hour at @nyu.edu’s Global AI Frontier Lab last week.
kyunghyuncho.bsky.social
everyone says google will lose search biz to llm, but in reality, all llm co's are so busy trying to reproduce google search. irony
kyunghyuncho.bsky.social
i meant each entry to a leaderboard is a hypothesis. in LM Arena, each entry is a chatbot product without any details, and i don’t believe we can consider it a scientific hypothesis.
kyunghyuncho.bsky.social
but it's not a leaderboard in the context of a scientific experimental paradigm, since no scientific hypothesis needs to be stated anywhere to participate. who knows what these models are, how they were trained, etc.?
kyunghyuncho.bsky.social
LMArena is a nice way to identify models in real time that are preferred by participants to this scheme. totally legit under this context.

...
kyunghyuncho.bsky.social
recently gave a talk on <Reality Checks> at two venues, and discussed (and rambled) about how leaderboard chasing is awesome (and we want it to continue) but that this isn't easy because everyone (me! me! me!) wants to write more papers.

the link to the slide deck in the reply.
kyunghyuncho.bsky.social
can’t really vibe code healthcare. a thoughtful and experience-rich talk by amazing Lavender Jiang at NYU Global AI Frontier Lab!
kyunghyuncho.bsky.social
vibe coding continues with a MS Edge extension: the world's most expensive, slow and inaccurate ad blocker, featuring Gemini!
kyunghyuncho.bsky.social
the perfect response from Google cannot be simpler: just halve the price of everything.

please start with the Gemini Pro subscription. pretty please?
Reposted
lampinen.bsky.social
These biases can lead to dramatic downstream effects that cause unexpected conclusions from analyses. For example, RSA may identify two models computing the same, complex task as much less representationally-similar than either of them is to a model computing a much simpler task (right panel)!
RSA within and between different sets of models can give surprising results due to representation biases. This plot shows similarities within and between different models computing different types of features. Ideally the similarities would be highest in blocks on the diagonal (i.e. models computing the same features), and the blocks off the diagonal would show graded similarity corresponding to the functional overlap. However, that is not the case. (Left) When comparing a model trained to output both easy and hard features to ones that are trained on only one feature, the multi-task model appears very similar to the easy-task only model (cf. Hermann and Lampinen, 2020). In fact, the models trained only on the hard task do not even appear particularly similar to other models trained on the same exact task. (Right) When models are trained on multiple easy or multiple hard tasks, the models trained on only hard tasks appear less similar to other models trained on exactly the same tasks than they do to models trained on strictly easier tasks that use the same input units.
Reposted
furiosa.ai
🛬 Next week, we’re hosting our inaugural happy hour: Powered by Furiosa.

🍹 On August 8, we’re bringing together industry leaders and practitioners at the forefront of AI to discuss, share knowledge, and enjoy great food and drinks.
Powered by Furiosa | LinkedIn
Join us at our inaugural happy hour: Powered by Furiosa. This event will bring together industry practitioners and leaders at the forefront of AI inference. Connect with NYC researchers, engineers, a...
www.linkedin.com
Reposted
eugenevinitsky.bsky.social
You know how RL is that whole big thing nowadays?

Present your work at the first-ever New York Reinforcement Learning Workshop (NYRL), co-organized by Amazon, Columbia Business School & NYU Tandon School of Engineering.

ny-rl.com!
Poster announcing NYC RL day!
kyunghyuncho.bsky.social
😂 i think this “challenging problem” may have been finally solved after five years.

===
Understanding and creating mathematics using natural mathematical language … used by humans is a challenging and important problem for driving progress in machine learning.
===