🥈 Spreading science over hype in #ML & #NLP
Proud shareLM💬 Donor
@IBMResearch & @MIT_CSAIL
In RLDM 1900BC or so, I was practising handstands before the conference and he passed and encouraged me.
He is a role model - friendly and ambitious with strong ideas, some wrong
In RLDM 1900BC or so, I was practising handstands before the conference and he passed and encouraged me.
He is a role model - friendly and ambitious with strong ideas, some wrong
Thanks you for your "attention"
Thanks you for your "attention"
With 25k people in the audience, who is even reading this tweet?
Can anyone estimate how many of the researchers in percentage attend neurips? 10%? how many researchers are there now?
With 25k people in the audience, who is even reading this tweet?
Can anyone estimate how many of the researchers in percentage attend neurips? 10%? how many researchers are there now?
There are some approaches but...
There are some approaches but...
We need reliable continual learning (yes! bsky.app/profile/lcho... )
We need reliable continual learning (yes! bsky.app/profile/lcho... )
Consider animal play, they set subproblems for themsleves to figure (action->) cause and effects.
Consider animal play, they set subproblems for themsleves to figure (action->) cause and effects.
You can’t have knowledge that’s only at the low level, you need high-level knowledge to integrate it all. And again big world🌍
You can’t have knowledge that’s only at the low level, you need high-level knowledge to integrate it all. And again big world🌍
maybe not everything should be called so, it is just a network, but we should aim for the networks to really model; To encompass everything. We should strive it to represent the coffee☕️, my hand the movement and everything else.
maybe not everything should be called so, it is just a network, but we should aim for the networks to really model; To encompass everything. We should strive it to represent the coffee☕️, my hand the movement and everything else.
There are atoms, physics, you are a tiny spec on earth etc.
You need a model that abstracts
There are atoms, physics, you are a tiny spec on earth etc.
You need a model that abstracts
If we look at current research most of it didn’t learn the bitter lesson (~algorithms scale and are not domain specific)
"but it is fine", 9 out of 10 studies are useless anyway, but it is hard to predict which one is the tenth one.
If we look at current research most of it didn’t learn the bitter lesson (~algorithms scale and are not domain specific)
"but it is fine", 9 out of 10 studies are useless anyway, but it is hard to predict which one is the tenth one.
LLMs do not learn from explicit corrections
LLMs do not learn from being told the answer
LLMs do not learn from being shown how to solve it
We study Machine Learning, these are opportunities!
A gold mine of research.
LLMs do not learn from explicit corrections
LLMs do not learn from being told the answer
LLMs do not learn from being shown how to solve it
We study Machine Learning, these are opportunities!
A gold mine of research.
@neuripsconf.bsky.social ), babyLMs, etc.
Open science and open research (happy to help such initiatives)
And humans, they all connect, but that's a different story
@neuripsconf.bsky.social ), babyLMs, etc.
Open science and open research (happy to help such initiatives)
And humans, they all connect, but that's a different story
So I also work on evaluation (btw anyone here care about agent evals? know someone? reach out plz)
@asaf-yehudai.bsky.social
Check @eval-eval.bsky.social for the open eval research community
So I also work on evaluation (btw anyone here care about agent evals? know someone? reach out plz)
@asaf-yehudai.bsky.social
Check @eval-eval.bsky.social for the open eval research community
We are working on learning from interaction through games as well (textarena.ai check the github etc.)
bsky.app/profile/lcho...
Explore theory of mind, game intelligence, and multi-agent LLMs in interactive game environments.
🗓 Sunday, December 7
⏰ 8:00–10:45 AM
📍 San Diego Convention Center, Ballroom 6CF
🤖📈🧠
We are working on learning from interaction through games as well (textarena.ai check the github etc.)
bsky.app/profile/lcho...
It calls learning from human-model interaction (sharelm.github.io use the data or contribute) but what else?
Have thoughts? let's talk
It calls learning from human-model interaction (sharelm.github.io use the data or contribute) but what else?
Have thoughts? let's talk
They already interact and use more compute
Yes, some scenarios require learning conflicting things (e.g. personalization)
Ok, let's start training models that fit our needs, but also share some of this knowledge across them?
They already interact and use more compute
Yes, some scenarios require learning conflicting things (e.g. personalization)
Ok, let's start training models that fit our needs, but also share some of this knowledge across them?
LLMs do not learn from explicit corrections
LLMs do not learn from being told the answer
LLMs do not learn from being shown how to solve it
We study Machine Learning, these are opportunities!
A gold mine of research.
LLMs do not learn from explicit corrections
LLMs do not learn from being told the answer
LLMs do not learn from being shown how to solve it
We study Machine Learning, these are opportunities!
A gold mine of research.
Language is just a latent variable, it shows sometimes outputs that required a skill (call it thinking if you must) to be produced.
Predicting this apparently creates another system that mimics thinking like behaviors
Language is just a latent variable, it shows sometimes outputs that required a skill (call it thinking if you must) to be produced.
Predicting this apparently creates another system that mimics thinking like behaviors