Restoring activity in the mediodorsal thalamus rescued the behavior, validating the circuit mechanism.
Restoring activity in the mediodorsal thalamus rescued the behavior, validating the circuit mechanism.
Uncertainty suppresses plasticity to prevent errors, while certainty releases the brakes to encode new rules.
Uncertainty suppresses plasticity to prevent errors, while certainty releases the brakes to encode new rules.
Biological heuristics like dynamic sparsity modulation allowed for faster adaptation than rigid mathematical formulas.
Biological heuristics like dynamic sparsity modulation allowed for faster adaptation than rigid mathematical formulas.
This circuit maintained a low-dimensional representation of likelihood that tracked the changing environment.
This circuit maintained a low-dimensional representation of likelihood that tracked the changing environment.
They showed the basal ganglia encodes uncertainty as a probability distribution used for exploration (Fig 2f).
Synapses representing different value quantiles allowed the model to balance risk and reward efficiently.
They showed the basal ganglia encodes uncertainty as a probability distribution used for exploration (Fig 2f).
Synapses representing different value quantiles allowed the model to balance risk and reward efficiently.
Hyperactive D2 receptors in BG unbalance activity in Thalamus, making the "Belief keeper" unstable.
This results in a unstable attractor, sensitive to noise and that cannot hold onto evidence.
Hyperactive D2 receptors in BG unbalance activity in Thalamus, making the "Belief keeper" unstable.
This results in a unstable attractor, sensitive to noise and that cannot hold onto evidence.
This hierarchy allows the mouse to instantly swap strategies without unlearning everything when the world changes.
E.g. PFC has evidence of a wrong choice.
This hierarchy allows the mouse to instantly swap strategies without unlearning everything when the world changes.
E.g. PFC has evidence of a wrong choice.
PFC helped by the mediodorsal (MD) thalamus "belief keeper" who uses attractor dynamics to hold a stable belief until enough evidence pushes it to a new conclusion.
PFC helped by the mediodorsal (MD) thalamus "belief keeper" who uses attractor dynamics to hold a stable belief until enough evidence pushes it to a new conclusion.
The basal ganglia "notebook": stores payout probabilities as full distributions, not just averages.
Premotor cortex: samples from these notes to explore options when the outcome is uncertain.
Motor cortex: compares values and takes action.
The basal ganglia "notebook": stores payout probabilities as full distributions, not just averages.
Premotor cortex: samples from these notes to explore options when the outcome is uncertain.
Motor cortex: compares values and takes action.
which arm pays out right now?, and
have the hidden rules changed?
which arm pays out right now?, and
have the hidden rules changed?
This model tries to explain decision impairments by hyperactive D2 receptors
We can clarify the paper by building a mouse with schizophrenia.
A🧵with my toy model and notes:
#neuroskyence #compneuro #NeuroAI
This model tries to explain decision impairments by hyperactive D2 receptors
We can clarify the paper by building a mouse with schizophrenia.
A🧵with my toy model and notes:
#neuroskyence #compneuro #NeuroAI
Upon reward delivery, it stops predicting force and starts predicting the *rate of licking* (Fig 9h, 9j).
Upon reward delivery, it stops predicting force and starts predicting the *rate of licking* (Fig 9h, 9j).
Inhibiting DA neurons during the task *did not impair* learning (Fig 10b, 10i).
Inhibiting DA neurons during the task *did not impair* learning (Fig 10b, 10i).
It even held true during an aversive air puff, proving it's not about "reward" (Fig 3e, 3h).
It even held true during an aversive air puff, proving it's not about "reward" (Fig 3e, 3h).
They identified two distinct DA neuron types: "Forward" and "Backward" populations.
These cells fire to drive movement in a specific direction (Fig 1e, 1h, 1k).
They identified two distinct DA neuron types: "Forward" and "Backward" populations.
These cells fire to drive movement in a specific direction (Fig 1e, 1h, 1k).
RPE view: DA activity scales with reward probability, encoding the *strength* of the prediction.
New view: Probability changes the animal's *effort*, and the DA signal simply tracks that performance.
RPE view: DA activity scales with reward probability, encoding the *strength* of the prediction.
New view: Probability changes the animal's *effort*, and the DA signal simply tracks that performance.
RPE view: A larger reward causes a larger DA spike because it's a bigger "positive error."
New view: A larger reward makes the mouse push *harder*, and the DA spike just tracks that *vigor*.
RPE view: A larger reward causes a larger DA spike because it's a bigger "positive error."
New view: A larger reward makes the mouse push *harder*, and the DA spike just tracks that *vigor*.
RPE view: The famous "dip" in DA when a reward is omitted is a negative prediction error.
New view: The dip simply reflects the animal *abruptly stopping* its forward movement.
RPE view: The famous "dip" in DA when a reward is omitted is a negative prediction error.
New view: The dip simply reflects the animal *abruptly stopping* its forward movement.
Why does the DA signal "move" to the cue?
RPE view: This helps an animal learn what things are important.
New view: the signal shifts because the animal's *action* (pushing forward) shifts to the cue.
Why does the DA signal "move" to the cue?
RPE view: This helps an animal learn what things are important.
New view: the signal shifts because the animal's *action* (pushing forward) shifts to the cue.
The classical RPE model says DA neurons encode the *difference* between expected and actual rewards.
A surprise spikes DA activity, while disappointment causes a dip.
This helps a mouse 🐭 learn what to pay attention to.
The classical RPE model says DA neurons encode the *difference* between expected and actual rewards.
A surprise spikes DA activity, while disappointment causes a dip.
This helps a mouse 🐭 learn what to pay attention to.
If DA RPE is the emperor, this work SCREAMS it was running naked all the time.
This paper got quite some attention recently. Let's simplify it a bit.
A🧵with my toy model and notes:
#neuroskyence #compneuro #NeuroAI
If DA RPE is the emperor, this work SCREAMS it was running naked all the time.
This paper got quite some attention recently. Let's simplify it a bit.
A🧵with my toy model and notes:
#neuroskyence #compneuro #NeuroAI
The same neural architecture used to *remember* a sequence can be used to *infer* a plan (Fig 2B).
The same neural architecture used to *remember* a sequence can be used to *infer* a plan (Fig 2B).
Sensory input about a wall specifically inhibited the neural representation for that impossible transition (Fig 6H).
Sensory input about a wall specifically inhibited the neural representation for that impossible transition (Fig 6H).