▶️ www.biorxiv.org/content/10.1...
#Neuroscience
▶️ www.biorxiv.org/content/10.1...
#Neuroscience
osf.io/preprints/ps...
osf.io/preprints/ps...
Reach out — we’d love to hear from you! 🙌
Reach out — we’d love to hear from you! 🙌
Our work aims at bridging cognitive science and machine learning, showing how human-inspired principles like reward normalization can improve reinforcement learning AI systems!
Our work aims at bridging cognitive science and machine learning, showing how human-inspired principles like reward normalization can improve reinforcement learning AI systems!
We further extend the RA model by integrating a temporal difference component to the dynamic range updates. With this extension, we demonstrate that the magnitude invariance capabilities of the RA model persist in multi-step tasks.
We further extend the RA model by integrating a temporal difference component to the dynamic range updates. With this extension, we demonstrate that the magnitude invariance capabilities of the RA model persist in multi-step tasks.
Thus, to achieve high performance, the ABS model requires tuning the 𝛽 value to the magnitudes at stake.
Thus, to achieve high performance, the ABS model requires tuning the 𝛽 value to the magnitudes at stake.
In contrast, the RA model maintains a consistent, scale-invariant performance.
In contrast, the RA model maintains a consistent, scale-invariant performance.
As expected the standard model is highly dependent on the tasks levels, while the RA model achieves high accuracy over the whole range of values tested!
As expected the standard model is highly dependent on the tasks levels, while the RA model achieves high accuracy over the whole range of values tested!
Let's now dive into the study!
Let's now dive into the study!