Projects per year
Abstract
Reinforcement learning is a powerful framework for modelling the cognitive and neural substrates of learning and decision making. Contemporary research in cognitive neuroscience and neuroeconomics typically uses value-based reinforcement-learning models, which assume that decision-makers choose by comparing learned values for different actions. However, another possibility is suggested by a simpler family of models, called policy-gradient reinforcement learning. Policy-gradient models learn by optimizing a behavioral policy directly, without the intermediate step of value-learning. Here we review recent behavioral and neural findings that are more parsimoniously explained by policy-gradient models than by value-based models. We conclude that, despite the ubiquity of ‘value’ in reinforcement-learning models of decision making, policy-gradient models provide a lightweight and compelling alternative model of operant behavior.
| Original language | English |
|---|---|
| Pages (from-to) | 114-121 |
| Number of pages | 8 |
| Journal | Current Opinion in Behavioral Sciences |
| Volume | 41 |
| DOIs | |
| Publication status | Published - Oct 2021 |
Projects
- 1 Finished
-
Towards a neurocomputational model of mood instability in psychiatric illness
Bennett, D. (Primary Chief Investigator (PCI)), Sundram, S. (Supervisor) & Niv, Y. (Supervisor)
NHMRC - National Health and Medical Research Council (Australia)
1/01/19 → 31/12/22
Project: Research