| dc.contributor.author | Tomov, Momchil S | |
| dc.contributor.author | Tsividis, Pedro A | |
| dc.contributor.author | Pouncy, Thomas | |
| dc.contributor.author | Tenenbaum, Joshua B | |
| dc.contributor.author | Gershman, Samuel J | |
| dc.date.accessioned | 2023-04-04T17:13:32Z | |
| dc.date.available | 2023-04-04T17:13:32Z | |
| dc.date.issued | 2023-03 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/150405 | |
| dc.description.abstract | Humans learn internal models of the world that support planning and generalization in complex environments. Yet it remains unclear how such internal models are represented and learned in the brain. We approach this question using theory-based reinforcement learning, a strong form of model-based reinforcement learning in which the model is a kind of intuitive theory. We analyzed fMRI data from human participants learning to play Atari-style games. We found evidence of theory representations in prefrontal cortex and of theory updating in prefrontal cortex, occipital cortex, and fusiform gyrus. Theory updates coincided with transient strengthening of theory representations. Effective connectivity during theory updating suggests that information flows from prefrontal theory-coding regions to posterior theory-updating regions. Together, our results are consistent with a neural architecture in which top-down theory representations originating in prefrontal regions shape sensory predictions in visual areas, where factored theory prediction errors are computed and trigger bottom-up updates of the theory. | en_US |
| dc.language.iso | en | |
| dc.publisher | Elsevier BV | en_US |
| dc.relation.isversionof | 10.1016/j.neuron.2023.01.023 | en_US |
| dc.rights | Creative Commons Attribution-NonCommercial-NoDerivs License | en_US |
| dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/4.0/ | en_US |
| dc.source | bioRxiv | en_US |
| dc.title | The neural architecture of theory-based reinforcement learning | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Tomov, Momchil S, Tsividis, Pedro A, Pouncy, Thomas, Tenenbaum, Joshua B and Gershman, Samuel J. 2023. "The neural architecture of theory-based reinforcement learning." Neuron. | |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences | en_US |
| dc.relation.journal | Neuron | en_US |
| dc.eprint.version | Original manuscript | en_US |
| dc.type.uri | http://purl.org/eprint/type/JournalArticle | en_US |
| eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
| dc.date.updated | 2023-04-04T17:05:19Z | |
| dspace.orderedauthors | Tomov, MS; Tsividis, PA; Pouncy, T; Tenenbaum, JB; Gershman, SJ | en_US |
| dspace.date.submission | 2023-04-04T17:05:37Z | |
| mit.license | OPEN_ACCESS_POLICY | |
| mit.metadata.status | Authority Work and Publication Information Needed | en_US |