Monkey Prefrontal Cortex Learns to Minimize Sequence Prediction Error

bioRxiv the preprint server for biology(2024)

Cited 0|Views4
No score
Abstract
In this study, we develop a novel recurrent neural network (RNN) model of pre-frontal cortex that predicts sensory inputs, actions, and outcomes at the next time step. Synaptic weights in the model are adjusted to minimize sequence prediction error, adapting a deep learning rule similar to those of large language models. The model, called Sequence Prediction Error Learning (SPEL), is a simple RNN that predicts world state at the next time step, but that differs from standard RNNs by using its own prediction errors from the previous state predictions as inputs to the hidden units of the network. We show that the time course of sequence prediction errors generated by the model closely matched the activity time courses of populations of neurons in macaque prefrontal cortex. Hidden units in the model responded to combinations of task variables and exhibited sensitivity to changing stimulus probability in ways that closely resembled monkey prefrontal neurons. Moreover, the model generated prolonged response times to infrequent, unexpected events as did monkeys. The results suggest that prefrontal cortex may generate internal models of the temporal structure of the world even during tasks that do not explicitly depend on temporal expectation, using a sequence prediction error minimization learning rule to do so. As such, the SPEL model provides a unified, general-purpose theoretical framework for modeling the lateral prefrontal cortex. ### Competing Interest Statement The authors have declared no competing interest.
More
Translated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined