Chrome Extension
WeChat Mini Program
Use on ChatGLM

Common population codes produce extremely nonlinear neural manifolds

Anandita De,Rishidev Chaudhuri

PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA(2023)

Cited 1|Views5
No score
Abstract
Populations of neurons represent sensory, motor and cognitive variables via patterns of activity distributed across the population. The size of the population used to encode a variable is typically much greater than the dimension of the variable itself, and thus the corresponding neural population activity occupies lower-dimensional subsets of the full set of possible activity states. Given population activity data with such lower-dimensional structure, a fundamental question asks how close the low-dimensional data lies to a linear subspace. The linearity or non-linearity of the low-dimensional structure reflects important computational features of the encoding, such as robustness and generalizability. Moreover, identifying such linear structure underlies common data analysis methods such as Principal Component Analysis. Here we show that for data drawn from many common population codes the resulting point clouds and manifolds are exceedingly nonlinear, with the dimension of the best-fitting linear subspace growing at least exponentially with the true dimension of the data. Consequently, linear methods like Principal Component Analysis fail dramatically at identifying the true underlying structure, even in the limit of arbitrarily many data points and no noise. ### Competing Interest Statement The authors have declared no competing interest.
More
Translated text
Key words
nonlinear neural manifolds,common population codes
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined