Monte Carlo Predictive Coding (MCPC): Learning probability distributions of sensory inputs with local synaptic plasticity
This repository houses the implementation and experiments supporting the findings presented in our paper, where we propose a novel computational framework called Monte Carlo Predictive Coding (MCPC). Our work is motivated by the hypothesis that the brain uses probabilistic generative models to interpret sensory information optimally. Previous theories in neuroscience, namely predictive coding and neural sampling, have offered partial explanations for how the brain might achieve this. Predictive coding suggests that networks of neurons can learn generative models through local synaptic plasticity, while neural sampling theories demonstrate how stochastic dynamics enable the representation of posterior distributions of the environment's latent states.
MCPC bridges these two theoretical frameworks, presenting a comprehensive model that leverages the strengths of both predictive coding and neural sampling. Our findings show that MCPC-equipped neural networks can learn accurate generative models using local computation and synaptic plasticity mechanisms. These networks are capable of inferring the posterior distributions of latent states based on sensory inputs and can simulate likely sensory inputs in their absence. Importantly, our model also aligns with experimental observations of neural activity variability during perceptual tasks, providing a unified theory of cortical computation that accounts for a broad range of neural data.
- Integration of Predictive Coding and Neural Sampling: We detail the theoretical underpinnings and practical implementation of combining predictive coding with neural sampling into a coherent framework, MCPC, which enhances our understanding of cortical computation.
- Local Computation and Plasticity: Our model demonstrates how precise generative models can be learned through mechanisms that are implementable within the biological constraints of neural networks.
- Alignment with Experimental Observations: MCPC not only offers theoretical contributions but also provides a compelling match to experimental data on neural variability, supporting its relevance and applicability to understanding brain function.
The repository includes:
figure_2.py
,figure_3.py
,figure_4.py
,figure_5.py
,figure_6.py
, andtable_1.py
contain the code to recreate the figures and the table from our paper.requirements.txt
contains the python dependencies of this repository.figures/
contains the output figures of the code.models/
contains trained models to generate figures.utils/
contains utility functionspredictive_coding/
contains the code to simulate MCPC- and PC- models.Deep_Latent_Gaussian_Models/
contains code to simulate DLGMs.ResNet.py
contains the code to simulate ResNet-9 models.
Follow the following steps to clone the code and setup the necessary python libraries:
git clone https://github.com/gaspardol/MonteCarloPredictiveCoding.git
cd MonteCarloPredictiveCoding
pip install -r requirements.txt
To generate the figures of the paper please run
python figure_2.py
python figure_3.py
python figure_4.py
python figure_5.py
python figure_6.py
python table_1.py
For those who find our work useful, here is how you can cite it:
@article {Oliviers2024,
author = {Gaspard Oliviers and Rafal Bogacz and Alexander Meulemans},
title = {Learning probability distributions of sensory inputs with Monte Carlo Predictive Coding},
elocation-id = {2024.02.29.581455},
year = {2024},
doi = {10.1101/2024.02.29.581455},
publisher = {Cold Spring Harbor Laboratory},
URL = {https://www.biorxiv.org/content/early/2024/03/02/2024.02.29.581455},
eprint = {https://www.biorxiv.org/content/early/2024/03/02/2024.02.29.581455.full.pdf},
journal = {bioRxiv}
}
For any inquiries or questions regarding the project, please feel free to contact Gaspard Oliviers at gaspard.oliviers@pmb.ox.ac.uk.
This repository builds upon the following repositories/codes: