A Python package used for simulating spiking neural networks (SNNs) using PyTorch GPU functionality.
BindsNET is a spiking neural network simulation software, a critical component enabling the modeling of neural systems and the development of biologically inspired algorithms in the machine learning domain.
This package is used as part of ongoing research on applying SNNs to machine learning (ML) and reinforcement learning (RL) problems in the Biologically Inspired Neural & Dynamical Systems (BINDS) lab.
- Python 3.6
BindsNET is available on PyPI. Issue
pip install bindsnet
to get the most recent stable release. Or, to build the bindsnet
package from source, clone the GitHub repository, change directory to the top level of this project, and issue
pip install .
Or, to install in editable mode (allows modification of package without re-installing):
pip install -e .
To install the packages necessary to interface with the OpenAI gym RL environments library, follow their instructions for installing the packages needed to run the RL environments simulator (on Linux / MacOS).
To run a near-replication of the SNN from this paper, issue
cd examples/mnist
python eth_mnist.py
There are a number of optional command-line arguments which can be passed in, including --plot
(displays useful monitoring figures), --n_neurons [int]
(number of excitatory, inhibitory neurons simulated), --mode ['train' | 'test']
(sets network operation to the training or testing phase), and more. Run the script with the --help
or -h
flag for more information.
A number of other examples are available in the examples
directory that are meant to showcase BindsNET's functionality. Take a look, and let us know what you think!
Issue the following to run the tests:
python -m pytest test/
Some tests will fail if Open AI gym
is not installed on your machine.
The simulation of biologically plausible spiking neuron dynamics can be challenging. It is typically done by solving ordinary differential equations (ODEs) which describe said dynamics. PyTorch does not explicitly support the solution of differential equations (as opposed to brian2
, for example), but we can convert the ODEs defining the dynamics into difference equations and solve them at regular, short intervals (a dt
on the order of 1 millisecond) as an approximation. Of course, under the hood, packages like brian2
are doing the same thing. Doing this in PyTorch
is exciting for a few reasons:
-
We can use the powerful and flexible
torch.Tensor
object, a wrapper around thenumpy.ndarray
which can be transferred to and from GPU devices. -
We can avoid "reinventing the wheel" by repurposing functions from the
torch.nn.functional
PyTorch submodule in our SNN architectures; e.g., convolution or pooling functions.
The concept that the neuron spike ordering and their relative timing encode information is a central theme in neuroscience. Markram et al. (1997) proposed that synapses between neurons should strengthen or degrade based on this relative timing, and prior to that, Donald Hebb proposed the theory of Hebbian learning, often simply stated as "Neurons that fire together, wire together." Markram et al.'s extension of the Hebbian theory is known as spike-timing-dependent plasticity (STDP).
We are interested in applying SNNs to ML and RL problems. We use STDP to modify weights of synapses connecting pairs or populations of neurons in SNNs. In the context of ML, we want to learn a setting of synapse weights which will generate data-dependent spiking activity in SNNs. This activity will allow us to subsequently perform some ML task of interest; e.g., discriminating or clustering input data. In the context of RL, we may think of the spiking neural network as an RL agent, whose spiking activity may be converted into actions in an environment's action space.
We have provided some simple starter scripts for doing unsupervised learning (learning a fully-connected or convolutional representation via STDP), supervised learning (clamping output neurons to desired spiking behavior depending on data labels), and reinforcement learning (converting observations from the Atari game Space Invaders to input to an SNN, and converting network activity back to actions in the game).
Hananel Hazan, Daniel J. Saunders, Hassaan Khan, Darpan T. Sanghavi, Hava T. Siegelmann, Robert Kozma, BindsNET: A machine learning-oriented spiking neural networks library in Python. 2018, Arxiv.
GNU Affero General Public License v3.0