This code is for the runs in our work, "Transformer as a hippocampal memory consolidation model based on NMDAR-inspired nonlinearity". In our work, all runs are performed on a single NVIDIA TITAN V GPU.
Supported platforms: MacOS and Ubuntu, Python 3.8
Installation using Miniconda:
conda create -y --name nmda python=3.8
conda activate nmda
pip install -r requirements.txt
python main.py --run_dir ./runs --group_name nmda_experiments --alpha 0.1 --num_envs 32 --log_to_wandb
-
alpha
is the parameter for the$\text{NMDA}_\alpha$ activation function we used in our experiment. -
num_envs
is the number of training maps ($N$ in our paper).