Official code for the AAAI 2020 paper "Proximal Distilled Evolutionary Reinforcement Learning".
Use the following to cite:
@article{Bodnar2019ProximalDE,
title={Proximal Distilled Evolutionary Reinforcement Learning},
author={Cristian Bodnar and Ben Day and Pietro Lio'},
journal={ArXiv},
year={2019},
volume={abs/1906.09807}
}
First, you will have to install all the dependencies by running pip install -r requirements.txt
.
Additionally, for installing mujoco-py 2.0.2.2, follow the instructions on the official github.
To run PDERL with proximal mutations and distillation-based crossover use:
python run_pderl.py -env=$ENV_NAME$ -distil -proximal_mut -mut_mag=$MUT_MAG$ -logdir=$LOG_DIR$
To evaluate and visualise a trained model in an environment use:
python play_pderl.py -env=$ENV_NAME$ -model_path=$MODEL_PATH$ -render
'Hopper-v2'
'HalfCheetah-v2'
'Swimmer-v2'
'Ant-v2'
'Walker2d-v2'
Our code is largely based on the code of Khadka and Tumer and we would like to thank them for making their code publicly available. The proximal mutations code is also relying on the safe mutations code of Lehman et al. from Uber Research.