/ddpg-collaboration

Using a DDPG algorithm to solve a multi-agent reinforcement learning problem with collaboration

Primary LanguageJupyter Notebook

DDPG Algorithm - Colaboration


This work is part of the Udacity Deep Reinforcement Learning Nanodegree third assignment, which consists on solving the Tennis environment.

tennis-env

Setup


To setup your local Python environment for running Unity environments checkout the instructions on this Github repository. On this work we'll use PyTorch to build the networks. On requirements.txt you'll also find some other packages required.

Environment


This work does not require to install Unity, the environment is already been built, and you can download it from the link below:

Then you must place the environment inside the env folder, or update the path on the notebook, if you wish to reproduce the report.ipynb.

Development


You should follow report.ipynb for the detailed implementation process. The models folder holds all the model files that was used, and utils folder has the support files, such as the noise and replay buffer implementations. On the agent.py file the main agent is implemented, the one responsible for creating and training both the actor and critic networks.

Trained model


You can use the trained model by loading the parameters from model_parameters folder to both the actor and critic network, and then acting on the environment.

Future work


This work can be improved by testing another noise functions, as well as implementing Prioritized Experience Replay.