A python implementation of Tangled Program Graphs. A graph based genetic programming algorithm.
Trainer Parameters: rTeamPopSize removed, just use teamPopSize and use the rootBasedPop parameter (boolean) to affect whether the population is based on all teams or just root teams.
- numba
- numpy
- clone this repo
- install with
pip install -e .
We will go over basic usage here. For more examples and details go to the examples page.
For practically all uses, you must import Trainer and Agent.
from tpg.trainer import Trainer
from tpg.agent import Agent
Trainer is the class used for training, to do all the standard tasks of an evolutionary algorithm. Such as getting individuals, evolving, and reporting stats. Agent is essentially a wrapper for an individual of the population that the user (you) gets access to. Its API allows for things like selecting an action based on the world state, and giving a reward.
Create an instance of the trainer, which initializes the necessary populations for the algorithm. There are many parameters to choose for the initializer, actions
is the only required one, which is a list of actions available to you from the chosen environment, or an int representing the number of continuous action ranging from
0-1.
trainer = Trainer(actions=range(7))
Withdraw the agents from the population to be used in some task.
agents = trainer.getAgents()
Some agent can act on the current state of the environment by the act method.
action = agent.act(state) # call every timestep
TPG works with final score only (for now), so track the score of whatever environment is being used, and reward it to the agent after the episode.
agent.reward(score, environmentName)
After all agents performed, call evolution on the trainer, and repeat for next generation if desired. environmentName
must be in a list because there is support for multiple environments/tasks to consider when evolving.
trainer.evolve([environmentName])
The above were just some of the important functions, and left out some necessary code for the environment, and a few other ways to use this API (some perhaps better). There are different ways to do things like withdrawing agents, rewarding agents, and evolving. And things become a bit tricky if you wish to work with multiprocessing, but its relatively straight forward to make it work. See the examples page for details.
- Implement a sort of memory module with some learners reading and some writing.
- Add convenience code to the utils file, such a simple method to run an agent in a gym environment with certain parameters, a method to transform the state, etc.
- https://github.com/Ryan-Amaral/general-game-playing-tpg (old API version)
- https://github.com/Ryan-Amaral/prosthetic-challenge-tpg (old API version)
- https://github.com/Ryan-Amaral/nips-l2m-2019 (private untill competition end)
- Send me a message or a pull request for your projects to be included.
- Dr. Malcolm Heywood: My undergrad and Masters research supervisor.
- Robert Smith: Provided an initial Java implementation of TPG, whiched I based this version on.
- Richard Wardin: Helped fix a few bugs in early development.