/tuner

An open source AutoML toolkit for neural architecture search and hyper-parameter tuning implementing multiple GPU Coordination

Primary LanguageC++OtherNOASSERTION

Asynchronous Hyperparameter Optimization on Distributed N-Node CPU/GPUs

Requirements

  • cmake 3.8.0 (minimum)
  • gcc-7.2 (minimum)
  • boost 1.65.1 (minimum)

Research Plan

The plan of this package is to build an automatic tool to do hyperparameter optimization. During our time at MILA, before and after coding up research ideas we always do hyperparameter optimization to set up a baseline on SOTA. Previous methods which are grid search and random search requires tedious work on grad student part to handpicking parameters. We realize that this process could be automated.

Another point is that we want a tool that does not necessarily work for big labs like MILA, but for labs with just 5 computers with GPUs together connected with a router or a startup company in Silicon Valley that could only afford low amount of AWS machines/GPUs, that doesn't have one reliable central node to control node coordination, new hyperparameter choice and node pruning. We plan to use a decentralize system that use effective broadcast of hyperparameter choice to each node, and also saving the hyperparameter history on a TinyDB/levelDB. Although we would not hash all of the hyperparameter blocks to ensure we didn't choose the same hyperparameters across two different nodes in the network.

The choice of new hyperparameters would be done using Bayesian optimization algorithms or evolutionary strategies/genetic algorithms. We would want to implement early stopping conditions such as Freeze-Thaw.

Execution of Research Plan

We plan to build this package that would be language agnostic, and also operation system agnostic, although at the beginning we would want to use it for mainly python scripts. This package is written mainly in C++17 and we would want to accomodate C++20.

Usage

evolve <python-script> --hyperparameter-argc

Copyright

Dendi Suhubdy