/george-LRA-scheduler

George: Learning to Place Long-Lived Containers in Large Clusters with Operation Constraints

Primary LanguagePythonApache License 2.0Apache-2.0

George

George: Learning to Place Long-Lived Containers in Large Clusters with Operation Constraints

This repository contains the tensorflow implementation for reinforcement learning based Long Running Application scheduling.

DOI

Dependencies

  1. Python 3.5 or above
  2. Tensorflow 1.12.0
  3. scipy
  4. numpy
  5. pandas
  6. matplotlib
  7. sklearn
  8. ray
  9. Matlab 2019b or above

Installation

Install Requirements

git clone https://github.com/George-RL-based-container-sche/George.git

Clone from Github source:

pip install -r requirement.txt

Content

Our project includes three parts:

  1. Cluster: Implementation of our seven real-world LRAs that exhibit inter-container interferences.
  2. Experiments: George scheduling workflow based on our real-world LRA setting.

Experiment workflow

Real-World LRA cluster setup and profiler establishment

  1. Launch a cluster of tens of nodes on Amazon EC2 consisting of at least 1 manager node, several worker nodes, and some client nodes. Record the public DNS address or ip address of each node and make sure the manager and client nodes are accessible through the SSH key ~/.ssh/id_rsa.

    export MANAGER=ec2-xxx-xxx-xxx-100.us-west-2.compute.amazonaws.com
    export WORKER1=ec2-xxx-xxx-xxx-101.us-west-2.compute.amazonaws.com
    # ......
    export CLIENT1=ec2-xxx-xxx-xxx-201.us-west-2.compute.amazonaws.com
    # ......
    
    $ ssh ubuntu@$MANAGER
    (manager)$ git clone https://github.com/George-RL-based-container-sche/George.git
  2. If Docker has not been installed, install Docker on each manager, worker, and client node:

    (manager)$ sudo Cluster/scripts/install_docker.sh
    (worker1)$ sudo Cluster/scripts/install_docker.sh
    # ......
  3. Coordinating manager and worker nodes through Docker Swarm to form a Swarm cluster.

    $ ssh $MANAGER docker swarm init
    
    # Output
    To add a worker to this swarm, run the following command:
    
        docker swarm join --token SWMTKN-1-1bu27zw7lzh6pnu1l0981bu1m6nqq2pcgk25kovuh565319cah-8480smxu3kp7cj5nfkck2itax 192.168.99.100:2377

    Then execute the aforementioned command on each worker.

    $ ssh $WORKER1 docker swarm join --token SWMTKN-1-1bu27zw7lzh6pnu1l0981bu1m6nqq2pcgk25kovuh565319cah-8480smxu3kp7cj5nfkck2itax 192.168.99.100:2377
    
    # Output
    This node joined a swarm as a worker.
  4. Build docker images of different workloads on the manager node.

    (manager)$ cd Cluster/workloads
    (manager)$ ./build-all.sh
  5. Launch certain workloads on certain worker node from the manager node.

    (manager)$ cd Cluster/scripts
    
    # Launching 1 container for each workload on WORKER1:
    # '0' indicates idle; '1' ~ '7' indicate different workloads.
    
    (manager)$ ./service-launching.sh $WORKER1 0 1 2 3 4 5 6 7
    
    # e.g., launch 3 workload-1, 2 workload-2, and 1 workload-3 container:
    # ./service-launching.sh $WORKER2 0 0 1 1 1 2 2 3
  6. Sending requests from client node to the worker node

    $ ssh ubuntu@$CLIENT1
    (client1)$ git clone https://github.com/George-RL-based-container-sche/George.git
    
    # export WORKER1=ec2-xxx-xxx-xxx-101.us-west-2.compute.amazonaws.com    
    # Send single request for testing
    (client1)$ cd Cluster/scripts
    (client1)$ curl $WORKER1:8081
    
    # Or pressure the application with locust or other tools, e.g.,
    (client1)$ cd Cluster/scripts
    (client1)$ python3 parallel_locust.py $WORKER1
    # Default log path lies in Cluster/scripts/log
  7. (Optional) Collecting performance benchmark datasets through automatically deploying, profiling, terminating, and re-deploying.

    # Here the "0-1-2-3-4-5-6-7" or "0-0-1-1-1-2-2-3" means combination
    # of different workloads (as described step 5).
    (manager)$ ./profiling-go.sh $WORKER1 "0-1-2-3-4-5-6-7 0-0-1-1-1-2-2-3 0-0-0-0-0-0-0-7 0-0-0-0-1-1-1-7 2-4-4-6-6-6-6-7"
  8. Terminate the swarm cluster on the manager node.

    (manager)$ docker swarm leave --force

Operation Constraints

Operation constraints are set by the operators in real-world LRA to guarantee the safety and stability of the cluster. As stated in our manuscript, we consider two kinds of constraints in our cluster, namely soft constraints and hard constraints.

Soft constraints

Soft constraints should be guaranteed within a pre-defined violation threshold (e.g., 5%). We consider the following soft constraints in our setting.

  • Incremental deployment: The containers of the ISR application should be placed on a specific machine subset for the new-version deployment. The machine subset consists of 50% of the machine in the cluster.

  • Hardware affinity: To mimic the hardware requirement, we assume one LRA, CKM requires a particular kernel version and can only be deployed on a machine subset with 30% of the machines.

Hard constraints

Hard constraints should be strictly enforced in the cluster and cannot be violated.

  • Node Capacity: The number of containers placed on each hosting machine should not exceed the node capacity, which is 8 in our experiment.

  • Deployment Spreading: For each LRA, it can launch at most one container on a physical machine for load-balance.

George: RL model training for the real-world LRA cluster

  1. First check the data collected by Real-World LRA cluster is stored in the folder:

    $ cd Experiments/
    $ ls ./simulator/datasets/
    
    ***_sample_collected.npz 
    
  2. Train sub-schedulers in a 27-node sub-cluster:

    $ cd Experiments/
    $ ./shell/TrainSubScheduler.sh
    

    Output: the well-trained sub-scheduler models, as well as corresponding log files will be store in the folder:

    $ ls ./checkpoint/
    
    subScheduler_*/  
    
    
  3. High-level model training based on previously well-trained sub-schedulers.

    (0) Check the sub-scheduler models are stored in the folder:

    $ cd Experiments/
    $ ls ./checkpoint/
    
    subScheduler_*/   
    

    Check the container batches data is stored in the folder or create your own batches:

    $ ls ./data
    
    batch_set_200.npz batch_set_300.npz batch_set_400.npz
    batch_set_1000.npz batch_set_2000.npz batch_set_3000.npz
    

    (1) High-level training in a medium-sized cluster of 81 nodes:

    $ ./shell/RunHighLevelTrainingMedium.sh 200
    $ ./shell/RunHighLevelTrainingMedium.sh 300
    $ ./shell/RunHighLevelTrainingMedium.sh 400
    

    Output: the training log files including the RPS, placement matrix, training time duration .etc will be store in the folder:

    $ ls ./checkpoint/
    
    81nodes_*_*/
    

    (2) High-level training in a large cluster of 729 nodes:

    $ ./shell/RunHighLevelTrainingLarge.sh 1000
    $ ./shell/RunHighLevelTrainingLarge.sh 2000
    $ ./shell/RunHighLevelTrainingLarge.sh 3000
    

    Output: the training log files including the RPS, placement matrix, training time duration .etc will be store in the folder:

    $ ls ./checkpoint/
    
    729nodes_*_*/
    

Baseline Method: Vanilla RL

Vanilla RL is built directly upon Policy Gradient without our Hierarchical designs.

  1. High-level training in a medium-sized cluster of 81 nodes:

    $ ./shell/RunVanillaRLMedium.sh 200
    

    Output: the training log files including the RPS, placement matrix, training time duration, etc will be store in the folder:

    $ ls ./checkpoint/
    
    Vanilla_81_*_*/
    

Baseline Method: Divide-Conquer (DC)

DC Method does not use sub-schedulers. Our code below shows its behaviors in a medium-sized cluster of 81 nodes. Each cluster is hierarchically divided into three subsets.

  1. High-level training in a medium-sized cluster of 81 nodes:

    $ ./shell/RunDCMedium.sh 200
    

    Output: the training log files including the RPS, placement matrix, training time duration, etc will be store in the folder:

    $ ls ./checkpoint/
    
    DC_81_*_*/
    

Baseline Method: Medea

Medea is implemented using Matlab, due to its outstanding performance in solving the Integer Linear Programming (ILP) Problem.

  1. Generate the performance-constraints used in Medea:

    $ cd Experiments
    $ ./shell/GenerateInterference.sh
    $ ls
    
    interference_applist.csv interference_rpslist.csv
    
  2. Run Medea in the folder:

    $ cd testbed/Medea
    $ Matlab Medea.m
    

    Output: the scheduling decision log files including the allocation matrix, constraint violations, time duration .etc will be store in the folder.

Baseline Method: Paragon

Paragon is re-implemented in Python. For the sake of fair comparison, we feed it with the full interference matrix information as Medea.

  1. Make sure interference_applist.csv has been generated in former Medea setup:

    $ ls interference_applist.csv
    

    Otherwise, generate the performance-constraints used in Medea:

    $ cd Experiments
    $ ./shell/GenerateInterference.sh
    $ ls interference_applist.csv
    
  2. Run Paragon of Medium size or Large size:

    $ cd Experiments/shell
    $ # Medium size
    $ ./RunParagonMedium.sh 200
    $
    $ # Large size
    $ ./RunParagonLarge.sh 2000
    

    Output: the default output shows the average throughput for each testing group as well as the scheduling latency.

    For detailed output including container placement and per-container throughput breakdown for each node, please add -v after each python script:

    $ cd Experiments
    $ python3 ParagonExp.py --batch_set_size 200 --batch_choice 0 --size medium --verbose
    

Baseline Method: FPO

FPO is implemented based on Metis (GitHub Repo) and the Fixed Penalty Optimization proposed in the CPO paper.

  1. Run FPO of Small size:

    $ cd Experiments/shell
    $ # Small size
    $ ./shell/RunFPO.sh
    

    Output: the training log files including the RPS, constraint violations, placement matrix, training time duration, etc. will be store in the folder:

    $ cd Experiments
    $  ls ./checkpoint/
    FPO_27_*_*/
    

Baseline Method: CPO

CPO is implemented based on thier GitHub Repo. In order to achieve fast training with Ray, we package the CPO source code in the testbedlib module.

  1. Install the testbedlib module with the following command.

    $ cd Experiments
    $ pip install ./dist/testbedlib-0.1.0-py3-none-any.whl
    

    It is much more convenient to install it in your Python virtual environment or conda environment.

  2. Run CPO of Small size:

    $ cd Experiments/shell
    $ # Small size
    $ ./shell/RunCPO.sh
    

    Output: the training log files including the RPS, constraint violations, placement matrix, training time duration, etc. will be store in the folder:

    $ cd Experiments
    $  ls ./checkpoint/
    CPO_27_*_*/
    

Our Method: PPPO

PPPO is implemented in Python with TensorFlow on our own. In order to achieve fast training with Ray, we package the PPPO source code in the testbedlib module.

  1. Install the testbedlib module with the following command.

    $ cd Experiments
    $ pip install ./dist/testbedlib-0.1.0-py3-none-any.whl
    

    It is much more convenient to install it in your Python virtual environment or conda environment.

  2. Run PPPO of Small size:

    $ cd Experiments/shell
    $ # Small size
    $ ./shell/RunPPPO.sh
    

    Output: the training log files including the RPS, constraint violations, placement matrix, training time duration, etc. will be store in the folder:

    $ cd Experiments
    $  ls ./checkpoint/
    PPPO_27_*_*/
    

References

[1] Medea: Panagiotis Garefalakis, Konstantinos Karanasos, Peter Pietzuch, Arun Suresh, and Sriram Rao. 2018. Medea: scheduling of long running applications in shared production clusters. In Proceedings of the Thirteenth EuroSys Conference (EuroSys ’18). Association for Computing Machinery, New York, NY, USA, Article 4, 1–13. DOI:https://doi.org/10.1145/3190508.3190549

[2] Paragon: Christina Delimitrou and Christos Kozyrakis. 2013. Paragon: QoS-aware scheduling for heterogeneous datacenters. In Proceedings of the eighteenth international conference on Architectural support for programming languages and operating systems (ASPLOS ’13). Association for Computing Machinery, New York, NY, USA, 77–88. DOI:https://doi.org/10.1145/2451116.2451125

[3] CPO: Joshua Achiam, David Held, Aviv Tamar, and Pieter Abbeel. 2017. Constrained Policy Optimization. In Proceedings of the 34th International Conference on Machine Learning (ICML), 2017.

[4] Redis: an open source, in-memory data structure store. https://redis.io

[5] Model Server for Apache MXNet. https://github.com/awslabs/mxnet-model-server

[6] Image Super Resolution. https://github.com/idealo/image-super-resolution

[7] Locust: an open source load testing tool. https://locust.io

[8] Yahoo! Cloud Streaming Benchmark: Brian F. Cooper, Adam Silberstein, Erwin Tam, Raghu Ramakrishnan, and Russell Sears. 2010. Benchmarking cloud serving systems with YCSB. In Proceedings of the 1st ACM symposium on Cloud computing (SoCC ’10). Association for Computing Machinery, New York, NY, USA, 143–154. DOI:https://doi.org/10.1145/1807128.1807152

[9] Ray: Fast and Simple Distributed Computing. https://ray.io/