hongzimao/decima-sim

A trained model

hongzimao opened this issue · 9 comments

Recently received many requests for our trained model. Here's one after 20,000 steps using the code in the master branch. Hope this helps saving some training time when reproducing our results. https://www.dropbox.com/sh/62niiuaa7103cth/AABxu3ekjOYakmr86gMECZ3Ca?dl=0

This link seems expired, would you please have a check?

The link is indeed expired. I lost the access to MIT dropbox to retrieve the model since graduation. I searched my local storage but unfortunately I couldn't find the exact trained model. However,I believe there were others downloaded this model before. Can someone upload the model here? Thanks a lot!

Here are some models I was able to retrieve from our local machine. Although the creation time of the model is a few months prior to this post, It should be in a similar setting. I'm attaching a few model snapshot at after 20,000 iterations. Check if the performance is good: models.zip

Still, if someone has the original model in this post, please do contact us with a copy. We will upload it here. Thanks!

Thanks for your reply.

I spend some time testing the performance (RL models with 20000 + epochs are given by you):

Scheduler Avg. JCT Executor Usage
FIFO 1842803 0.8197
Dynamic Partition 62783 0.7074
RL 20000 61190 0.6456
RL 24000 64663 0.6469
RL 25000 64044 0.6399
RL 26000 63494 0.6462
RL 10000 (Trained by us) 57241 0.6634

All settings are as default. Does this performance meet expectations?

Looks like your trained model performs better :) would you mind sharing the model so that others may use it too? Thank you!

Here are our trained_models. We trained 23000 epochs for 10 days on a server with 56 cores CPU and 4 V100 GPUs. (For training reference. All settings are as default.)

We also perform more evaluations on our trained models: (For testing reference. Each evaluation takes about 30 minutes.)

Scheduler Avg. JCT Executor Usage
RL 100 99540 0.6392
RL 500 89455 0.6397
RL 1000 62610 0.6456
RL 2000 58918 0.6493
RL 6000 56739 0.6665
RL 10000 57241 0.6634
RL 10900 105744 0.7236
RL 11000 66811 0.6396
RL 12000 62434 0.6886
RL 15000 73518 0.6029
RL 16000 61234 0.6694
RL 18000 61145 0.6629
RL 20000 65519 0.6518
RL 22000 61105 0.6689

It seems not stable for RL training and the performances are not much better than Dynamic Partition. I want to know how to choose the best model checkpoint without testing (which metric in the tensorboard is the most significant one)? Do you have any insight? Thank you.

@Tonyhao96 Would you like to give me some instructions on which command syntax did you used to train your model and how did you compare the performance?

@jahidhasanlinix I just use the command provided by the authors without modification. I trained Decima several months ago and I totally forget the details.

@Tonyhao96 Thank you for your response.