1st place solution
Main files:
make_features.py
- builds features from source datainput_pipe.py
- TF data preprocessing pipeline (assembles features into training/evaluation tensors, performs some sampling and normalisation)model.py
- the modeltrainer.py
- trains the model(s)hparams.py
- hyperpatameter sets.submission-final.ipynb
- generates predictions for submission
How to reproduce competition results:
- Download input files from https://www.kaggle.com/c/web-traffic-time-series-forecasting/data :
key_2.csv.zip
,train_2.csv.zip
, put them intodata
directory. - Run
python make_features.py data/vars --add_days=63
. It will extract data and features from the input files and put them intodata/vars
as Tensorflow checkpoint. - Run trainer:
python trainer.py --name s32 --hparam_set=s32 --n_models=3 --name s32 --no_eval --no_forward_split --asgd_decay=0.99 --max_steps=11500 --save_from_step=10500
. This command will simultaneously train 3 models on different seeds (on a single TF graph) and save 10 checkpoints from step 10500 to step 11500 todata/cpt
. Note: training requires GPU, because of cuDNN usage. CPU training will not work. If you have 3 or more GPUs, add--multi_gpu
flag to speed up the training. One can also try different hyperparameter sets (described inhparams.py
):--hparam_set=definc
,--hparam_set=inst81
, etc. Don't be afraid of displayed NaN losses during training. This is normal, because we do the training in a blind mode, without any evaluation of model performance. - Run
submission-final.ipynb
in a standard jupyter notebook environment, execute all cells. Prediction will take some time, because it have to load and evaluate 30 different model weights. At the end, you'll getsubmission.csv.gz
file indata
directory.
See also detailed model description