Kaggle_Amazon_2017
Kaggle Amazon
Data
Image: 4 channels (red, green, blue, and near infrared)
Labels: There are 17 possible tags: agriculture, artisinal_mine, bare_ground, blooming, blow_down, clear, cloudy, conventional_mine, cultivation, habitation, haze, partly_cloudy, primary, road, selective_logging, slash_burn, water
Problems/Solution:
- Something wrong in the test file naming:
Solution: https://www.kaggle.com/robinkraft/fix-for-test-jpg-vs-tif-filenames
- Use Better Metric to see the performance
Suggested Solution: Use F2 Score directly as the metric.
-
GPU machines?
-
explore and decide trade off in optimizing GRAM usage
- larger batch size -> need more gram
- larger input image size -> need more gram
- larger model -> need more gram
- freeze some weights -> reduce gram usage?
-
freeze low level weights in pretained model?
-
use jpeg vs TIFF
- tiff has 4 channels: RGB + near IR, better?
- problem: most pretrained CNN model only have 3channels + tiff has large size
-
how to combine models?
- use XGBoost? example code: https://www.kaggle.com/opanichev/xgb-starter
Experiences:
Kin:
- used p8 (8 GPU) to run Keras model on ResNet50, but the runnung time didn't improve.
Jupyter notebook in AWS: https://blog.keras.io/running-jupyter-notebooks-on-gpu-on-aws-a-starter-guide.html
References:
-
Hierarchical Probabilistic Neural Network Language Model (Hierarchical Softmax) http://cpmarkchang.logdown.com/posts/276263--hierarchical-probabilistic-neural-networks-neural-network-language-model
-
'Working' Models https://www.kaggle.com/c/planet-understanding-the-amazon-from-space/discussion/33559
Graphs:
Meeting Summary:
20170702
Roy
- Freeze all layers but the last one fully connected layers
- Freeze some layers
- [the best] no-freezing and train with pre-train model
2016-6-17
Homework
- KL Divergence
- Reading: Distilling the knowledge in a neural network (by Hinton, Jeff Dean)
- Explore using GPU instance
- TIFF format and details (open the file and play around)
Individual work
Steve:
- Summary of Hinton's paper
Roy present:
- Task: Label image class, can have multiple label per image
- photos: 40000 train, 60000 test
- 256x256 jpg, 10k per image
- TIFF is 20x larger in size and TIFF has 4 channels R, G, B, NearIR
Problem: From 3 channel model to 4 channel
Roy's experience on Invasive Competition
- Run on InceptionV3, ResNet50, VGG.
combine 3 models output q1, q2, q3:
find output p then minimize sum of KL divergence:
KL(p, q1) + KL(p, q2) + KL(p, q3)
Result becomes better
Paper sharing
Distilling the knowledge in a neural network (by Hinton, Jeff Dean)
Ideas from paper
-
Train general CNN model
-
Make a covariance matrix of prediction categories
-
Make confusion matrix
-
inctrased concentration of relevant class in train samples, irrlevent class merged to dustbin class
-
minimize sum of KL divergence (from generel model + expert model)
To Do
- biased training set