/carnd_traffic_sign_classifier

German traffic signs classifier in TensorFlow

Primary LanguageJupyter Notebook

carnd_traffic_sign_classifier

This is a German traffic signs classifier written in TensorFlow, which I created as part of Udacity's Self-Driving Car Engineer Nanodegree (carnd).

##The phases of this project are the following:

  • Loading the dataset
  • Dataset summary, exploration and visualization
  • Design, train and test a model architecture
  • Use the model to make predictions on new images
  • Analyze the softmax probabilities of the new images

Loading the dataset

First thing's first: download the dataset

$ wget https://d17h27t6h515a5.cloudfront.net/topher/2017/February/5898cd6f_traffic-signs-data/traffic-signs-data.zip

The zip file contains 3 pickle files with training, validation and test images, all resized to 32x32 pixels.
The zip file also contains a CSV file (signnames.csv) with the first column containing the class ID (an integer spanning 0-42), and the second column containing a descriptive name of the sign
Here are the first 4 rows:

ClassId SignName
0 Speed limit (20km/h)
1 Speed limit (30km/h)
2 Speed limit (50km/h)
3 Speed limit (60km/h)

A quick examination of the data sets yields these stats:

Number of training examples = 34799
Number of validation examples = 4410
Number of testing examples = 12630
Image data shape = (32, 32, 3)
Number of classes = 43

Each class represents a differnet traffic sign.
I think it is odd that the test set is larger than the validation set and that it is so large: it is 1/3 the size of the training set and 3 times the size of the validation set. The validation set is in fact only about 12.7% of the training set, while 20% is recommended. When I used part of the training set for validation (instead of using the supplied validation set), I received very good validation results. However, I chose to use the suuplied validation set, as it seemed more appropriate.

X_train, X_valid, y_train, y_valid = train_test_split(train['features'], train['labels'], test_size=0.2, random_state=0)

Dataset summary, exploration and visualization

I plotted the distribution of the training samples between the classes It is evident that the classes do not have equal representation in the training dataset. Because there are 43 classes, had the samples been distributed equally, we would have 2.33% (100%/43) of the samples in each class. However, in the actual dataset distribution, most classes comprise less than 2%. In fact, although the mean representation is 2.33%, the median is 1.55%, which means that half the classes have a representation below 1.55%
Class 2 has the largest number of samples (5.78%), and classes 0 and 19 have the lowest representation (0.52% each). I also plotted a histogram of the percent-of-samples per class. This graph provides another look at how the samples distribute between the classes and it also shows that most classes have a very low representation.
The validation dataset also doesn't distribute the samples between the classes in an equal manner.
It is interesting to look at how well the validation dataset represents the training set, and that's the reason I plotted the distributions of the training and validation set in the same plot. Looking at the plot we see q large resemblence in the distributions of the two sets, but I wanted to make sure this is the case. I divided the percent representation of each class in the training set, by its representation in the validation set and called this the "ratio" A ratio close to 1 indicates that there about the same fraction of validation samples as training samples, in the specific class. In the iPython notebook I print the "ratio" value for each class, but here it sufficient to note that the median=1.03 and mean=0.94. Although the two datasets distribute similarly enough, the plot below shows that there are enough outliers to raise my curiousity if the classes that have low/high ratio will show poor validation accuracy. We shall see.

Next, let's have a quick look at an image from each class, just to start getting familiar with the images themselves.

We see here several very dark images, some blurry images. Images are mostly centered, and mostly occupy the same amount of area within each image. Images are mostly upright, but there are a few with a bit of rotation. Backgrounds vary, but are mostly yellowish-brownish.

Now I want to dig deeper into one of the classes. It is important to spend time reviewing the finer details of the images and watch for different qualities and characteristics they may have. This will help in deciding what augmentations may help during the training of the model.
Class 0 is small and might need augmenting, so let's look at it.
This is a bit surprising: I expected more variance, but it looks like there's a small set of signs that appear multiple times with some differences. Not exactly what I was expecting. Next we look at class 0 images from the validation dataset.

Another surprise! These images look almost all the same - it's as if they were taken from the window of a slow moving car.
I don't have much experience, but this kind of repetitivity seems problematic since its mostly testing for a small number of features, because these images are highly corrolated. On the other hand, I suppose that in a real driving scenario, this is exaclty what we would expect to see: the same sign with some small variance due to the car movement.
I also examined class 19, since it is also small enough to display (180 training samples). It exhibited similar chracteristics, but the training set has almost 30 very dark images (~15% of the images).

Even when I look at these closer (larger) they look extremely dark.
I converted them to grayscale, and look at the improvement! Grayscale is just a dot product of [R,G,B] and [0.299, 0.587, 0.114], so I imagine this "feature" can be learned by the network. I.e. the network can learn to exract the luminance feature if it determines it is important enough. If in the pre-processing phase I will convert the dataset images to grayscale, I will bascially make the decision that the chroma channels are not important and that the luminance data is sufficient to represent the inputs. These dark images give me confidence that there's a lot of information in the luminance data.

###Design, train and test a model architecture The code for preprocessing the image datasets is in the iPython cell which has "Preprocessing utilities" as the first remark.
I first convert the RGB images to grayscale. There's little color information in the images, and I think that the structure of the signs is enough. Yann LeCun reported in his paper that this helped increase the predcition accuracy by a small bit, while decreasing the complexity of the model (we only have to deal with one third of the input features).
Next, I perform min-max scaling (also known as scale normalization) to rescale the pixels to the 0-1 floating-point range. It is common practice to do this in cases where the input features have different scales, but in 8-bit integer grayscale images the pixels are already within the same value range (0-255), so it would seem unnecessary to do this rescaling. Finally, I standardized the data using Z-score normalization which centers (mean=0) and normalizes (std=1) all of the images.
As a sanity check, I display the same 4 dark left-turn signs from the pre-processed image set and display them:

Before applying the image preprocessing, I need to augment the dataset. I'm augmenting because I want to make the classes have a bit better representation distribution. I'm augmenting a class by taking each image from that class in the dataset and add augmented copies of it. The augmentation is performed with scipy's ndimage library. I only augment the classes that have "low" representation (1% or lower). For classes with representation below 0.7% I apply extra augmentation (three times as much augmented images). I could have chosen a more elaborate augmentation policy, and there's more experimenting that I can do, but will skip due to lack of time. After the augmentation the distribution is much better:

I tried two different augmentations: random rotation of angles chosen in the range {-15, 15}, and random bluring levels. I choose between the two randomly. Later on, feeling that I needed more augmentation, I added random shifting (on x and y axes) and mixed rotation with shifting. I didn't see much improvement, but I didn't do enough experimnetation. Finally, I apply the pre-processing on the newly augmented training dataset, and then plot 5 random images, for some sanity checks:


I tried four different network models which I named netanet1 thru netanet4. All are very similar and are some variant of the LeNet architecture:

After trials and tribulations I settled on netanet4. As I describe below, I ran training & validation runs and plotted the results against one another, but I don't show all of these runs. Instead, the data below refers to netanet4 only. I use three convolution layers which I made quite deep - all in an attempt to extract more features. The last convolution layer uses a smaller 3x3 kernel as the IFMs were getting pretty small due to the two max-pool layers. I only used 2 FC layers hoping that this would be enough to capture the interaction between the feature-maps. A 50% dropout layer should prevent overfitting.
Hyper-parameters which I considered included the learning-rate, the batch size, the number of epochs, and the network topology itself. Along the way I changed other "hyper-parameters" such as the types of the augmentations I need to apply, the classes I applied the augmentations, and the types of pre-processing. That's quite a lot of tuning and I feel that I've got a long way to go before I understand the interactions between these hyper-parameters. One of the first things I tried, was adding the pre-processing step which added approximately 2% of accuracy improvement. Adding normalization to the preprocessing did wonders to the Loss behavior. Instead of the Loss jerking around, it became much smoother. The graphs below shows this:
Loss before normalization:

Loss after normalization:

I also slowly increased the number of epochs, since I understood that I needed to give the training process enough time to reach the minimum, or get out of a local minimum and continue searching for the global loss minimum.
Eventually, I had to finish this task and settle on some values for these hyper-parameters. However, it became obvious along the way, that I need create some kind of light framework to capture the training process and results. At minimum, I needed to collect and archive the results of each run, so that I'll be able to load them to a single graph and do some comparisons.
I wrote three classes to aid in the training. One class encapsulates the training configuration (TrainingConfig), another collects results (TrainingResults), and one performs the actual training (CnnTrainer). I supplement this with some code to archive all of the training artificats. I do this so that I can graph all of the results and compare them to one another. I placed this code in a separate IPython notebook because things were getting messy for me (results_comparison.ipynb).

Comparing the models

So what am I comparing? Obviously, the goal is to maximize the validation accuracy.
I also graphed the errors distribution between the classes, and something I call (maybe I heard it somewhere) the class confusion graph. The confusion graph of a class, show what classes are predicted for a specific class.
Anrdej Karpathy's Stanford CS231n notes provide some information on how to interpret the loss function behavior. The shape of the Loss graph gives clues into tuning the learning rate and batch size (the batch size might be a little too low if the loss is too noisy). Following Karpathy's advice I plot the Loss on both a linear and log scale and I found them both interesting. The linear scale makes the "hockey stick" shape more pronounced which help determine if this shape is "healthy" (not too steep and not too gradual); while the log scale "magnifies" the details of the Loss function behavior as the Loss decreases, making comparisons between inference runs easier.
I wanted to use a decaying learning rate because it looked like after a certain amount of epochs, the Loss kind of bounces around in some small range. I thought this might be due to the size of the learning rate. From the TensorFlow documentation it appeared that tf.contrib.layers.xavier_initializer is a good candidate for the decay function, but from reading around I understood that tf.train.AdamOptimizer already takes care of learning rate decay. According to Karpathy: "In practice Adam is currently recommended as the default algorithm to use" so I didn't want to change that. I wanted to try out different values for 'epsilon' as recommended ("The default value of 1e-8 for epsilon might not be a good default in general"), but I didn't have enough time to do that.


Karpathy also provides insights into interpreting the validation vs. training accuracy graph, which is used to look for overfitting.
The graphs below display some of the statistics I collected, while training the networks and comparing them.
![](loss_accuracy_summary.png)
This graph shows the error distribution of one of the runs. It shows for example, that class 24 has the highest error rate.
![](errors_per_class.png)
Let's take a look at class 24's confusion graph and see what class is giving it problems:
![](class_24_errors.png) ### Use the model to make predictions on new images I collected 6 German traffic signs by cropping an [image](http://electronicimaging.spiedigitallibrary.org/data/journals/electim/927109/jei_22_4_041105_f010.png) I found on the Web and ran predictions and top-5 softmax results for them.
The images I chose are represented in the training set. They are bright, clear and centered. Class 0 (speed limit 20) is cropped at the edges, and that may throw off the prediction because I don't have such images in the training dataset. When I plot the images in the training set next to the images I downloaded, they look similar enough.
Class 0 has a a very high error rate in the validation process (23.3%) so it has a high chance of being misclassified. It has a high representation in the training dataset, but most of the images are augmentations, so they might have a high correlation - especially if the augmentations didn't change them enough. Class 33 has the next highest error rate, and a low representation in the training set, so it might also be misclassified.
I plotted a few of them, against their training images:
![](web_vs_training.png)
Image Verification error rep in training
17 1.7% 2.05%
33 5.6% 1.24%
0 23.3% 2.23%
28 0.0% 0.99%
14 3.3% 1.43%
10 0.0% 3.72%

Image Prediction 1 2 3 4 5
17 Correct 17 (100%) 14 41 9 20
33 Correct 33 (99.45%) 42 (0.55%) 10 7 13
0 Incorrect 31 (85.82%) 37 (14.11%) 14 25 38
28 Incorrect 11 (68.67%) 30 (31.0%) 34 42 23
14 Correct 14 (99.1%) 17 (0.8%) 33 1 38
10 Correct 10 (100%) 42 41 0 1

The model is very certain of its predictions for classes 17, 33, 14, and 10 - and indeed these are predicted correctly!
For the image from class 0, the model is leaning towards class 31 (85.82%) and 37 (14.11%). This is strange: class 31 is "Wild animals crossing", and has a red triangle frame; and class 37 is "Go straight or left", and has a blue circle frame; while class 0 is "Speed limit (20km/h)", and has a red circle frame. The network has failed to learn class 0 very well and this is also noticeable in the verifitcation error rate for this class (23%).
For class 28 ("Children crossing"), the model predicts class 11 ("Right-of-way at the next intersection", 68.67%) and class 30 ("Beware of ice/snow", 31.0%). This confusion makes a lot of sense: all three classes have a dominant red traingle with a black "blob" in the middle. In fact, it is this black blob which is what distinguishes the classes from one another - but it is small and blurry, making the prediction much harder. The network performs max-pooling right after the first convolution, and this subsampling process even further "compresses" the receptive field.
Over all, the model predicted correctly 4/6 of the images, or 67%. Well below the expected 95-96% correct prediction rate. So am I overfitting? I don't think we can tell, by looking at just 6 test images. I also used the test dataset I described at the top to make predictions:
Test dataset: 923 errors of 12630 samples (7.31%)
Looking at the results of testing a much larger test dataset, the error rate goes down to 7.3%. This is a lot closer to the error rate I've got on the verification dataset - the model might be overfitting by a small amount. More images might help generalize.