/Compare_Activation_Function

Compare vanishing gradient problem case by case.

Primary LanguagePython

Compare Activation Function


The log of training process.
Two(Accuracy and Cross entropy loss) from the left.
ReLU, Sigmoid, ReLU with He initialization.


The loss graph of ReLU and Sigmoid.


The loss graph of ReLU with He initialization and without.