Compare Activation Function The log of training process. Two(Accuracy and Cross entropy loss) from the left. ReLU, Sigmoid, ReLU with He initialization. The loss graph of ReLU and Sigmoid. The loss graph of ReLU with He initialization and without.