When you implement back propagation for your NN, you need to compute the slop/derivative of the activation function. Let's take a look at this.
- Sigmoid activation function
We have:
Following is the derivation:
- Tanh activation function
We have:
Here's the derivation:
- ReLU activation function
We have:
- Leaky ReLU activation function
<end>