Deep Learning Tutorial 3
Deep Learning Tutorial 3
Scratch
Theory + Practical
FAHAD HUSSAIN
MCS, MSCS, DAE(CIT)
∑ 𝒘𝒊 𝒙𝒊
𝒊= 𝒙
Applying Activation
Function
The activation function is the non linear transformation that we do over the input
signal. This transformed output is then seen to the next layer of neurons as input.
What we have done here is that we have simply replaced the horizontal line with a non-zero, non-horizontal line. Here
a is a small value like 0.01 or so.
The main advantage of using Softmax is the output probabilities range. The range will 0 to 1, and the sum of all
the probabilities will be equal to one. If the softmax function used for multi-classification model it returns the
probabilities of each class and the target class will have the high probability.
The formula computes the exponential (e-power) of the given input value and the sum of exponential values of
all the values in the inputs. Then the ratio of the exponential of the input value and the sum of exponential
values is the output of the softmax function.