Ajai Chemmanam
Activation Functions
Date: 2022-04-12
Written by Ajai Chemmanam

Activation Functions

In an artificial neural network, activation functions decide the output of a node for a given input. It decides whether a node should be activated or not.

Importance & Properties Of Activation Functions

  • It adds a non-linearity to Neural Networks

  • It puts upper and lower bounds in the output of each node

  • It should be Continuous, Monotonic & Differentiable.

Some commonly used Activation Functions are listed below

Sigmoid Function

It is a continuous 'S' curve ranging from 0 to 1.

Equation: Z = 1/(1+e^-x)

Tanh Function

It has similar structure of sigmoid function ranging from -1 to 1.

Equation: tanh(x) =2/(1+e^-2x) -1

Relu Function

The most widely used activation function ranges from 0 - inf.It returns 0 if the input is less than or equal to 0.

Equlation: Relu(x) = max(0,x)