SSU | [GD]

Optimization algorithm solving solving by moving through the parameter space steps given by the gradient.

Each step is given by where is called the learning rate or step size at iteration . The learning rate exist in order to prevent algorithm get stuck at local minimum.

Backpropagation is the standart way to compute gradient .

Properties