Search Results

Rectified Linear Unit (ReLU)

A Rectified Linear Unit is a common name for a neuron (the “unit”) with an activation function of \(f(x) = \max(0,x)\).

Neural networks built with ReLU have the following advantages:

  • gradient computation is simpler because the activation function is computationally similar than comparable activation functions like \(\tanh(x)\).
  • Neural networks with ReLU are less susceptible to the vanishing gradient problem but may suffer from the dying ReLU problem.