Meri Leeworthy

activation function

Perceptrons traditionally used the (binary) Heaviside step function. The binary threshold was meant to model the way biological neurons ‘fire’. Alternatively, sigmoid can be used.

In the context of perceptrons and neural networks, using different activation functions like the step function and the sigmoid function serves different purposes. Here’s a detailed explanation:

Step Function:

Sigmoid Function:

Why Use the Sigmoid Function?

  1. Gradient-Based Learning:

    • Neural networks are typically trained using gradient-based optimization algorithms like backpropagation.
    • The sigmoid function is differentiable, allowing the calculation of gradients needed for these algorithms.
  2. Probabilistic Interpretation:

    • The sigmoid function maps any real-valued number into the range (0, 1), which can be interpreted as a probability.
    • This is useful in applications like binary classification, where the output can be interpreted as the probability of belonging to a certain class.
  3. Smooth Transitions:

    • Unlike the step function, the sigmoid function allows for smooth transitions between output values, which can be beneficial in learning complex patterns in the data.

Practical Considerations

Summary

I live and work on the land of the Wurundjeri people of the Kulin Nation. I pay respect to their elders past and present and acknowledge that sovereignty was never ceded. Always was, always will be Aboriginal land.

This site uses open source typefaces, including Sligoil by Ariel Martín Pérez, and Vercetti by Filippos Fragkogiannis