Mathematical functions (like ReLU, Sigmoid, Softmax) that introduce non-linearity into the network, allowing it to learn complex patterns.
Essential for deep learning to work.
ReLU is the default for hidden layers; Softmax for output.