Activation functions are mathematical functions applied to a neural network node's input to introduce non-linearity, enabling the network to learn complex patterns. They transform the weighted sum of inputs into an output, which is then passed to subsequent layers. Common examples include ReLU, sigmoid, and tanh, each influencing the network's ability to model various data relationships.