Gradient Descent is a fundamental optimization algorithm in deep learning that adjusts model parameters to minimize the cost function. It works by calculating the gradient (slope) of the loss with respect to parameters and updating them in the opposite direction of the gradient, iteratively reducing error and improving model accuracy. Variants include batch, stochastic, and mini-batch gradient descent.