## Math Terms

A simple method for minimizing a function by changing its parameters in the direction the function decreases.

Gradient descent is one of the most used optimization methods. The goal is to find the parameters of a function `w` that produces the lowest value of the function. This is done by changing the parameters in small steps in the direction the function decreases until a (local) minimum is reached. Simple gradient descent implementations are very likely to get stuck in a local minimum - there are several techniques to mitigate this, like for example using momentum.

Gradient descent and its variants are often used in machine learning to minimize a loss function during training.

## Detailed formula explanation

The gradient descent formula can be written like this: The basis is this simple formula defines an iterative optimization method. We have some weights (parameters) `w` and we iteratively update them in some way to minimize the value of the function. Iterative methods are used when we cannot compute the solution directly. In machine learning, we define a loss function describing how good our model is. We want to find the weights that minimize the loss (make the model better). We compute the gradient of the function and update the weights by a small amount (learning rate) against the gradient. Here is an illustration of how it works. The gradient tells us if the function will decrease (negative gradient) or increase (positive gradient) if we increase the weight. The learning rate defines how far along the gradient we will jump in the current step of the optimization.

Large steps will lead us to the minimum more quickly, but we may jump over it and the optimization process may start to oscillate. Small steps may mean that we need many iterations to reach the minimum. Therefore, setting the learning rate correctly is very important for efficient optimization. With this, we get the final formula: Back to Math Terms