Gradient Descent

  • Optimization algorithm
  • Trains Machine Learning Model
  • Minimizes errors between predicted value and actual value

Iterative process:

  • Trains over data over time
  • Users parameters with loss function
  • Until the function is close or equal to zero to get smallest possible error

Goal of gradient descent: MINIMIZE the cost function (error between predicted and actual value)

Requirements: a Learning Rate and the direction where to go.

Terms:

  1. Cost function: average error across whole dataset
  2. Loss Function: can be for a single sample of dataset

Types of Gradient Descents:

  1. Batch
  2. Stochastic
  3. Mini-Batch

https://www.ibm.com/think/topics/gradient-descent

Linear Regression: Gradient Descent (One of the best source)
https://developers.google.com/machine-learning/crash-course/linear-regression/gradient-descent
Math behind gradient descent:

  1. Initial model weights
  2. Calculate MSE loss with current model parameters
  3. Get weights values
  4. Move according to the learning rate for new weight values
  5. Perform for number of iterations

https://sebastianraschka.com/faq/docs/gradient-optimization.html

https://www.khanacademy.org/math/multivariable-calculus/applications-of-multivariable-derivatives/optimizing-multivariable-functions/a/what-is-gradient-descent

https://www.geeksforgeeks.org/gradient-descent-algorithm-and-its-variants

https://en.wikipedia.org/wiki/Gradient_descent

Leave a Reply

Your email address will not be published. Required fields are marked *

error: