Convergence rates

  1. Stochastic Gradient Descent has a faster convergence rate due to the use of single training examples in each iteration.
  2. Gradient Descent as a slower convergence rate, as it uses the entire dataset for each iteration.