1.

What Are The Benefits Of Mini-batch Gradient Descent?

Answer»
  1. Computationally efficient compared to stochastic GRADIENT descent.
  2. Improve generalization by finding flat minima.
  3. Improving convergence, by using mini-batches we approximating the gradient of the entire training SET, which MIGHT help to AVOID local minima.



Discussion

No Comment Found