Mini-batch gradient descent
Citation, DOI, disclosures and article data
At the time the article was created David John Wang had no recorded disclosures.View David John Wang's current disclosures
At the time the article was last revised Andrew Murphy had no recorded disclosures.View Andrew Murphy's current disclosures
The mini-batch gradient descent is a technique that combines properties from batch gradient descent and also stochastic gradient descent to optimize efficiency and accuracy of the gradient descent algorithm. In each iteration, a certain number of examples (a batch) within a data set will undergo batch gradient descent. This batch could be as small as 2 or larger than 200 examples. Compared to batch gradient descent it is significantly faster, and compared with stochastic gradient descent good vectorisation of the number of examples allows the computation to parallelised, hence it can perform faster than a stochastic gradient descent as well.
- 1. NG A. "Machine Learning | Coursera". Coursera, 2018. [Link].