Batch size (machine learning)
Citation, DOI, disclosures and article data
Citation:
Gaillard F, Murphy A, Batch size (machine learning). Reference article, Radiopaedia.org (Accessed on 14 Dec 2024) https://doi.org/10.53347/rID-56140
Permalink:
rID:
56140
Article created:
Disclosures:
At the time the article was created Frank Gaillard had no recorded disclosures.
View Frank Gaillard's current disclosures
Last revised:
Disclosures:
At the time the article was last revised Andrew Murphy had no recorded disclosures.
View Andrew Murphy's current disclosures
Revisions:
2 times, by
2 contributors -
see full revision history and disclosures
Sections:
Tags:
Batch size is a term used in machine learning and refers to the number of training examples utilised in one iteration. The batch size can be one of three options:
- batch mode: where the batch size is equal to the total dataset thus making the iteration and epoch values equivalent
- mini-batch mode: where the batch size is greater than one but less than the total dataset size. Usually, a number that can be divided into the total dataset size.
- stochastic mode: where the batch size is equal to one. Therefore the gradient and the neural network parameters are updated after each sample.
Quiz questions
{"containerId":"expandableQuestionsContainer","displayRelatedArticles":true,"displayNextQuestion":true,"displaySkipQuestion":true,"articleId":56140,"questionManager":null,"mcqUrl":"https://radiopaedia.org/articles/batch-size-machine-learning/questions/1650?lang=gb"}
Incoming Links
Multiple choice questions:
Related articles: Artificial intelligence
- artificial intelligence (AI)
- imaging data sets
- computer-aided diagnosis (CAD)
- natural language processing
- machine learning (overview)
- visualising and understanding neural networks
- common data preparation/preprocessing steps
- DICOM to bitmap conversion
- dimensionality reduction
- scaling
- centring
- normalisation
- principal component analysis
- training, testing and validation datasets
- augmentation
- loss function
-
optimisation algorithms
- ADAM
- momentum (Nesterov)
- stochastic gradient descent
- mini-batch gradient descent
-
regularisation
- linear and quadratic
- batch normalisation
- ensembling
- rule-based expert systems
- glossary
- activation function
- anomaly detection
- automation bias
- backpropagation
- batch size
- computer vision
- concept drift
- cost function
- confusion matrix
- convolution
- cross validation
- curse of dimensionality
- dice similarity coefficient
- dimensionality reduction
- epoch
- explainable artificial intelligence/XAI
- feature extraction
- federated learning
- gradient descent
- ground truth
- hyperparameters
- image dataset normalisation
- image registration
- imputation
- iteration
- jaccard index
- linear algebra
- noise reduction
- normalisation
- R (Programming language)
- radiomics quality score (RQS)
- Python (Programming language)
- segmentation
- semi-supervised learning
- synthetic and augmented data
- overfitting
- underfitting
- transfer learning