Transfer learning

Changed by Dimitrios Toumpanakis, 16 Apr 2021

Updates to Article Attributes

Body was changed:

The concept of transfer learning in artificial neural networks is taking knowledge acquired from training on one particular domain and applying it in learning a separate task.

In recent years, a well-established paradigm has been to pre-train models using large-scale data (e.g., ImageNet) and then to fine-tune the models on target tasks that often have less training data 3. For example, a neural network that has previously been trained to recognise pictures of animals may more effectively learn how to categorise pathology on a chest x-ray. In this example, the initial training of the network in animal image recognition is known as “pre-training”, while training on the subsequent data set of chest x-rays is known as “fine tuning”. This tool is most useful when the number of training examples in the pre-training data set is relatively large (e.g. 100,000 animal images) while the fine-tuning data set is relatively small (e.g. 200 chest x-rays).

The most popular dataset used for pre-training is the ImageNet dataset5, a very large dataset containing more than 14 million annotated images 4.

Intuition

The initial layers in a neural network for most image recognition tasks are involved in recognising simple features such as edges and curves. As such, a network which has been pre-trained on an unrelated image recognition task has already learned to see these lower level features. A network already pre-trained on images of animals does not need to re-learn such features, and is, therefore, able to train for the task of recognising chest x-ray pathology with fewer training examples.

  • -<![endif]--><!--StartFragment-->The concept of<strong> transfer learning</strong> in artificial <a href="/articles/neural-network-overview-1">neural networks</a> is taking knowledge acquired from training on one particular domain and applying it in learning a separate task.</p><p>For example, a neural network that has previously been trained to recognise pictures of animals may more effectively learn how to categorise pathology on a chest x-ray. In this example, the initial training of the network in animal image recognition is known as “pre-training”, while training on the subsequent data set of chest x-rays is known as “fine tuning”. This tool is most useful when the number of training examples in the pre-training data set is relatively large (e.g. 100,000 animal images) while the fine-tuning data set is relatively small (e.g. 200 chest x-rays).</p><p>The most popular dataset used for pre-training is the <a title="ImageNet dataset" href="/articles/imagenet-dataset">ImageNet dataset</a>.</p><h4>Intuition</h4><p><!--[if gte mso 9]><xml>
  • +<![endif]--><!--StartFragment-->The concept of<strong> transfer learning</strong> in artificial <a href="/articles/neural-network-overview-1">neural networks</a> is taking knowledge acquired from training on one particular domain and applying it in learning a separate task.</p><p>In recent years, a well-established paradigm has been to pre-train models using large-scale data (e.g., ImageNet) and then to fine-tune the models on target tasks that often have less training data <sup>3</sup>. For example, a neural network that has previously been trained to recognise pictures of animals may more effectively learn how to categorise pathology on a chest x-ray. In this example, the initial training of the network in animal image recognition is known as “pre-training”, while training on the subsequent data set of chest x-rays is known as “fine tuning”. This tool is most useful when the number of training examples in the pre-training data set is relatively large (e.g. 100,000 animal images) while the fine-tuning data set is relatively small (e.g. 200 chest x-rays).</p><p>The most popular dataset used for pre-training is the <a href="/articles/imagenet-dataset">ImageNet dataset</a> <sup>5</sup>, a very large dataset containing more than 14 million annotated images <sup>4</sup>.</p><h4>Intuition</h4><p><!--[if gte mso 9]><xml>

References changed:

  • 3. He K, Girshick R, Dollar P. Rethinking ImageNet Pre-Training. 2019 IEEE/CVF International Conference on Computer Vision (ICCV). 2019. <a href="https://doi.org/10.1109/iccv.2019.00502">doi:10.1109/iccv.2019.00502</a>
  • 4.O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 2015
  • 5.What makes ImageNet good for transfer learning? M Huh, P Agrawal, AA Efros - arXiv preprint arXiv:1608.08614, 2016 - arxiv.org

ADVERTISEMENT: Supporters see fewer/no ads

Updating… Please wait.

 Unable to process the form. Check for errors and try again.

 Thank you for updating your details.