Stay organized with collections
Save and categorize content based on your preferences.
Preventing Overfitting
As with any machine learning model, a key concern when training a convolutional
neural network is overfitting: a model so tuned to the specifics of the
training data that it is unable to generalize to new examples. Two techniques to
prevent overfitting when building a CNN are:
Data augmentation: artificially boosting the diversity and number of
training examples by performing random transformations to existing images to
create a set of new variants (see Figure 7). Data augmentation is especially
useful when the original training data set is relatively small.
Dropout regularization: Randomly removing units from the neural network
during a training gradient step.
Figure 7. Data augmentation
on a single dog image (excerpted from the "Dogs vs. Cats" dataset
available on Kaggle). Left: Original dog image from training set.
Right: Nine new images generated from original image using random
transformations.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2022-07-18 UTC."],[[["Overfitting in convolutional neural networks can be mitigated by using techniques like data augmentation and dropout regularization."],["Data augmentation involves creating variations of existing training images to increase dataset diversity and size, which is particularly helpful for smaller datasets."],["Dropout regularization randomly removes units during training to prevent the model from becoming overly specialized to the training data."],["When dealing with large datasets, the need for dropout regularization diminishes and the impact of data augmentation is reduced."]]],[]]