If my model is overfitting the training dataset, does adding noise to training dataset help regularizing the machine learning model

datascience.stackexchange https://datascience.stackexchange.com//questions/62928

Question

I would like to know if this is a best practice or not. Can we add noise to the training data to help the model "fit less the training data"; as a result, hoping to generalize better on new unseen data?

Was it helpful?

Solution

Yes, adding noise can help to regularize a model.

It is well known that the addition of noise to the input data of a neural network during training can, in some circumstances, lead to significant improvements in generalization performance

from Training with Noise is Equivalent to Tikhonov Regularization

In particular, adding structured noise that is consistent with natural perturbations of the data could help regularized a model. This is a form of data augmentation.

OTHER TIPS

No, adding noise will not help to regularise the model. This won't help your model generalise better to unseen inputs. It will probably just make your model perform generally worse. Any modifications you make to the training instances should be learnable.

Take an image classification problem as an example (images of a cat or not cat): randomly converting pixels to white in every image will not help your model learn generalisable features of cats; random white noise is not a 'learnable' feature. However by rotating, flipping, cropping, or adjusting contrast of the images, and adding these as new training instances, this would force the model to be more tolerant of differences in each image of cats, and therefore generalise better.

There are many methods for regularising models, such as l1 or l2 regularisation in linear regression or neural nets, and also plenty of other methods specific to other model types.

Making sure your training data is as varied as possible, and using model specific regularisation techniques, are probably closer to 'best practise' (though I don't really like the term since 'best practises' often change, such as new regularisation techniques, optimisations etc etc).

Licensed under: CC-BY-SA with attribution
Not affiliated with datascience.stackexchange
scroll top