Question

In a linear model, regularization decreases the slope. Do we just assume that fitting a lin model on training data overfits by almost always creating a slope which is higher than it would be with infinite observations instead? What is the intuition?

Was it helpful?

Solution

Regularization is used to help smooth multi-dimensional models. Take this example,

y = x_1 + eps*(x_2 + ... + x_100)

Let's say eps is very small. It doesnt seem very useful to store those 99 coefficients, isn't it? How do we manage to fit a model in such a way that we drop negligible coefficients? This is exactly what L1-regularisation does!

Each other type of regularisation has another geometric intuition.

Licensed under: CC-BY-SA with attribution
Not affiliated with datascience.stackexchange
scroll top