top of page

Regularization: The magic sauce for a fit model

1 min read

2

61

1

Need for regularization


Machine learning models tend to overfit, making accurate predictions on trained data but inaccurate ones on new examples. Regularization helps prevent overfitting by shrinking the parameters through a regularized term added to the cost function.


Check out my comic that explains regularization in a fun and simple way!



Regularization calculation


Let's take an example of a linear regression model.

From my cost function blog, we can define the minimization of the linear regression cost function as:

Where,

J(θ) - cost function of linear regression

M - number of training examples

hθ​(xi) - predicted output for the i-th input x, given the parameters θ

θ - parameters like model weights

yi - actual output for the input xi

As we can see, in the overfitting graph there are too many parameters.


We need to multiply the θ3 and θ4 values with a number to make it really small.

Therefore, θ3 and θ4 become negligible as the min function is present.


Now, the generalized equation for linear regression regularization can be written as:

Where,

λ - regularization parameter (penalizing value)

n - number of parameters


Similar to the linear regression example we've taken, any function can be regularized by adding a shrinking term.


Hope you've understood the concept of regularization! Let me know what you think :)

1 min read

2

61

1

Related Posts

Comments (1)

Unknown member
Oct 04, 2024

Awesome!

Like

Follow Me

  • LinkedIn
  • Instagram
github-mark-white.png
bottom of page