-
Book Overview & Buying
-
Table Of Contents
-
Feedback & Rating

Scala for Machine Learning
By :

The ordinary least squares method for finding the regression parameters is a specific case of the maximum likelihood. Therefore, regression models are subject to the same challenge in terms of overfitting as any other discriminative model. You are already aware that regularization is used to reduce model complexity and avoid overfitting as stated in the Overfitting section of Chapter 2, Hello World!.
Regularization consists of adding a penalty function J(w)
to the loss function (or RSS in the case of a regressive classifier) in order to prevent the model parameters (or weights) from reaching high values. A model that fits a training set very well tends to have many features variable with relatively large weights. This process is known as shrinkage. Practically, shrinkage involves adding a function with model parameters as an argument to the loss function:
The penalty function is completely independent from the training set {x,y}. The penalty term is usually...
Change the font size
Change margin width
Change background colour