Linear Models for Regression in Machine Learning
📂Machine LearningLinear Models for Regression in Machine Learning
Definition
Simple Model
Let’s define the target function f:X→Y between the data set X={xi} and the label set Y={yi} as follows.
yi=f(xi)
In machine learning, linear regression refers to finding a linear function f^ that satisfies the following equation for w.
yi≈y^i=f^(xi,w)=w0+w1x1+⋯+wnxn=w0+j∑wjxj
In this case, x∈Rn,w∈Rn+1.
Extended Model
Let’s say a non-linear function ϕj for input data is given.
yi≈y^i=f^(xi,w)=w0+j∑wjϕj(xi)
In this case, ϕj is called basis functions.
Explanation
Realistically, we can’t even be sure of the existence of the target function f. Therefore, the goal is to find f^ as similar as possible to f, assuming the simplest yet most explanatory linear function.
f≈f~
Even when introducing non-linear basis functions ϕ, it is called a linear model because f^ is linear with respect to the weights w.
Notation
If we denote by x0=1, ϕj=1, we can represent the two models more simply.
y^=f^(x,w)=j=0∑nwjxj=wTx
In this case, x=[x0…xn]T, w=[w0…wn]T.
y^=f^(x,w)=j=0∑nwjϕj(x)=wTϕ(x)
In this case, ϕ=[ϕ0…ϕn]T.
Linear Regression in Statistics
In statistics, instead of wi, it is denoted as βi, and these are called regression coefficients. In machine learning, wi is called weights.
| Statistics | Machine Learning |
---|
x | Independent Variable | Data |
y | Dependent Variable | Label |
w,β | Regression Coefficient | Weight |
Especially in machine learning, b=w0 is called bias.
Learning Methods
The methods to train a linear model include:
- [Gradient Descent]
- [Least Squares Method]
- Bayesian
- [Maximum Likelihood (ML)]
- [Maximum A Posteriori Probability (MAP)]