Normal Equations
- Linear prediction in multi-dimensional spaces
- Normal equations and projections
- Different cases and pseudo-inversion
Polynomial Approximation:
In case of a polynomial function:
we can see it as a liner multi-variable function:
Normal Equation - Solutions
To solve a normal equation we will need more points than unknown variables, which means having enough data to learn the weights.
Given the matrix : “example” matrix, or data matrix, we put in each row one observation, plus the bias input (usually 1). The number of columns correspond to the total number of observation we took.
- More rows = More parameters = More weights to learn but a more “complete” solution, i.e. where the edge cases of the problem are taken into account
- More columns = More data = More precise weights
Number of columns Number of rows
Given:
- : information matrix
- : “perfect” weight matrix: such that
- : target Matrix
We have that: where is the maximum rank And we can say that:
So: