Normal Equations

  • Linear prediction in multi-dimensional spaces
  • Normal equations and projections
  • Different cases and pseudo-inversion

Polynomial Approximation:

In case of a polynomial function:

we can see it as a liner multi-variable function:


Normal Equation - Solutions

To solve a normal equation we will need more points than unknown variables, which means having enough data to learn the weights.

Given the matrix : “example” matrix, or data matrix, we put in each row one observation, plus the bias input (usually 1). The number of columns correspond to the total number of observation we took.

  • More rows = More parameters = More weights to learn but a more “complete” solution, i.e. where the edge cases of the problem are taken into account
  • More columns = More data = More precise weights

Number of columns Number of rows

Given:

  • : information matrix
  • : “perfect” weight matrix: such that
  • : target Matrix

We have that: where is the maximum rank And we can say that:

So: