Skip to content

Least Square Problems⚓︎

Given ARm×n,mn;bRm, the Least Square problem is defined as: Find xRn, such that:

x=argminxbAx2

where r=bAx is called residual. Residual r0.

We will introduce three different methods to solve the Least Square problem.

Normal Equation⚓︎

Traditional Method⚓︎

We solve AAx=Ab, where AARn×n. If A is full rank, then AA is invertible.

The Least Square solution is:

x=(AA)1Ab

Question: How about A is not full rank?

Comment: This method is not recommended is κ(A) is large.

Note: κ(AA)=(κ(A))2

Usually in practice, the number of matrix rows are way larger than that of columns.

QR Factorization⚓︎

A=QRQA=R
Axb2=Q(Axb)2
=QAxQb2=RxQb2

Let c=Qb, then:

x=R1c=R1Qb

Assume A=UΣV as full SVD. We can get:

A=UΣVUAV=Σ;
Axb2=UAVVxUb2=ΣVxUb2

Define Vx=y, w=Ub, then:

Axb2=Σyw2;
argminyΣyw2
y=Σ1w=Σ1Ub,
x=Vy=VΣ1Ub

We get the formula:

x=VΣ1Ub

However, Σ1 may not exist! Assume:

Σ=[σ1σr00],σ1σ2σr>σr+1==σm=0

The pseudo inverse Σ1 is defined as:

Σ1=[1σ11σr00]

Then x=VΣ1Ub can be applied to the reduced SVD.

We can rewrite another formula of the result:

x=i=1ruiTbσivi

where ui,vi are columns of U,V respectively.

Comments: If σi is small, the error is enlarged by σi.

We can compute the approximate value ("the best approximation of x in Rk "):

y=i=1kuiTbσivi,k<r

This is the idea behind Principal Component Analysis (where to cut off?).