2-Linear Transformations and least squares.ppt

ssuser362a24 12 views 32 slides Sep 10, 2024
Slide 1
Slide 1 of 32
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25
Slide 26
26
Slide 27
27
Slide 28
28
Slide 29
29
Slide 30
30
Slide 31
31
Slide 32
32

About This Presentation

Linear Transformations and least squares


Slide Content

Professor A G Constantinides© 1
AG
C
DSP
Hilbert Spaces
Linear Transformations and Least Squares:
Hilbert Spaces

Professor A G Constantinides© 2
AG
C
DSP
Linear Transformations
A transformation from a vector space to a vector space
with the same scalar field denoted by
is linear when

Where
We can think of the transformation as an operator

YXL:
X
Y
2121)( xxxxL 
)()( xaLaxL
Xxxx 
21,,

Professor A G Constantinides© 3
AG
C
DSP
Linear Transformations …

Example: Mapping a vector space from to
can be expressed as a mxn matrix.

Thus the transformation
can be written as
n
R
m
R
)43,2().,(
3221321 xxxxxxxL 























3
2
1
2
1
430
021
x
x
x
y
y

Professor A G Constantinides© 4
AG
C
DSP
Range space & Null space

The range space of a transformation
is the set of all vectors that can be reached by the
transformation

The null space of the transformation is the set of all
vectors in X that are transformed to the null vector in
Y.
 XLLR  xxy :)()(
YXL:
 XLLN  x0x:)()(

Professor A G Constantinides© 5
AG
C
DSP
Range space & Null space …

If is a projection operator then so is

Hence we have

Thus the vector is decomposed into two disjoint
parts. These parts are not necessarily orthogonal

If the range and null space are orthogonal then the
projections is said to be orthogonal
P PI
xP)x(IPx 
x

Professor A G Constantinides© 6
AG
C
DSP
Linear Transformations
Example: Let
and let the transformation a nxm matrix
Then
Thus, the range of the linear transformation (or column
space of the matrix ) is the span of the basis vectors.
The null space is the set which yields
 
T
mxxxx ...
321x
mm332211 ppppAx xxxx  ...
A
0Ax

Professor A G Constantinides© 7
AG
C
DSP
A Problem

Given a signal vector in the vector space S, we
want to find a point in the subset V of the space ,
nearest to
x
v
x
x
0v1v
2v
0w
V
W
00 vxw 

Professor A G Constantinides© 8
AG
C
DSP
A Problem …
Let us agree that “nearest to” in the figure is taken in
the Euclidean distance sense.
The projection orthogonal to the set V gives the
desired solution.
Moreover the error of representation is
This vector is clearly orthogonal to the set V (More
on this later)
00 vxw 
0v

Professor A G Constantinides© 9
AG
C
DSP
Another perspective …

We can look at the above problem as seeking to find
a solution to the set of linear equations
where the given vector is not in the range of
as is the case with an overspecified set of equations.

There is no exact solution. If we project orthogonally
the given vector into the range of then we have
the “shortest norm solution” in terms of the Euclidean
distance of the “error”.
xAv
v
x A
A

Professor A G Constantinides© 10
AG
C
DSP
Another perspective …
The least error is then orthogonal to the data into
which we are projecting
Set
Then as in the above figure we can write
Where is the error, which is orthogonal to each
of the members of above.
 
m321 ppppA ...
m321 ppppv
mvvvv  ...
321
wAvx 
w
A

Professor A G Constantinides© 11
AG
C
DSP
Another perspective …

Thus we can write

Or
mj
j ,...,3,2,1,0, pAvx
0Avx
p
p
p













)(
.
2
1
H
m
H
H
0AvxA )(
H
xAAAv
HH 1
)(

Professor A G Constantinides© 12
AG
C
DSP
Another perspective …

Thus
and hence the projection matrix is
ie this is the matrix that projects orthogonally into the
column space of
HH
AAAP
1
)(


PxxAAAv 
HH 1
)(
A

Professor A G Constantinides© 13
AG
C
DSP
Another perspective …

If we adopt the weighted form

The induced norm is

Then the projection matrix is

Where is positive definite
WAWAAAP
HH 1
)(


Wyxyx,
W
H

Wxxx
W
H

2
W

Professor A G Constantinides© 14
AG
C
DSP
Least Squares Projection
PROJECTION THEOREM
In a Hilbert space the orthogonal projection of a
signal into a smaller dimensional space minimises the
norm of the error, and the error vector is orthogonal
to the data (ie the smaller dimensional space).

Professor A G Constantinides© 15
AG
C
DSP
Orthogonality Principle

Let be a set of
independent vectors in a vector space S.

We wish to express any vector in S as

If is in the span of the independent vectors
then the representation will be exact.

If on the other hand it is not then there will be an
error
}...{
m321 pppp
x
mm332211 ppppx xxxx  ...
x

Professor A G Constantinides© 16
AG
C
DSP
Orthogonality Principle

In the latter case we can write

Where
is an approximation to given vector with error

We wish to find that approximation which minimises
the Euclidean error norm (squared)
i
m
i
i
xpx
1
ˆ
exxˆ
e
i
m
i
ii
m
i
im
xxxxJ pxpx 
 11
1
,),...,(

Professor A G Constantinides© 17
AG
C
DSP
Orthogonality Principle

Expand to

Where

 
T
m
xxx ...
21
χ








m
i
iji
m
j
ji
m
i
im
xxxxxJ
1
*
11
*
1
,,Re2,),...,( pppxxx
 χRχpχx
THH
m
xxJ  Re2),...,(
2
1
 
T
m
pxpxpxp ,...,,
21

Professor A G Constantinides© 18
AG
C
DSP
Reminders


0ax
x



*
aax
x



H
*
2/)Re(
*
aax
x



H RxRxx
x



H
*

Professor A G Constantinides© 19
AG
C
DSP
Orthogonality Principle

On setting this to zero we obtain the solution

This is a minimum because on differentiating we
have a positive definite matrix

  RχpχRχpχx
χ



THH
Re2
2
*
pRχ
1

Professor A G Constantinides© 20
AG
C
DSP
Alternatively …

The norm squared of the error is

where

We note that
and
ee
T
J
ee
T
ii
x
J
x











2
i
i
x
pe

Professor A G Constantinides© 21
AG
C
DSP
Orthogonality Principle

At the minimum

Thus we have
and hence
Thus,
1) At the optimum the error is orthogonal to the data (Principle of
orthogonality)
2)
0ˆ xx,pe,p
ii
022 










epee
T
i
T
ii
x
J
x
mix
m
j
jjiii
,...,1,ˆ,
1


ppxpx,p

Professor A G Constantinides© 22
AG
C
DSP
Orthogonality Principle

Thus for
Hence or
 
T
m
pxpxpxp ,...,,
21

 
T
m
xxx ...
21
χ













mmmm
m
m
pppppp
pppppp
pppppp
R
,..,,
........
,..,,
,..,,
21
22212
12111
pRχ
1
Rχp

Professor A G Constantinides© 23
AG
C
DSP
Orthogonalisation

A signal may be projected into any linear space.

The computation of its coefficients in the various
vectors of the selected space is easier when the
vectors in the space are orthogonal in that they are
then non-interacting, ie the evaluation of one such
coefficient will not influence the others

The error norm is easier to compute

Thus it makes sense to use an orthogonal set of
vectors in the space into which we are to project a
signal

Professor A G Constantinides© 24
AG
C
DSP
Orthogonalisation
Given any set of linearly independent vectors that
span a certain space, there is another set of
independent vectors of the same cardinality, pair-
wise orthogonal, that spans the same space
We can think of the given set as a linear
combination of orthogonal vectors
Hence because of independence, the orthogonal
vectors is a linear combination of the given vectors
This is the basic idea behind the Gram-Schmidt
procedure

Professor A G Constantinides© 25
AG
C
DSP
Gram-Schmidt Orthogonalisation

The problem: (we consider finite dimensional spaces
only)

Given a set of linearly independent vectors
to determine a set of vectors that are pair-
wise orthogonal

Write the ith vector as
}{x
}{p
mixxxx
m
i
m
iii
i ,...,1...
)(
3
)(
32
)(
21
)(
1  ppppx

Professor A G Constantinides© 26
AG
C
DSP
Gram-Schmidt Orthogonalisation

If we knew the orthogonal set then the
coefficients of the expression can be determined as
the inner product


Step(1) The unknown orthogonal vector can be
oriented such that one of its members coincides with
one of the members of the given set

Choose to be coincident with
}{x
}{p
2
,
jjji xppx 
1p
1x

Professor A G Constantinides© 27
AG
C
DSP
Gram-Schmidt Orthogonalisation

Step (2) Each member of has a projection
onto given by

Step(3) We construct

Step(4) Repeat the above on
}{x
2
1
)(
11
, ppx
i
i
x
1p
}{u
mix
i
ii ,...,2
1
)(
1  pxu

Professor A G Constantinides© 28
AG
C
DSP
Gram-Schmidt Orthogonalisation

Example: Let

Then

And the projection of onto is







1
1
1x







2
1
2x







1
1
1p
 1
1
1
21 







2x
1p

Professor A G Constantinides© 29
AG
C
DSP
Gram-Schmidt Orthogonalisation

Form

Then





















5.1
5.1
2/
1
1
1
2
1
/1
2
112 ppx







5.1
5.1
2p

Professor A G Constantinides© 30
AG
C
DSP
Gram-Schmidt
-2 -1.5 -1 -0.5 0 0.5 1
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
1x
212
xxp a
11
xp
2
x
Projection
of in
2x
1
x
2p
1xa

Professor A G Constantinides© 31
AG
C
DSP
3-D G-S Orthogonalisation
-1
-0.5
0
0.5
1
-1
-0.5
0
0.5
1
-1
-0.5
0
0.5
1

Professor A G Constantinides© 32
AG
C
DSP
Gram-Schmidt Orthogonalisation

Note that in the previous 4 steps we have
considerable freedom at Step 1 to choose
any vector not necessarily coincident with
one from the given set of data vectors.

This enables us to avoid certain numerical
ill-conditioning problems that may arise in
the Gram-Schmidt case.

Can you suggest when we are likely to have
ill-conditioning in the G-S procedure?
Tags