LinearAlgebra_2016updatedFromwikyyyi.ppt

srmvalliammaicse2 7 views 110 slides Aug 28, 2024
Slide 1
Slide 1 of 110
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25
Slide 26
26
Slide 27
27
Slide 28
28
Slide 29
29
Slide 30
30
Slide 31
31
Slide 32
32
Slide 33
33
Slide 34
34
Slide 35
35
Slide 36
36
Slide 37
37
Slide 38
38
Slide 39
39
Slide 40
40
Slide 41
41
Slide 42
42
Slide 43
43
Slide 44
44
Slide 45
45
Slide 46
46
Slide 47
47
Slide 48
48
Slide 49
49
Slide 50
50
Slide 51
51
Slide 52
52
Slide 53
53
Slide 54
54
Slide 55
55
Slide 56
56
Slide 57
57
Slide 58
58
Slide 59
59
Slide 60
60
Slide 61
61
Slide 62
62
Slide 63
63
Slide 64
64
Slide 65
65
Slide 66
66
Slide 67
67
Slide 68
68
Slide 69
69
Slide 70
70
Slide 71
71
Slide 72
72
Slide 73
73
Slide 74
74
Slide 75
75
Slide 76
76
Slide 77
77
Slide 78
78
Slide 79
79
Slide 80
80
Slide 81
81
Slide 82
82
Slide 83
83
Slide 84
84
Slide 85
85
Slide 86
86
Slide 87
87
Slide 88
88
Slide 89
89
Slide 90
90
Slide 91
91
Slide 92
92
Slide 93
93
Slide 94
94
Slide 95
95
Slide 96
96
Slide 97
97
Slide 98
98
Slide 99
99
Slide 100
100
Slide 101
101
Slide 102
102
Slide 103
103
Slide 104
104
Slide 105
105
Slide 106
106
Slide 107
107
Slide 108
108
Slide 109
109
Slide 110
110

About This Presentation

LinearAlgebra_2016updatedFromwiki.ppt


Slide Content

Introduction to Linear Algebra
Mark Goldman
Emily Mackevicius

Outline
1. Matrix arithmetic
2. Matrix properties
3. Eigenvectors & eigenvalues
-BREAK-
4. Examples (on blackboard)
5. Recap, additional matrix properties, SVD

Part 1: Matrix Arithmetic
(w/applications to neural networks)

Matrix addition

Scalar times vector

Scalar times vector

Scalar times vector

Product of 2 Vectors
•Element-by-element
•Inner product
•Outer product
Three ways to multiply

Element-by-element product
(Hadamard product)
•Element-wise multiplication (.* in MATLAB)

Multiplication:
Dot product (inner product)

Multiplication:
Dot product (inner product)

Multiplication:
Dot product (inner product)

Multiplication:
Dot product (inner product)

Multiplication:
Dot product (inner product)
1 X N N X 1 1 X 1
•MATLAB: ‘inner matrix dimensions must agree’Outer dimensions give
size of resulting matrix

Dot product geometric intuition:
“Overlap” of 2 vectors

Example: linear feed-forward network
Input neurons’
Firing rates
r
1
r
2
r
n
r
i

Example: linear feed-forward network
Input neurons’
Firing rates
Synaptic w
eights
r
1
r
2
r
n
r
i

Example: linear feed-forward network
Input neurons’
Firing rates
Output neuron’s
firing rate
Synaptic w
eights
r
1
r
2
r
n
r
i

Example: linear feed-forward network
•Insight: for a given input
(L2) magnitude, the
response is maximized
when the input is parallel
to the weight vector
•Receptive fields also can
be thought of this way
Input neurons’
Firing rates
Output neuron’s
firing rate
Synaptic w
eights
r
1
r
2
r
n
r
i

Multiplication: Outer product
N X 1 1 X M N X M

Multiplication: Outer product

Multiplication: Outer product

Multiplication: Outer product

Multiplication: Outer product

Multiplication: Outer product

Multiplication: Outer product

Multiplication: Outer product
•Note: each column or each row is a multiple of the others

Matrix times a vector

Matrix times a vector

Matrix times a vector
M X 1 M X N N X 1

Matrix times a vector:
inner product interpretation
•Rule: the i
th
element of y is the dot product of
the i
th
row of W with x

Matrix times a vector:
inner product interpretation
•Rule: the i
th
element of y is the dot product of
the i
th
row of W with x

Matrix times a vector:
inner product interpretation
•Rule: the i
th
element of y is the dot product of
the i
th
row of W with x

Matrix times a vector:
inner product interpretation
•Rule: the i
th
element of y is the dot product of
the i
th
row of W with x

Matrix times a vector:
inner product interpretation
•Rule: the i
th
element of y is the dot product of
the i
th
row of W with x

Matrix times a vector:
outer product interpretation
•The product is a weighted sum of the columns
of W, weighted by the entries of x

Matrix times a vector:
outer product interpretation
•The product is a weighted sum of the columns
of W, weighted by the entries of x

Matrix times a vector:
outer product interpretation
•The product is a weighted sum of the columns
of W, weighted by the entries of x

Matrix times a vector:
outer product interpretation
•The product is a weighted sum of the columns
of W, weighted by the entries of x

Example of the outer product method

Example of the outer product method
(3,1)
(0,2)

Example of the outer product method
(3,1)
(0,4)

Example of the outer product method
(3,5)
•Note: different
combinations of
the columns of
M can give you
any vector in
the plane
(we say the columns of M
“span” the plane)

Rank of a Matrix
•Are there special matrices whose columns
don’t span the full plane?

Rank of a Matrix
•Are there special matrices whose columns
don’t span the full plane?
(1,2)
(-2, -4)
•You can only get vectors along the (1,2) direction
(i.e. outputs live in 1 dimension, so we call the
matrix rank 1)

Example: 2-layer linear network
•W
ij is the connection strength
(weight) onto neuron y
i from
neuron x
j.

Example: 2-layer linear network:
inner product point of view
•What is the response of cell y
i
of the second layer?
•The response is the dot
product of the i
th
row of
W with the vector x

Example: 2-layer linear network:
outer product point of view
•How does cell x
j contribute to the pattern of firing of
layer 2?
Contribution
of x
j to
network output
1
st
column
of W

Product of 2 Matrices
•MATLAB: ‘inner matrix dimensions must agree’
•Note: Matrix multiplication doesn’t (generally) commute, AB  BA
N X P P X M N X M

Matrix times Matrix:
by inner products
•C
ij is the inner product of the i
th
row with the j
th
column

Matrix times Matrix:
by inner products
•C
ij is the inner product of the i
th
row with the j
th
column

Matrix times Matrix:
by inner products
•C
ij is the inner product of the i
th
row with the j
th
column

Matrix times Matrix:
by inner products
•C
ij is the inner product of the i
th
row of A with the j
th
column of B

Matrix times Matrix:
by outer products

Matrix times Matrix:
by outer products

Matrix times Matrix:
by outer products

Matrix times Matrix:
by outer products
•C is a sum of outer products of the columns of A with the rows of B

Part 2: Matrix Properties
• (A few) special matrices
• Matrix transformations & the determinant
• Matrices & systems of algebraic equations

Special matrices: diagonal matrix
•This acts like scalar multiplication

Special matrices: identity matrix
for all

Special matrices: inverse matrix
•Does the inverse always exist?

How does a matrix transform a square?
(1,0)
(0,1)

How does a matrix transform a square?
(1,0)
(0,1)

How does a matrix transform a square?
(3,1)
(0,2)
(1,0)
(0,1)

Geometric definition of the determinant:
How does a matrix transform a square?
(1,0)
(0,1)

Example: solve the algebraic equation

Example: solve the algebraic equation

Example: solve the algebraic equation

•Some non-zero vectors are sent to 0
Example of an underdetermined system

•Some non-zero vectors are sent to 0
Example of an underdetermined system

Example of an underdetermined system

Example of an underdetermined system
•Some non-zero x are sent to 0 (the set of all x with Mx=0 are called the “nullspace” of M)
•This is because det(M)=0 so M is not invertible. (If det(M) isn’t 0, the only solution is x = 0)

Part 3: Eigenvectors & eigenvalues

What do matrices do to vectors?
(3,1)
(0,2)
(2,1)

Recall
(3,5)
(2,1)

What do matrices do to vectors?
(3,5)
•The new vector is:
1) rotated
2) scaled
(2,1)

Are there any special vectors
that only get scaled?

Are there any special vectors
that only get scaled?
Try (1,1)

Are there any special vectors
that only get scaled?
= (1,1)

Are there any special vectors
that only get scaled?
= (3,3)
= (1,1)

Are there any special vectors
that only get scaled?
•For this special vector,
multiplying by M is like
multiplying by a scalar.
•(1,1) is called an
eigenvector of M
•3 (the scaling factor) is
called the eigenvalue
associated with this
eigenvector
= (1,1)
= (3,3)

Are there any other eigenvectors?
•Yes! The easiest way to find is with MATLAB’s
eig command.
•Exercise: verify that (-1.5, 1) is also an
eigenvector of M.
•Note: eigenvectors are only defined up to a
scale factor.
–Conventions are either to make e’s unit vectors,
or make one of the elements 1

Step back:
Eigenvectors obey this equation

Step back:
Eigenvectors obey this equation

Step back:
Eigenvectors obey this equation

Step back:
Eigenvectors obey this equation
•This is called the characteristic equation for 
•In general, for an N x N matrix, there are N
eigenvectors

BREAK

Part 4: Examples (on blackboard)
• Principal Components Analysis (PCA)
• Single, linear differential equation
• Coupled differential equations

Part 5: Recap & Additional useful stuff
• Matrix diagonalization recap:
transforming between original & eigenvector coordinates
• More special matrices & matrix properties
• Singular Value Decomposition (SVD)

Coupled differential equations
•Calculate the eigenvectors and eigenvalues.
–Eigenvalues have typical form:
•The corresponding eigenvector component has
dynamics:

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx
eig(M) in MATLAB

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx

•Step 1: Find the
eigenvalues and
eigenvectors of M.
•Step 2: Decompose x
into its eigenvector
components
•Step 3: Stretch/scale
each eigenvalue
component
•Step 4: (solve for c
and) transform back
to original
coordinates.
Practical program for approaching
equations coupled through a term Mx

Where (step 1):
MATLAB:
Putting it all together…

Step 2: Transform
into eigencoordinates
Step 3: Scale by 
i

along the i
th

eigencoordinate
Step 4: Transform
back to original
coordinate system
Putting it all together…

Left eigenvectors
-The rows of E inverse are called the left eigenvectors
because they satisfy E
-1
M =  E
-1
.
-Together with the eigenvalues, they determine how x is
decomposed into each of its eigenvector components.

Matrix in
eigencoordinate
system
Original Matrix
Where:
Putting it all together…

•Note: M and Lambda look very different.
Q: Are there any properties that are preserved between them?
A: Yes, 2 very important ones:
Matrix in eigencoordinate
system
Original Matrix
Trace and Determinant
2.
1.

Special Matrices: Normal matrix
•Normal matrix: all eigenvectors are orthogonal
 Can transform to eigencoordinates (“change basis”) with a
simple rotation* of the coordinate axes
 A normal matrix’s eigenvector matrix E is a *generalized
rotation (unitary or orthonormal) matrix, defined by:
E
Picture:
(*note: generalized means one can also do reflections of the eigenvectors through a line/plane”)

Special Matrices: Normal matrix
•Normal matrix: all eigenvectors are orthogonal
 Can transform to eigencoordinates (“change basis”)
with a simple rotation of the coordinate axes
 E is a rotation (unitary or orthogonal) matrix, defined
by:
where if: then:

Special Matrices: Normal matrix
•Eigenvector decomposition in this case:
•Left and right eigenvectors are identical!

•Symmetric Matrix:
Special Matrices
•e.g. Covariance matrices, Hopfield network
•Properties:
–Eigenvalues are real
–Eigenvectors are orthogonal (i.e. it’s a normal matrix)

SVD: Decomposes matrix into outer products
(e.g. of a neural/spatial mode and a temporal mode)
t = 1 t = 2 t = T
n = 1
n = 2
n = N

SVD: Decomposes matrix into outer products
(e.g. of a neural/spatial mode and a temporal mode)
n = 1
n = 2
n = N
t = 1 t = 2 t = T

SVD: Decomposes matrix into outer products
(e.g. of a neural/spatial mode and a temporal mode)
Rows of V
T
are
eigenvectors of
M
T
M
Columns of U
are eigenvectors
of MM
T
•Note: the eigenvalues are the same for M
T
M and MM
T

SVD: Decomposes matrix into outer products
(e.g. of a neural/spatial mode and a temporal mode)
Rows of V
T
are
eigenvectors of
M
T
M
Columns of U
are eigenvectors
of MM
T
•Thus, SVD pairs “spatial” patterns with associated
“temporal” profiles through the outer product

The End