the rank of a product of two matrices x and y is equal to the smallest of the rank of x and y:
DESCRIPTION
S. A. C. The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:. Rank (X Y) =min (rank (X) , rank (Y)). =. Eigenvectors and Eigenvalues. 0. v. R. I. R. v. = l. - l. v. For a symmetric, real matrix, R , an eigenvector v is obtained from:. - PowerPoint PPT PresentationTRANSCRIPT
![Page 1: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/1.jpg)
The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:
Rank (X Y) =min (rank (X) , rank (Y))
00.05
0.1
0.150.2
0.250.3
0.350.4
400 450 500 550 600Wavelength (nm)
Abso
rban
ce
0
0.5
1
1.5
2
2.5
0 2 4 6 8 10 12
Time (s)
Con
cent
ratio
n
A = CS
![Page 2: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/2.jpg)
![Page 3: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/3.jpg)
![Page 4: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/4.jpg)
![Page 5: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/5.jpg)
![Page 6: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/6.jpg)
Eigenvectors and EigenvaluesFor a symmetric, real matrix, R, an eigenvector v is obtained from:
Rv = v is an unknown scalar-the eigenvalue
Rv – v= 0 (R – Iv= 0The vector v is orthogonal to all of the row
vector of matrix (R-I)
R v = v
0v- R I =
![Page 7: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/7.jpg)
0.1 0.2 0.3
0.2 0.4 0.6A= R=ATA =
0.14 0.28
0.28 0.56
Rv = v(R – Iv= 00.14 0.28
0.28 0.56
1 0
0 1-
v1
v2
0
0=
=0.14 - 0.28
0.28 0.56 - =0
0
v1
v2
0
0
0.14 0.28
0.28 0.56-
v1
v2
![Page 8: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/8.jpg)
(0.14 – ) (0.56 – ) – (0.28) (0.28) = 0
= 0
= 0
&
0.14 - 0.28
0.28 0.56 - = 0
![Page 9: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/9.jpg)
For
=0
0
0.14 – 0.28
0.28 0.56 –
v11
v21
-0.56 0.28
0.28 -0.14=
v11
v21
-0.56 v11 + 0.28v21 = 0
0.28 v11 - 0.14 v21 = 0v21 = 2 v11
Normalized vector v1 =0.4472
0.8944
If v11 = 1 v21 = 2
![Page 10: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/10.jpg)
0.14 0.28
0.28 0.56 =0
0
v12
v22
0.14 v12 + 0.28 v22 = 0
0.28 v12 +0.56 v22 = 0
v12 = -2 v22
If v22 = 1 v12 = -2
Normalized vector v1 =-0.8944
0.4472
For
![Page 11: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/11.jpg)
0.1 0.2 0.3
0.2 0.4 0.6A=
Rv = vRV = V
V =-0.8944
0.4472
0.4472
0.8944
0.7 0
0 0v1v2 =0
R=ATA =0.14 0.28
0.28 0.56
More generally, if R (p x p) is symmetric of rank r≤p then R posses r positive eigenvalues and (p-r) zero eigenvalues
tr(R) = i= 0.7 + 0.0 =0.7∑
![Page 12: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/12.jpg)
![Page 13: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/13.jpg)
![Page 14: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/14.jpg)
![Page 15: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/15.jpg)
Example
![Page 16: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/16.jpg)
Consider 15 sample each contain 3 absorbing components
![Page 17: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/17.jpg)
![Page 18: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/18.jpg)
![Page 19: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/19.jpg)
![Page 20: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/20.jpg)
![Page 21: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/21.jpg)
?Show that in the presence of random noise the number of non-zero eigenvalues is larger than numbers of components
![Page 22: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/22.jpg)
Variance-Covariance Matrixx11 – mx1
…
x21 – mx1
(xn1 – mx1)
x12 – mx2
x22 – mx2
xn2 – mx2
…
……
…
…
x1p – mxp
x1p – mxp
xnp – mxp
…X =
Column mean centered matrix
XTX =
var(x1)
… ……
…
…… …
var(x2)
var(xp)
covar(x1x2) covar(x1xp)
covar(x2x1) covar(x2xp)
covar(xpx1) covar(xpx2)
![Page 23: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/23.jpg)
mmcn.m file for mean centering a matrix
![Page 24: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/24.jpg)
![Page 25: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/25.jpg)
![Page 26: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/26.jpg)
![Page 27: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/27.jpg)
?
Use anal.m file and mmcn.m file and verify that each eigenvalue of an absorbance data matrix is correlated with variance of data
![Page 28: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/28.jpg)
Singular Value DecompositionSVD of a rectangular matrix X is a method which yield at the same time a diagnal matrix of singular values S and the two matrices of singular vectors U and V such that :
X = U S VT UTU = VTV =Ir
The singular vectors in U and V are identical to eigenvectors of XXT AND XTX, respectively and the singular values are equal to the positive square roots of the corresponding eigenvalues
X = U S VT XT = V S UT
X XT= U S VT VSUT= US2UT
(X XT) U = US2
![Page 29: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/29.jpg)
=
X = U S VT = s1u1v1T + … + srurvr
T
X
m
n
U
m
n
Sn
n
VT
n
n
If the rank of matrix X=r then;
X
m
n
= U
m
r
Sr
r
VT
r
n
![Page 30: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/30.jpg)
Singular value decomposition with MATLAB
![Page 31: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/31.jpg)
![Page 32: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/32.jpg)
![Page 33: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/33.jpg)
Consider 15 sample containing 2 component with strong spectral overlapping and construct their absorbance data matrix accompany with random noise
Ideal dataA
Noised datand
Reconstructed datard
residualR1
Ideal dataA
residualR2
- =
- =
It can be shown that the reconstructed data matrix is closer to ideal data matrix
![Page 34: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/34.jpg)
Anal.m file for constructing the data matrix
![Page 35: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/35.jpg)
Spectral overlapping of two absorbing species
![Page 36: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/36.jpg)
Ideal data matrix A
![Page 37: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/37.jpg)
Noised data matrix, nd, with 0.005 normal distributed random noise
![Page 38: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/38.jpg)
nf.m file for investigating the noise filtering property of svd reconstructed data
![Page 39: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/39.jpg)
![Page 40: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/40.jpg)
![Page 41: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/41.jpg)
![Page 42: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/42.jpg)
![Page 43: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/43.jpg)
![Page 44: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/44.jpg)
?Plot the %relative standard error as a function of number of eigenvectors
![Page 45: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/45.jpg)
x11 x12 x114
x21 x21 x214…
…
Principal Component Analysis (PCA)
• • • • • • • • • • • • • •
x1
x2
![Page 46: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/46.jpg)
PCA
• • • • • • • • • • • • • •
u 1
u 2
u11
u12
u114
…
![Page 47: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/47.jpg)
•
• ••
••
• •• •
•• ••
x1
x2
x11 x12 x114
x21 x21 x214…
…
![Page 48: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/48.jpg)
•
• ••
••
• •• •
•• •• u 1u 2 u11
u12
u114
…
u21
u22
u214
…
![Page 49: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/49.jpg)
Principal Components in two Dimensions
u1 = ax1 + bx2
u2 = cx1 + dx2
0.10.2
0.20.4
0.30.6
1 2 s1
s2 s3
In principal components model new variables are found which give a clear picture of the variability of the data. This is best achieved by giving the first new variable maximum variance, the second new variable is then selected so as to be uncorrelated with the first one, and so on
![Page 50: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/50.jpg)
The new variables can be uncorrelated if:ac + bd =0
a=1 b=2 c=-1 d=0.50.1
0.2
0.3
x1 =
0.2
0.4
0.6
x2 =
0.5
1.0
1.5
u1 = var(u1)=0.25
a=2 b=4 c=-2 d=1
1.0
2.0
3.0
u1 = var(u1)=1.0
Orthogonality constraint
![Page 51: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/51.jpg)
Normalizing constraint a2 + b2 = 1c2 + d2 = 1
a=1 b=2
c=-1 d=0.5
a=0.4472 b=0.8944
c=-0.8944 d=0.4472Normalizing
a=2 b=4
c=-2 d=1Normalizing a=0.4472 b=0.8944
c=-0.8944 d=0.4472
![Page 52: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/52.jpg)
Maximum variance constraint
u1 = ax1 + bx2
2u1 = a2 2
x1 + b2 2x2 + 2ab x1-x2
2u1 = [ a b ]
2x1 x1-x2
x1-x2 2x2
a
b
= 2u1
2x1 x1-x2
x1-x2 2x2
a
b
a
b
![Page 53: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/53.jpg)
Principal Components in m Dimensionsx11 x12 … x1m
x21 x22 … x2m
xn1 xn2 … xnm
…
…
… X=
u1 = v11x1 + v12x2 + … + v1mxm
u2 = v21x1 + v22x2 + … + v2mxm
um = vm1x1 + vm2x2 + … + vmmxm
…
…
…
…
![Page 54: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/54.jpg)
var(x1)… …
……
…
… …
var(x2)
var(xm)
covar(x1x2) covar(x1xm)
covar(x2x1) covar(x2xm)
covar(xmx1) covar(xmx2)
v11
v21
vm1
…
var(u1)=
v11
v21
vm1
…
C V V=
Xn
m m
mV U
n
m
=
![Page 55: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/55.jpg)
X V =UX
n
m m
mV U
n
m
=
Loading vectors Score vectors
X VTV = UVT
VT V = I X = UVT = S LT
Xn
ms
n
mm
mLT
=
X = USVT S = US
![Page 56: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/56.jpg)
More generally, when one analyzes a data matrix consisting of n objects for which m variables have been determined, m principal components can then be extracted (as long as m<n.
PC1 represents the direction in the data containing the largest variation. PC2 is orthogonal to PC1 and represents the direction of the largest residual variation around PC1. PC3 is orthogonal to the first two and represents the direction of the highest residual variation around the plane formed by PC1 and PC2.
![Page 57: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/57.jpg)
PCA.m file
![Page 58: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/58.jpg)
![Page 59: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/59.jpg)
![Page 60: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/60.jpg)
10 mixtures of two componentsanal.m file
![Page 61: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/61.jpg)
![Page 62: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/62.jpg)
![Page 63: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/63.jpg)
![Page 64: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/64.jpg)
![Page 65: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/65.jpg)
![Page 66: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/66.jpg)
![Page 67: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/67.jpg)
![Page 68: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/68.jpg)
![Page 69: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/69.jpg)
![Page 70: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/70.jpg)
![Page 71: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/71.jpg)
![Page 72: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/72.jpg)
![Page 73: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/73.jpg)
![Page 74: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/74.jpg)
![Page 75: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/75.jpg)
![Page 76: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/76.jpg)
?
Perform PCA on data matrix obtained from an evolutionary process, such as kinetic data (kin.m file) and interpret the score vectors.
![Page 77: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/77.jpg)
Classification with PCA
The most informative view of a data set, in terms of variance at least, will be given by consideration of the first two PCs. Since the scores matrix contains a value for each sample corresponding to each PC, it is possible to plot these values against one another to produce a low dimensional picture of a high-dimensional data set.
![Page 78: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/78.jpg)
Suppose there are 20 sample from two different class
0
0.05
0.1
0.15
0.2
0.25
0.3
400 450 500 550 600
Wavelength (nm)
Abso
rban
ce
Class I Class II
0
0.04
0.08
0.12
0.16
400 450 500 550 600
Wavelength (nm)
Abso
rban
ce
00.020.040.060.08
0.10.12
0 0.05 0.1 0.15Abs. (1)
Abs
. (2)
![Page 79: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/79.jpg)
0
0.025
0.05
0.075
0.1
0.125
0.15
400 450 500 550 600Wavelength (nm)
Abso
rban
ce
00.020.040.060.08
0.10.120.140.16
0 0.05 0.1 0.15Abs. ( 1)
Abs
. (
2)0
0.05
0.1
0.15
0.2
0.25
0.3
400 450 500 550 600Wavelength (nm)
Abso
rban
ce
Class I Class II
![Page 80: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/80.jpg)
![Page 81: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/81.jpg)
-0.12-0.1
-0.08-0.06-0.04-0.02
00.020.040.060.080.1
-0.8 -0.6 -0.4 -0.2 0
PC1
PC2
![Page 82: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/82.jpg)
Multiple Linear Regression (MLR)
y = b1 x1 + b2 x2 + … + bp xp
x11 …
x21
xn1
x12
x22
xn2
…
……
……
x1p
x2p
xnp…
b1b2
bp
…
y1
y2
yn
… = y = X b
= y
n
1
X n
pb1
p
![Page 83: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/83.jpg)
If p>na1 = 1 c11 + 2 c12 + 3 c13
a2 = 1 c21 + 2 c22 + 3 c23
There is an infinite number of solution for , which all fit the equation
If p=na1 = 1 c11 + 2 c12 + 3 c13
a2 = 1 c21 + 2 c22 + 3 c23
a3 = 1 c31 + 2 c32 + 3 c33
It gives a unique solution for provided that the X matrix has ful rank
![Page 84: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/84.jpg)
If p<n
y1 = 1 c11 + 2 c12 + 3 c13
a2 = 1 c21 + 2 c22 + 3 c23
a3 = 1 c31 + 2 c32 + 3 c33
a4= 1 c41 + 2 c42 + 3 c43
This does not allow an exact solution for , but one can get a solution by minimizing the length of the residual vector eThe least squares solution is
= (CTC)-1 CT a
![Page 85: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/85.jpg)
Least Squares in Matrix Equations
= y
n
1
X n
p
y = X b
y
n
1
= x1
n
1
x2
n
1
xp
n
1
b111
bp11
b211
+ + … +
For solving this system the Xb-y must be perpendicular to the column space of X
1
pb
![Page 86: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/86.jpg)
Suppose vector Xc is a linear combination of the columns of X :
(Xc)T [Xb – y]=0
c [XTXb –XT y]=0XTXb = XT y
b = (XTX)-1 XT yThe projection of y onto the column space of X is therefore
p=Xb = (X (XTX)-1 XT )y
![Page 87: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/87.jpg)
Least Squares Solution
![Page 88: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/88.jpg)
![Page 89: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/89.jpg)
![Page 90: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/90.jpg)
![Page 91: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/91.jpg)
Projection the y vector in column space of X
![Page 92: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/92.jpg)
The error vector
![Page 93: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/93.jpg)
The error vector is perpendicular to all columns of X matrix
![Page 94: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/94.jpg)
MLR with more than one dependent variable
= y1
n
1
X n
p 1
pb1y3
n
1
y2
n
1 1
pb2
1
pb3
=Y
n
m
X
n
p
Bp
m
Y= X B B= (XTX) -1 Y
![Page 95: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/95.jpg)
Classical Least Squares (CLS)A= C K
=A
n
m
C
n
p
Kp
m
Calibration step K = (CTC)-1 CT AThe number of calibration standards should at least be as large as the number of analytesThe rank of C must be equal to p
Prediction step aTun= cT
un K cun= (KKT)-1 K aun
Number of wavelengths mustbe equal or larger than number of components
![Page 96: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/96.jpg)
Advantages of CLSFull spectral domain is used for estimating each constituent. Using redundant information has an effect equivalent to replicated measurement and signal averaging, hence it improves the precision of the concentration estimates.
Disadvantages of CLSThe concentration of all the constituents in the calibration set have to be known
![Page 97: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/97.jpg)
Simultaneous determination of two
components with CLS
![Page 98: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/98.jpg)
Random design of concentration matrix
![Page 99: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/99.jpg)
Pure component spectra
![Page 100: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/100.jpg)
Absorbance data matrix
![Page 101: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/101.jpg)
Data matrices for mlr.m file
![Page 102: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/102.jpg)
mlr.m file for multiple linear regression
![Page 103: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/103.jpg)
![Page 104: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/104.jpg)
![Page 105: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/105.jpg)
Predicted concentrations
Real concentrations
![Page 106: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/106.jpg)
?Use CLS method for determination of one component in binary mixture samples
![Page 107: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/107.jpg)
Inverse Least Squares (ILS)c= A b
Calibration step b = (ATA)-1 AT c1
The number of calibration samples should at least be as large as the number of wavelengthsThe rank of A must be equal to p
Prediction step cTun= aT
un b
= c1
n
1
A n
p 1
pb
![Page 108: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/108.jpg)
Advantages of ILSIt is not necessary to know all the information on possible constituents, analyte of interest and interferents
Disadvantages of ILSThe number of calibration samples should at least be as large as the number of wavelengths
The method can work in principal when unknown chemical interferents are present. It is important that such interferents are present in calibration samples
![Page 109: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/109.jpg)
Determination of x in the presence of y by ILS
method
![Page 110: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/110.jpg)
![Page 111: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/111.jpg)
![Page 112: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/112.jpg)
15 x 9 absorbance data matrix
![Page 113: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/113.jpg)
ILS.m file
![Page 114: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/114.jpg)
ILS calibration
![Page 115: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/115.jpg)
Predicted concentrationsReal concentrations
![Page 116: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/116.jpg)
?Does in ILS method the accuracy of final results is dependent to number of wavelength?
![Page 117: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/117.jpg)
Principal Component Regression (PCR)PCR is simply PCA followed by a regression step
A= C E = S L
A C E= S L=
A= C E = (S R) (R-1 L)C = S R
C S R=
S r=c1
![Page 118: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/118.jpg)
A data matrix can be represented by its score matrixA regression of score matrix against one or several dependent variables is possible, provided that scores corresponding to small eigenvalues are omittedThis regression gives no matrix inversion problemPCR has the full-spectrum advantages of the CLS methodPCR has the ILS advantage of being able to perform the analysis one chemical components at a time while avoiding the ILS wavelength selection problem
![Page 119: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/119.jpg)
ValidationHow many meaningful principal components should be retained?
*Percentage of explained varianceIf all possible PCs are used in the model 100% of the variance is explained
sd2 =
∑ ii=1
d
∑ ii=1
p x 100
![Page 120: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/120.jpg)
Percentage of explained variance for determination of number of PCs
![Page 121: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/121.jpg)
Spectra of 20 samples of various amount of 2 components
![Page 122: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/122.jpg)
Pev.m file for
percentage of explained variance method
![Page 123: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/123.jpg)
Performing pev.m file on nd absorbance data matrix
![Page 124: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/124.jpg)
![Page 125: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/125.jpg)
![Page 126: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/126.jpg)
?Show the validity of results of Percentage Explained Variance method is dependent to spectral overlapping of individual components
![Page 127: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/127.jpg)
An
p
loadingn
p
n
n
score
PCA
![Page 128: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/128.jpg)
An
p
A’n-1
p
cp1 a
Cross-Validation
![Page 129: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/129.jpg)
Creating absorbance data for performing cross-validation
method
![Page 130: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/130.jpg)
Spectra of 15 samples of various amount of 3 components
![Page 131: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/131.jpg)
cross.m file for
PCR cross-validation
![Page 132: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/132.jpg)
PCR cross-validation
![Page 133: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/133.jpg)
01234567
0 5 10 15no of factors
PRES
Scross-validation plot
![Page 134: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/134.jpg)
c = S b
Calibration and Prediction Steps in PCR
=c1
n
1
Sn
rb
r
1
b = ( STS)-1 ST c
Calibration Step
Axm
p
Lp
rr
mSx =
Prediction StepSx = Ax L
cx = Sx b
![Page 135: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/135.jpg)
Pcr.m file for calibration and prediction by PCR method
![Page 136: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/136.jpg)
Spectra of 20 samples of various amount of 3 components
![Page 137: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/137.jpg)
Input data for pcr.m file
pcr.m function
![Page 138: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/138.jpg)
Predicted and real values for first component
![Page 139: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/139.jpg)
Predicted and real values for first component
![Page 140: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/140.jpg)
Predicted and real values for first component
![Page 141: The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y:](https://reader035.vdocuments.us/reader035/viewer/2022062521/56814a86550346895db79854/html5/thumbnails/141.jpg)
?Compare the CLS, ILS and PCR methods for prediction in a two components system with strong spectral overlapping