multivariate statistics [1em]principal component analysis...

41
Multivariate Statistics Principal Component Analysis (PCA) 9.10.2013

Upload: others

Post on 26-Oct-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Multivariate Statistics

Principal Component Analysis (PCA)

9.10.2013

Page 2: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Goals of Today’s Lecture

I Get familiar with the multivariate counterparts of theexpectation and the variance.

I See how principal component analysis (PCA) can be used as adimension reduction technique.

1 / 40

Page 3: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Introduction

I In your introductory course you started with univariatestatistics. You had a look at one random variable X at atime. E.g., X = “measurement of temperature”.

I A random variable can be characterized by its expectation µand the variance σ2 (or standard deviation σ).

µ = E [X ], σ2 = Var(X ) = E [(X − µ)2].

I A model that is often used is the normal distribution:X ∼ N (µ, σ2).

I The normal distribution is fully characterized by theexpectation and the variance.

2 / 40

Page 4: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I The unknown parameters µ and σ can be estimated fromdata.

I Say we observe n (independent) realizations of our randomvariable X : x1, . . . , xn.

I You can think of measuring n times a certain quantity, e.g.temperature.

I Usual parameter estimates are

µ̂ = x =1

n

n∑i=1

xi , σ̂2 = V̂ar(X ) =1

n − 1

n∑i=1

(xi − x)2.

3 / 40

Page 5: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Multivariate Data

I Now we are going to have a look at the situation where wemeasure multiple things simultaneously.

I Hence, we have a multivariate random variable (vector) Xhaving m components: X ∈ Rm.

I You can think of measuring temperature at two differentlocations or measuring temperature and pressure at onelocation (m = 2).

I In that case

X =

[X (1)

X (2)

],

where X (1) is temperature and X (2) is pressure.

4 / 40

Page 6: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I A possible data-set now consists of n vectors of dimension 2(or m in the general case):

x1, . . . , xn,

where x i ∈ R2 (or Rm).

Remark

I In multiple linear regression we already had multiple variablesper observation.

I There, we had one response variable and many predictorvariables.

I Here, the situation is more general in the sense that we don’thave a response variable but we want to model “relationships”between (any) variables.

5 / 40

Page 7: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Expectation and Covariance Matrix

I We need some new quantities to model/describe this kind ofdata.

I We are therefore looking for the multivariate counterparts ofthe expectation and the variance.

I The (multivariate) expectation of X is defined as

E [X ] = µ = (µ1, . . . , µm)T = (E [X (1)], . . . ,E [X (m)])T .

I It’s nothing else than the collection of the univariateexpectations.

6 / 40

Page 8: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

What about dependency?

● ●

●●

●●

●●

●●

●●

●●

●●

● ●

18 19 20 21 22 23

1819

2021

22

Temperature A

Tem

pera

ture

B●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

18 19 20 21 22

1819

2021

22

Temperature A

Tem

pera

ture

B

●●

●●

●●

● ●

●●

●●

●●

●●

●●

17 18 19 20 21 22 23

1718

1920

2122

23

Temperature A

Tem

pera

ture

B

● ●

●●

● ●

●●

●●●

●●

●●

●●

●●

● ●●

17 18 19 20 21 22

1718

1920

2122

Temperature A

Tem

pera

ture

B

7 / 40

Page 9: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I We need a measure to characterize the dependency betweenthe different components.

I The simplest thing one can think of is linear dependencybetween two components.

I The corresponding measure is the correlation ρ.

I ρ is dimensionless and it always holds that

−1 ≤ ρ ≤ 1

I |ρ| measures the strength of the linear relationship.

I The sign of ρ indicates the direction of the linear relationship.

8 / 40

Page 10: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I The formal definition of the correlation is based on thecovariance.

I The covariance is an unstandardized version of thecorrelation. It is defined as

Cov(X (j),X (k)) = E [(X (j) − µj)(X (k) − µk)].

I The correlation between X (j) and X (k) is then

ρjk = Corr(X (j),X (k)) =Cov(X (j),X (k))√

Var(X (j)) Var(X (k))

I You have seen the empirical version in the introductory course.

9 / 40

Page 11: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I The covariance matrix |Σ is an m ×m matrix with elements

|Σjk = Cov(X (j),X (k)) = E [(X (j) − µj)(X (k) − µk)].

I We also write Var(X ) or Cov(X ) instead of |Σ.

I The special symbol |Σ is used in order to avoid confusion withthe sum sign

∑.

10 / 40

Page 12: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I The covariance matrix contains a lot of information, e.g.

|Σjj = Var(X (j)).

This means that the diagonal consists of the individualvariances.

I We can also compute the correlations via

Corr(X (j),X (k)) =Cov(X (j),X (k))√

Var(X (j)) Var(X (k))=

|Σjk√|Σjj |Σkk

.

11 / 40

Page 13: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I Again, from a real data-set we can estimate these quantitieswith

µ̂ =[x (1), x (2), . . . , x (m)

]T|̂Σjk =

1

n − 1

n∑i=1

(x(j)i − µ̂j)(x

(k)i − µ̂k),

I Or more directly the whole matrix

|̂Σ =1

n − 1

n∑i=1

(x i − µ̂)(x i − µ̂)T .

I Remember: (Empirical) correlation only measures strength oflinear relationship between two variables.

12 / 40

Page 14: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Illustration: Empirical Correlation

Source: Wikipedia

13 / 40

Page 15: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Linear Transformations

The following table illustrates how the expectation and thevariance (covariance matrix) change when linear transformationsare applied to the univariate random variable X or the multivariaterandom vector X .

Univariate Multivariate

Y = a + bX Y = a + BX

E [Y ] = a + bE [X ] E [Y ] = a + BE [X ]

Var(Y ) = b2 Var(X ) Var(Y ) = B |ΣXBT

where a, b ∈ R and a ∈ Rm,B ∈ Rm×m.

14 / 40

Page 16: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Principal Component Analysis (PCA)

Goal: Dimensionality reduction.

I We have m different dimensions (variables) but we would liketo find “a few specific dimensions of the data that containmost variation”.

I If two specific dimensions of the data-set contain mostvariation, visualizations will be easy (plot these two!).

I Such a plot then can be used to check for any “structure”.

15 / 40

Page 17: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Illustration of Artificial 3-Dim Data-Set

−1.5 −1.0 −0.5 0.0 0.5 1.0 1.5

−1.

5−

1.0

−0.

5 0

.0 0

.5 1

.0 1

.5

−1.5

−1.0

−0.5

0.0

0.5

1.0

1.5

x(1)

x(2)x(3

)

●●

●●

●●

●●

●●

16 / 40

Page 18: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I We have to be more precise with what we mean with“variation”.

I We define variation in the data as the sum of all individualempirical variances

m∑j=1

V̂ar(X (j)) =m∑j=1

σ̂2(X (j)).

I How can we now find these “few dimensions”?

I We are looking for a new coordinate system with basisvectors b1, . . . , bm ∈ Rm.

17 / 40

Page 19: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I Of course, our data-points x i ∈ Rm will then have newcoordinates

z(k)i = xTi bk , k = 1, . . . ,m

(= projection on new basis vectors).

I How should we choose the new basis?

I The first basis vector b1 should be chosen such that V̂ar(Z (1))is maximal (i.e. it captures most variation).

I The second basis vector b2 should be orthogonal to the first

one (bT2 b1 = 0) such that V̂ar(Z (2)) is maximized.

I And so on...

18 / 40

Page 20: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I The new basis vectors are the so-called principalcomponents.

I The individual components of these basis vectors are calledloadings. The loadings tell us how to interpret the newcoordinate system (i.e., how the old variables are weighted toget the new ones).

I The coordinates with respect to the new basis vectors (thetransformed variable values) are the so-called scores .

19 / 40

Page 21: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

PCA: Illustration in Two Dimensions

log

(w

idth

)

0.65 0.75

0.5

0.6

10

1. Principal Component

2. P

rinci

pal C

ompo

nent

−0.1 0 0.1−

0.1

00.

1

20 / 40

Page 22: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I We could find the first basis vector b1 by solving the followingmaximization problem

maxb:‖b‖=1

V̂ar(Xb),

where X is the matrix that has different observations indifferent rows and different variables in different columns (likethe design matrix in regression).

I It can be shown that b1 is the (standardized) eigenvector of

|̂ΣX that corresponds to the largest eigenvalue.

I Similarly for the other vectors b2, . . . , bm.

21 / 40

Page 23: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I To summarize: We are performing a transformation to newvariables

z i = BT (x i − µ̂),

where the transformation matrix B is orthogonal and containsthe bk ’s as columns.

I In general we also subtract the mean vector to ensure that allcomponents have mean 0.

I B is the matrix of (standardized) eigenvectors corresponding

to the eigenvalues λk of |̂ΣX (in decreasing order).

22 / 40

Page 24: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I Hence, we have

V̂ar(Z ) = |̂ΣZ = BT |̂ΣXB =

λ1 0 . . . 00 λ2 . . . 0...

.... . .

0 0 . . . λm

λ1 ≥ λ2 ≥ · · · ≥ λm ≥ 0.

I Hence, the variance of the different components of Z is givenby the corresponding eigenvalue (values on the diagonal).

I Moreover, the different components are uncorrelated(because the off-diagonal elements of the covariance matrix ofZ are all zero).

23 / 40

Page 25: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Scaling Issues

I The variance is not invariant under rescaling.

I If we change the units of a variable, that will change thevariance (e.g. when measuring a length in [m] instead of[mm]).

I Therefore, if variables are measured on very different scales,they should first be standardized to comparable units.

I This can be done by standardizing each variable to variance 1.

I Otherwise, PCA can be misleading.

24 / 40

Page 26: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

PCA and Dimensionality Reduction

I At the beginning we were talking about dimensionalityreduction.

I We can achieve this by simply looking at the first p < mprincipal components (and ignoring the remainingcomponents).

I The proportion of the variance that is explained by the first pprincipal components is ∑p

j=1 λj∑mj=1 λj

.

25 / 40

Page 27: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

I Sometimes we see a sharp drop when plotting the eigenvalues(in decreasing order).

Comp.1 Comp.3 Comp.5 Comp.7 Comp.9

NIR−spectra without 5 outliers

Var

ianc

es0.

0e+

002.

0e−

054.

0e−

05 Variances of princ. components

I This plot is also known as the scree-plot.

26 / 40

Page 28: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

PCA and Dimensionality Reduction

I For visualization of our data, we can for example use ascatterplot of the first two principal components.

I It should show “most of the variation”.

27 / 40

Page 29: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Illustration

−1.5 −1.0 −0.5 0.0 0.5 1.0 1.5

−1.

5−

1.0

−0.

5 0

.0 0

.5 1

.0 1

.5

−1.5

−1.0

−0.5

0.0

0.5

1.0

1.5

x(1)

x(2)x(3

)

●●

●●

●●

●●

●●

28 / 40

Page 30: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

PC1

−1.0 −0.5 0.0 0.5 1.0

●●

●● ●

● −1.

0−

0.5

0.0

0.5

1.0

●●

●●●

−1.

0−

0.5

0.0

0.5

1.0

●●

●●

●●

●●

●●

● ●

●●

●●

●●

● ●●

PC2●

●●

●●

●●

● ●

●●

●●

●●

●●

●●●

−1.0 −0.5 0.0 0.5 1.0

●●

●● ●

●●

●●●●

●● ●

●●

●●

●● ● ●● ●●●

●●●

● ●

●●

●●

● ●

●● ●

●●

●● ●●

●● ●

●●

●●

● ● ●●●● ●●

● ●●

● ●

●●

●●

−1.0 −0.5 0.0 0.5 1.0

−1.

0−

0.5

0.0

0.5

1.0

PC3

29 / 40

Page 31: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Loadings

PC1 PC2 PC3

x1 -0.395118451 0.06887762 0.91604437

x2 -0.009993467 0.99680385 -0.07926044

x3 -0.918575822 -0.04047172 -0.39316727

ScreeplotV

aria

nces

0.0

0.1

0.2

0.3

0.4

0.5

30 / 40

Page 32: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Applying PCA to NIR-Spectra

I An NIR-spectrum can be thought of as a multivariateobservation (the different variables are the measurements atdifferent wavelengths).

I A spectrum has the property that the different variables are“ordered” and we can plot one observation as a “function” (seeplot on next slide).

I If we apply PCA to this kind of data, the individualcomponents of the bk ’s (the so called loadings) can again beplotted as spectra.

I As an example we have a look at spectra measured atdifferent time-points of a chemical reaction.

31 / 40

Page 33: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Illustration: Spectra (centered at each wavelength)

1200 1400 1600 1800 2000 2200 2400

−0.

4−

0.2

0.0

0.2

0.4

Wavelength

Cen

tere

d S

pect

ra

One observation is one spectrum, i.e. a whole “function”.32 / 40

Page 34: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

1200 1400 1600 1800 2000 2200 2400

−0.

20.

00.

2

1st Principal Component

Wavelength

Load

ing

1200 1400 1600 1800 2000 2200 2400

−0.

20.

00.

2

2nd Principal Component

Wavelength

Load

ing

1200 1400 1600 1800 2000 2200 2400

−0.

20.

00.

2

3rd Principal Component

Wavelength

Load

ing

33 / 40

Page 35: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Scatterplot of the First Two Principal Components

● ●●●●●●●●●●●●

●●●●●●●

●●●●●●●●

●●●●●●●●

●●●●●●●●●●●●

●●●●●●●●●●●●●●●●●●●●●●●●●●●

●●●●●●

●●

●●

●●●●●

●●●●●● ●●●●●

●●

−2 0 2 4 6 8

−1.

0−

0.5

0.0

0.5

1.0

1.5

1. PCA

2. P

CA

34 / 40

Page 36: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Scree-Plot

Var

ianc

es

02

46

810

1214

35 / 40

Page 37: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Alternative Interpretations

If we restrict ourselves to the first p < m principal components wehave

x i − µ̂ = x̂ i + e i

where

x̂ i =

p∑k=1

z(k)i b(k), e i =

m∑k=p+1

z(k)i b(k).

36 / 40

Page 38: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Linear Algebra

It can be shown that the data matrix consisting of the x̂ i is thebest approximation of our original (centered) data matrix ifwe restrict ourselves to matrices of rank p (with respect to theFrobenius norm), i.e. it has smallest∑

i ,j

(e(j)i

)2.

Statistics

It’s the best approximation in the sense that it has the smallestsum of variances

m∑j=1

V̂ar(E (j)).

37 / 40

Page 39: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

PCA via Singular Value Decomposition (SVD)

I We can also get the principal components from the singularvalue decomposition (SVD) of the data matrix X.

I For that reason we require X to have centered columns! I Why does this work?

SVD of X yields the decomposition

X = UDVT

whereI U is n × n and orthogonal

I D is n ×m and generalized diagonal (containing the so-calledsingular values in descending order)

I V is m ×m and orthogonal

38 / 40

Page 40: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Properties of SVD

I The (standardized) eigenvectors of XTX make up the columnsof V.

I The singular values are the square roots of the eigenvalues ofXTX.

But what is XTX? If the columns of X are centered, this is the

rescaled (empirical) covariance matrix |̂ΣX , because

(XTX)jk =n∑

i=1

(x ixTi )jk =

n∑i=1

x(j)i x

(k)i = (n − 1) |̂Σjk .

Hence, the singular value decomposition of the centereddata-matrix automatically gives us the principal components (inV).

The data-matrix in new coordinates is given in UD.

39 / 40

Page 41: Multivariate Statistics [1em]Principal Component Analysis (PCA)stat.ethz.ch/~meier/teaching/cheming/2013/4_multivariate.pdf · Multivariate Statistics Principal Component Analysis

Summary

I PCA is a useful tool for dimension reduction.

I New basis system is given by (standardized) eigenvectors ofcovariance matrix.

I Eigenvalues are the variances of the new coordinates.

I In the case of spectra, the loadings can again be plotted asspectra.

40 / 40