the general linear model (glm)

28
The General Linear Model (GLM) SPM Short Course, Wellcome Trust Centre for Neuroimaging October 2008 Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical Research in Economics University of Zurich Functional Imaging Laboratory (FIL) Wellcome Trust Centre for Neuroimaging University College London

Upload: aimee-terrell

Post on 03-Jan-2016

48 views

Category:

Documents


8 download

DESCRIPTION

The General Linear Model (GLM). Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical Research in Economics University of Zurich Functional Imaging Laboratory (FIL) Wellcome Trust Centre for Neuroimaging University College London. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: The General Linear Model (GLM)

The General Linear Model (GLM)

SPM Short Course, Wellcome Trust Centre for NeuroimagingOctober 2008

Klaas Enno Stephan Branco Weiss Laboratory (BWL)Institute for Empirical Research in EconomicsUniversity of Zurich

Functional Imaging Laboratory (FIL)Wellcome Trust Centre for NeuroimagingUniversity College London

Page 2: The General Linear Model (GLM)

Overview of SPM

RealignmentRealignment SmoothingSmoothing

NormalisationNormalisation

General linear modelGeneral linear model

Statistical parametric map (SPM)Statistical parametric map (SPM)Image time-seriesImage time-series

Parameter estimatesParameter estimates

Design matrixDesign matrix

TemplateTemplate

KernelKernel

Gaussian Gaussian field theoryfield theory

p <0.05p <0.05

StatisticalStatisticalinferenceinference

Page 3: The General Linear Model (GLM)

Passive word listeningversus rest7 cycles of rest and listeningBlocks of 6 scanswith 7 sec TR

Question: Is there a change in the BOLD response between listening and rest?

Stimulus function

One session

A very simple fMRI experiment

Page 4: The General Linear Model (GLM)

stimulus function

1. Decompose data into effects and error

2. Form statistic using estimates of effects and error

Make inferences about effects of interest

Why?

How?

datalinearmodel

effects estimate

error estimate

statistic

Modelling the measured data

Page 5: The General Linear Model (GLM)

Time

BOLD signalTim

esingle voxel

time series

single voxel

time series

Voxel-wise time series analysis

modelspecificati

on

modelspecificati

onparameterestimationparameterestimation

hypothesishypothesis

statisticstatistic

SPMSPM

Page 6: The General Linear Model (GLM)

BOLD signal

Tim

e =1 2+ +

err

or

x1 x2 e

Single voxel regression model

exxy 2211 exxy 2211

Page 7: The General Linear Model (GLM)

Mass-univariate analysis: voxel-wise GLM

=

e+yy XX

N

1

N N

1 1p

p

Model is specified by1. Design matrix X2. Assumptions about

e

Model is specified by1. Design matrix X2. Assumptions about

e

N: number of scansp: number of regressors

N: number of scansp: number of regressors

eXy eXy

The design matrix embodies all available knowledge about experimentally controlled factors and potential confounds.

),0(~ 2INe ),0(~ 2INe

Page 8: The General Linear Model (GLM)

• parametric• one sample t-test

• two sample t-test

• paired t-test

• Anova

• AnCova

• correlation

• linear regression

• multiple regression

• F-tests

• etc…

• non-parametric? SnPM

all cases of the

General Linear Model

GLM: mass-univariate parametric analysis

Page 9: The General Linear Model (GLM)

GLM assumes Gaussian “spherical” (i.i.d.) errors

sphericity = iid:error covariance is scalar multiple of identity matrix:Cov(e) = 2I

sphericity = iid:error covariance is scalar multiple of identity matrix:Cov(e) = 2I

10

01)(eCov

10

04)(eCov

21

12)(eCov

Examples for non-sphericity:

non-identity

non-identitynon-independence

Page 10: The General Linear Model (GLM)

Parameter estimation

eXy

= +

e

2

1

Ordinary least squares estimation

(OLS) (assuming i.i.d. error):

Ordinary least squares estimation

(OLS) (assuming i.i.d. error):

yXXX TT 1)(ˆ

Objective:estimate parameters to minimize

N

tte

1

2

y X

Page 11: The General Linear Model (GLM)

y

e

Design space defined by X

x1

x2

A geometric perspective on the GLM

PIR

Rye

PIR

Rye

ˆ Xy

yXXX TT 1)(ˆ yXXX TT 1)(ˆ

TT XXXXP

Pyy1)(

ˆ

TT XXXXP

Pyy1)(

ˆ

Residual forming matrix

R

Projection matrix P

OLS estimates

Page 12: The General Linear Model (GLM)

x1

x2x2*

y

Correlated and orthogonal regressors

When x2 is orthogonalized with regard to x1, only the parameter estimate for x1 changes, not that for x2!

Correlated regressors = explained variance is shared between regressors

121

2211

exxy

121

2211

exxy

1;1 *21

*2

*211

exxy

1;1 *21

*2

*211

exxy

Page 13: The General Linear Model (GLM)

What are the problems of this model?

1. BOLD responses have a delayed and dispersed form. HRF

2. The BOLD signal includes substantial amounts of low-frequency noise.

3. The data are serially correlated (temporally autocorrelated) this violates the assumptions of the noise model in the GLM

Page 14: The General Linear Model (GLM)

t

dtgftgf0

)()()(

The response of a linear time-invariant (LTI) system is the convolution of the input with the system's response to an impulse (delta function).

Problem 1: Shape of BOLD responseSolution: Convolution model

hemodynamic response function (HRF)

expected BOLD response = input function impulse response function (HRF)

Page 15: The General Linear Model (GLM)

Convolution model of the BOLD response

Convolve stimulus function with a canonical hemodynamic response function (HRF):

HRF

t

dtgftgf0

)()()(

Page 16: The General Linear Model (GLM)

Problem 2: Low-frequency noise Solution: High pass filtering

SeSXSy

discrete cosine transform (DCT)

set

discrete cosine transform (DCT)

set

S = residual forming matrix of DCT set

Page 17: The General Linear Model (GLM)

High pass filtering: example

blue = data

black = mean + low-frequency drift

green = predicted response, taking into account low-frequency drift

red = predicted response, NOT taking into account low-frequency drift

Page 18: The General Linear Model (GLM)

withwithttt aee 1 ),0(~ 2 Nt

1st order autoregressive process: AR(1)

)(eCovautocovariance

function

N

N

Problem 3: Serial correlations

Page 19: The General Linear Model (GLM)

Dealing with serial correlations

• Pre-colouring: impose some known autocorrelation structure on the data (filtering with matrix W) and use Satterthwaite correction for df’s.

• Pre-whitening:

1. Use an enhanced noise model with multiple error covariance components.

2. Use estimated autocorrelation to specify filter matrix W for whitening the data.

WeWXWy WeWXWy

Page 20: The General Linear Model (GLM)

How do we define V ?

• Enhanced noise model

• Remember linear transform for Gaussians

• Choose W such that error covariance becomes spherical

• Conclusion: W is a function of V so how do we estimate V ?

WeWXWy WeWXWy

),0(~ 2VNe ),0(~ 2VNe

),(~

),,(~22

2

aaNy

axyNx

),(~

),,(~22

2

aaNy

axyNx

2/1

2

22 ),0(~

VW

IVW

VWNWe

2/1

2

22 ),0(~

VW

IVW

VWNWe

Page 21: The General Linear Model (GLM)

Multiple covariance components

),0(~ 2VNe ),0(~ 2VNe

iiQV

eCovV

)(

iiQV

eCovV

)(

= 1 + 2

Q1 Q2

Estimation of hyperparameters with ReML (restricted maximum likelihood).

V

enhanced noise model error covariance components Qand hyperparameters

Page 22: The General Linear Model (GLM)

Contrasts &statistical parametric

maps

Q: activation during listening ?

Q: activation during listening ?

c = 1 0 0 0 0 0 0 0 0 0 0

Null hypothesis:Null hypothesis: 01

)ˆ(

ˆ

T

T

cStd

ct

)ˆ(

ˆ

T

T

cStd

ct

X

Page 23: The General Linear Model (GLM)

WeWXWy

c = 1 0 0 0 0 0 0 0 0 0 0c = 1 0 0 0 0 0 0 0 0 0 0

)ˆ(ˆ

ˆ

T

T

cdts

ct

cWXWXc

cdtsTT

T

)()(ˆ

)ˆ(ˆ

2

)(

ˆˆ

2

2

Rtr

WXWy

ReML-estimates

ReML-estimates

WyWX )(

)(2

2/1

eCovV

VW

)(WXWXIRX

t-statistic based on ML estimates

iiQ

V

TT XWXXWX 1)()( TT XWXXWX 1)()( For brevity:

Page 24: The General Linear Model (GLM)

• head movements

• arterial pulsations

• breathing

• eye blinks

• adaptation affects, fatigue, fluctuations in concentration, etc.

Physiological confounds

Page 25: The General Linear Model (GLM)

Outlook: further challenges

• correction for multiple comparisons

• variability in the HRF across voxels

• slice timing

• limitations of frequentist statistics Bayesian analyses

• GLM ignores interactions among voxels models of effective connectivity

These issues are discussed in future lectures.

Page 26: The General Linear Model (GLM)

Correction for multiple comparisons

• Mass-univariate approach: We apply the GLM to each of a huge number of voxels (usually > 100,000).

• Threshold of p<0.05 more than 5000 voxels significant by chance!

• Massive problem with multiple comparisons!

• Solution: Gaussian random field theory

Page 27: The General Linear Model (GLM)

Variability in the HRF

• HRF varies substantially across voxels and subjects

• For example, latency can differ by ± 1 second

• Solution: use multiple basis functions

• See talk on event-related fMRI

Page 28: The General Linear Model (GLM)

Summary

• Mass-univariate approach: same GLM for each voxel

• GLM includes all known experimental effects and confounds

• Convolution with a canonical HRF

• High-pass filtering to account for low-frequency drifts

• Estimation of multiple variance components (e.g. to account for serial correlations)

• Parametric statistics