statistical analysis rik henson with thanks to: karl friston, andrew holmes, stefan kiebel, will...

82
Statistical Analysis Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny

Upload: hollie-chandler

Post on 11-Jan-2016

232 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Statistical AnalysisStatistical Analysis

Rik Henson

With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny

Statistical AnalysisStatistical Analysis

Rik Henson

With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny

Page 2: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

Motioncorrection

Smoothing

kernel

Spatialnormalisation

Standardtemplate

fMRI time-series Statistical Parametric Map

General Linear Model

Design matrix

Parameter Estimates

Page 3: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Some TerminologySome TerminologySome TerminologySome Terminology

• SPM (“Statistical Parametric Mapping”) is a massively univariate approach - meaning that a statistic (e.g., T-value) is calculated for every voxel - using the “General Linear Model”

• Experimental manipulations are specified in a model (“design matrix”) which is fit to each voxel to estimate the size of the experimental effects (“parameter estimates”) in that voxel…

• … on which one or more hypotheses (“contrasts”) are tested to make statistical inferences (“p-values”), correcting for multiple comparisons across voxels (using “Gaussian Field Theory”)

• The parametric statistics assume continuous-valued data and additive noise that conforms to a “normal” distribution (“nonparametric” versions of SPM eschew such assumptions)

• SPM (“Statistical Parametric Mapping”) is a massively univariate approach - meaning that a statistic (e.g., T-value) is calculated for every voxel - using the “General Linear Model”

• Experimental manipulations are specified in a model (“design matrix”) which is fit to each voxel to estimate the size of the experimental effects (“parameter estimates”) in that voxel…

• … on which one or more hypotheses (“contrasts”) are tested to make statistical inferences (“p-values”), correcting for multiple comparisons across voxels (using “Gaussian Field Theory”)

• The parametric statistics assume continuous-valued data and additive noise that conforms to a “normal” distribution (“nonparametric” versions of SPM eschew such assumptions)

Page 4: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Some TerminologySome TerminologySome TerminologySome Terminology

• SPM usually focused on “functional specialisation” - i.e. localising different functions to different regions in the brain

• One might also be interested in “functional integration” - how different regions (voxels) interact

• Multivariate approaches work on whole images and can identify spatial/temporal patterns over voxels, without necessarily specifying a design matrix (PCA, ICA)...

• … or with an experimental design matrix (PLS, CVA), or with an explicit anatomical model of connectivity between regions - “effective connectivity” - eg using Dynamic Causal Modelling

• SPM usually focused on “functional specialisation” - i.e. localising different functions to different regions in the brain

• One might also be interested in “functional integration” - how different regions (voxels) interact

• Multivariate approaches work on whole images and can identify spatial/temporal patterns over voxels, without necessarily specifying a design matrix (PCA, ICA)...

• … or with an experimental design matrix (PLS, CVA), or with an explicit anatomical model of connectivity between regions - “effective connectivity” - eg using Dynamic Causal Modelling

Page 5: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 6: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 7: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

General Linear Model…General Linear Model…General Linear Model…General Linear Model…

• Parametric statisticsParametric statistics

• one sample one sample tt-test-test

• two sample two sample tt-test-test

• paired paired tt-test-test

• AnovaAnova

• AnCovaAnCova

• correlationcorrelation

• linear regressionlinear regression

• multiple regressionmultiple regression

• FF-tests-tests

• etc…etc…

all cases of the

General Linear Model

Page 8: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

General Linear ModelGeneral Linear ModelGeneral Linear ModelGeneral Linear Model

• Equation for single (and all) voxels:

yj = xj1 1 + … + xjP P + j j ~ N(0,2)

yj : data for scan, j = 1…N

xjp : explanatory variables / covariates / regressors, p = 1…P

p : parameters / regression slopes / fixed effects

j : residual errors, independent & identically (normally) distributed

• Equivalent matrix form:

y = X +

X : “design matrix” / model

• Equation for single (and all) voxels:

yj = xj1 1 + … + xjP P + j j ~ N(0,2)

yj : data for scan, j = 1…N

xjp : explanatory variables / covariates / regressors, p = 1…P

p : parameters / regression slopes / fixed effects

j : residual errors, independent & identically (normally) distributed

• Equivalent matrix form:

y = X +

X : “design matrix” / model

Page 9: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Matrix FormulationMatrix FormulationMatrix FormulationMatrix Formulation

Equation for scan j

Simultaneous equations forscans 1..N(J)

…that can be solvedfor parameters 1..P(L)

Regressors

Sca

ns

Page 10: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

General Linear Model (Estimation)General Linear Model (Estimation)General Linear Model (Estimation)General Linear Model (Estimation)

• Residual errors and estimated error variance are:

r = y - Y 2 = rTr / df

where df are the degrees of freedom (assuming iid):

df = N - rank(X) (=N-P if X full rank)

( R = I - XX+ r = Ry df = trace(R) )

• Residual errors and estimated error variance are:

r = y - Y 2 = rTr / df

where df are the degrees of freedom (assuming iid):

df = N - rank(X) (=N-P if X full rank)

( R = I - XX+ r = Ry df = trace(R) )

^

^

^

• Estimate parameters from least squares fit to data, y:

= (XTX)-1XTy = X+y (OLS

estimates)

• Fitted response is:

Y = X

Page 11: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

General Linear Model (Inference)General Linear Model (Inference)General Linear Model (Inference)General Linear Model (Inference)

^• Specify contrast (hypothesis), c, a linear combination

of parameter estimates, cT

^^ ^ ^

• Calculate T-stastistic for that contrast:

T(N-p) = cT / var(cT) = cT / sqrt(2cT(XTX)-1c)

(c is a vector), or an F-statistic:

F(p-p0,N-p) = [(r0Tr0 – rTr) / (p-p0)] / [rTr / (N-P)]

where r0 and p0 are parameters of the reduced model specified by c (which is a matrix)

• Prob. of falsely rejecting Null hypothesis, H0: cT=0 (“p-value”)

c = [1 -1 0 0]T

c = [ 2 -1 -1 0 -1 2 -1 0 -1 -1 2 0]

F

Page 12: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Example PET experimentExample PET experimentExample PET experimentExample PET experiment

• 12 scans, 3 conditions (1-way ANOVA)12 scans, 3 conditions (1-way ANOVA)

yj = x1j 1 + x2j 2 + x3j 3 + x4j 4 + j

where (dummy) variables:where (dummy) variables:

x1j = [0,1] = condition A (first 4 scans) = [0,1] = condition A (first 4 scans)

x2j = [0,1] = condition B (second 4 scans) = [0,1] = condition B (second 4 scans)

x3j = [0,1] = condition C (third 4 scans) = [0,1] = condition C (third 4 scans)

x4j = [1] = grand mean = [1] = grand mean

• T-contrast : T-contrast : [1 -1 0 0] tests whether A>B[1 -1 0 0] tests whether A>B

[-1 1 0 0] tests whether B>A[-1 1 0 0] tests whether B>A

• F-contrast: F-contrast: [ 2 -1 -1 0[ 2 -1 -1 0 -1 2 -1 0 -1 2 -1 0 -1 -1 2 0] tests main effect of A,B,C -1 -1 2 0] tests main effect of A,B,C

• 12 scans, 3 conditions (1-way ANOVA)12 scans, 3 conditions (1-way ANOVA)

yj = x1j 1 + x2j 2 + x3j 3 + x4j 4 + j

where (dummy) variables:where (dummy) variables:

x1j = [0,1] = condition A (first 4 scans) = [0,1] = condition A (first 4 scans)

x2j = [0,1] = condition B (second 4 scans) = [0,1] = condition B (second 4 scans)

x3j = [0,1] = condition C (third 4 scans) = [0,1] = condition C (third 4 scans)

x4j = [1] = grand mean = [1] = grand mean

• T-contrast : T-contrast : [1 -1 0 0] tests whether A>B[1 -1 0 0] tests whether A>B

[-1 1 0 0] tests whether B>A[-1 1 0 0] tests whether B>A

• F-contrast: F-contrast: [ 2 -1 -1 0[ 2 -1 -1 0 -1 2 -1 0 -1 2 -1 0 -1 -1 2 0] tests main effect of A,B,C -1 -1 2 0] tests main effect of A,B,C

c=[-1 1 0 0], T=10/sqrt(3.3*8)df=12-3=9, T(9)=1.94, p<.05

1191282119221831293228

+

1-1 2-2 1-1 2-2 1-1 2-2

1 0 0 11 0 0 11 0 0 11 0 0 10 1 0 10 1 0 10 1 0 10 1 0 10 0 1 10 0 1 10 0 1 10 0 1 1

=

-10 0 10 20

rank

(X)=

3

Page 13: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Global EffectsGlobal EffectsGlobal EffectsGlobal Effects

• May be variation in PET tracer dose May be variation in PET tracer dose from scan to scanfrom scan to scan

• Such “global” changes in image Such “global” changes in image intensity (gCBF) confound local / intensity (gCBF) confound local / regional (rCBF) changes of experiment regional (rCBF) changes of experiment

• Adjust for global effects by:Adjust for global effects by:

- AnCova (Additive Model) - AnCova (Additive Model) - PET?- PET?

- Proportional Scaling - Proportional Scaling - fMRI?- fMRI?

• Can improve statistics when orthogonal Can improve statistics when orthogonal to effects of interest (as here)…to effects of interest (as here)…

• ……but can also worsen when effects of but can also worsen when effects of interest correlated with global (as next)interest correlated with global (as next)

• May be variation in PET tracer dose May be variation in PET tracer dose from scan to scanfrom scan to scan

• Such “global” changes in image Such “global” changes in image intensity (gCBF) confound local / intensity (gCBF) confound local / regional (rCBF) changes of experiment regional (rCBF) changes of experiment

• Adjust for global effects by:Adjust for global effects by:

- AnCova (Additive Model) - AnCova (Additive Model) - PET?- PET?

- Proportional Scaling - Proportional Scaling - fMRI?- fMRI?

• Can improve statistics when orthogonal Can improve statistics when orthogonal to effects of interest (as here)…to effects of interest (as here)…

• ……but can also worsen when effects of but can also worsen when effects of interest correlated with global (as next)interest correlated with global (as next)

rCB

F

x

o

o

o

o

o

o

x

x

x

x

x

gCBF

rCB

F

x

o

o

o

o

o

o

x

x

x

x

x

global

gCBF

rCB

F

x

o

o

o

o

o

o

x

x

x

x

x

g..

k1

k2

k

global

AnCova

globalgCBF

rCB

F

x

oo

o

oo

o

xx

xx

x

0 50

rC

BF

(a

dj)

o

0

xxxx

xx

ooooo

Scaling

Page 14: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Global Effects (AnCova)Global Effects (AnCova)Global Effects (AnCova)Global Effects (AnCova)

• 12 scans, 3 conditions, 1 confounding covariate12 scans, 3 conditions, 1 confounding covariate

yj = x1j 1 + x2j 2 + x3j 3 + x4j 4 + x5j 5 + j

where (dummy) variables:where (dummy) variables:

x1j = [0,1] = condition A (first 4 scans) = [0,1] = condition A (first 4 scans)

x2j = [0,1] = condition B (second 4 scans) = [0,1] = condition B (second 4 scans)

x3j = [0,1] = condition C (third 4 scans) = [0,1] = condition C (third 4 scans)

x4j = grand mean = grand mean

x5j = global signal (mean over all voxels) = global signal (mean over all voxels)

(further mean-corrected over all scans)(further mean-corrected over all scans)

• Global correlated here with conditions (and time)Global correlated here with conditions (and time)

• Global estimate can be scaled to, eg, 50ml/min/dlGlobal estimate can be scaled to, eg, 50ml/min/dl

• 12 scans, 3 conditions, 1 confounding covariate12 scans, 3 conditions, 1 confounding covariate

yj = x1j 1 + x2j 2 + x3j 3 + x4j 4 + x5j 5 + j

where (dummy) variables:where (dummy) variables:

x1j = [0,1] = condition A (first 4 scans) = [0,1] = condition A (first 4 scans)

x2j = [0,1] = condition B (second 4 scans) = [0,1] = condition B (second 4 scans)

x3j = [0,1] = condition C (third 4 scans) = [0,1] = condition C (third 4 scans)

x4j = grand mean = grand mean

x5j = global signal (mean over all voxels) = global signal (mean over all voxels)

(further mean-corrected over all scans)(further mean-corrected over all scans)

• Global correlated here with conditions (and time)Global correlated here with conditions (and time)

• Global estimate can be scaled to, eg, 50ml/min/dlGlobal estimate can be scaled to, eg, 50ml/min/dl

c=[-1 1 0 0], T=3.3/sqrt(3.8*8)df=12-4=8, T(8)=0.61, p>.05

1191282119221831293228

+

1-1 2-2 1-1 2-2 1-1 2-2

1 0 0 1 -11 0 0 1 -11 0 0 1 -11 0 0 1 -10 1 0 1 00 1 0 1 0 0 1 0 1 00 1 0 1 00 0 1 1 10 0 1 1 10 0 1 1 10 0 1 1 1

=

1.75.08.3 15 6.7

1 2 3 4 5

rank

(X)=

4

Page 15: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Global Effects (fMRI)Global Effects (fMRI)Global Effects (fMRI)Global Effects (fMRI)

• Two types of scaling: Two types of scaling: Grand Mean Grand Mean scaling and scaling and Global Global scalingscaling

• Grand Mean scaling is automatic, global scaling is optionalGrand Mean scaling is automatic, global scaling is optional

• Grand Mean scales by 100/mean over all voxels and ALL scans Grand Mean scales by 100/mean over all voxels and ALL scans (i.e, single number per session) (i.e, single number per session)

• Global scaling scales by 100/mean over all voxels for EACH scan Global scaling scales by 100/mean over all voxels for EACH scan (i.e, a different scaling factor every scan)(i.e, a different scaling factor every scan)

• Problem with Problem with globalglobal scaling is that TRUE global is not (normally) known… scaling is that TRUE global is not (normally) known…

• ……we only estimate it by the mean over voxelswe only estimate it by the mean over voxels

• So if there is a large signal change over many voxels, the global So if there is a large signal change over many voxels, the global estimateestimate will will be confounded by local changesbe confounded by local changes

• This can produce artifactual deactivations in other regions after global scalingThis can produce artifactual deactivations in other regions after global scaling

• Since most sources of global variability in fMRI are low frequency (drift), Since most sources of global variability in fMRI are low frequency (drift), high-pass filteringhigh-pass filtering may be sufficient, and many people to not use global scaling may be sufficient, and many people to not use global scaling

• Two types of scaling: Two types of scaling: Grand Mean Grand Mean scaling and scaling and Global Global scalingscaling

• Grand Mean scaling is automatic, global scaling is optionalGrand Mean scaling is automatic, global scaling is optional

• Grand Mean scales by 100/mean over all voxels and ALL scans Grand Mean scales by 100/mean over all voxels and ALL scans (i.e, single number per session) (i.e, single number per session)

• Global scaling scales by 100/mean over all voxels for EACH scan Global scaling scales by 100/mean over all voxels for EACH scan (i.e, a different scaling factor every scan)(i.e, a different scaling factor every scan)

• Problem with Problem with globalglobal scaling is that TRUE global is not (normally) known… scaling is that TRUE global is not (normally) known…

• ……we only estimate it by the mean over voxelswe only estimate it by the mean over voxels

• So if there is a large signal change over many voxels, the global So if there is a large signal change over many voxels, the global estimateestimate will will be confounded by local changesbe confounded by local changes

• This can produce artifactual deactivations in other regions after global scalingThis can produce artifactual deactivations in other regions after global scaling

• Since most sources of global variability in fMRI are low frequency (drift), Since most sources of global variability in fMRI are low frequency (drift), high-pass filteringhigh-pass filtering may be sufficient, and many people to not use global scaling may be sufficient, and many people to not use global scaling

Page 16: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimability

• If any column of X is a linear If any column of X is a linear combination of any others (X is combination of any others (X is rank rank deficientdeficient), some parameters cannot be ), some parameters cannot be estimated uniquely (estimated uniquely (inestimableinestimable))

• … … which means some contrasts cannot which means some contrasts cannot be tested (be tested (eg,eg, only if sum to zero) only if sum to zero)

• This has implications for whether This has implications for whether “baseline” (constant term) is explicitly “baseline” (constant term) is explicitly or implicitly modelledor implicitly modelled

• If any column of X is a linear If any column of X is a linear combination of any others (X is combination of any others (X is rank rank deficientdeficient), some parameters cannot be ), some parameters cannot be estimated uniquely (estimated uniquely (inestimableinestimable))

• … … which means some contrasts cannot which means some contrasts cannot be tested (be tested (eg,eg, only if sum to zero) only if sum to zero)

• This has implications for whether This has implications for whether “baseline” (constant term) is explicitly “baseline” (constant term) is explicitly or implicitly modelledor implicitly modelled

cm = [1 0 0]

cd = [1 -1 0]

1 = 1.6

2 = 0.7

1 = 0.9

2 = 0.7

cd* = [1 -1]*= 0.9

cd* = [1 0]*= 0.9

A B A+B

cm = [1 0]

cd = [1 -1]

cd = [1 0]

A B

“implicit”

A A+B

“explicit”

rank(X)=2

Page 17: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimability

• If any column of X is a linear If any column of X is a linear combination of any others (X is combination of any others (X is rank rank deficientdeficient), some parameters cannot be ), some parameters cannot be estimated uniquely (estimated uniquely (inestimableinestimable))

• … … which means some contrasts cannot which means some contrasts cannot be tested (eg, only if sum to zero)be tested (eg, only if sum to zero)

• This has implications for whether This has implications for whether “baseline” (constant term) is explicitly “baseline” (constant term) is explicitly or implicitly modelledor implicitly modelled

• (rank deficiency might be thought of as (rank deficiency might be thought of as perfect correlation…)perfect correlation…)

• If any column of X is a linear If any column of X is a linear combination of any others (X is combination of any others (X is rank rank deficientdeficient), some parameters cannot be ), some parameters cannot be estimated uniquely (estimated uniquely (inestimableinestimable))

• … … which means some contrasts cannot which means some contrasts cannot be tested (eg, only if sum to zero)be tested (eg, only if sum to zero)

• This has implications for whether This has implications for whether “baseline” (constant term) is explicitly “baseline” (constant term) is explicitly or implicitly modelledor implicitly modelled

• (rank deficiency might be thought of as (rank deficiency might be thought of as perfect correlation…)perfect correlation…)

cm = [1 0 0]

cd = [1 -1 0]

X(1) * T = X(2)

A B A+B

A B

“implicit”

A A+B

“explicit”

rank(X)=2

T = 1 1 0 1

c(1) * T = c(2)

[ 1 -1 ] * = [ 1 0 ] 1 1 0 1

Page 18: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimabilityA word on correlation/estimability

• When there is high (but not perfect) When there is high (but not perfect) correlation between regressors, correlation between regressors, parameters can be estimated…parameters can be estimated…

• ……but the estimates will be but the estimates will be inefficientinefficient estimated (ie highly variable)estimated (ie highly variable)

• ……meaning some contrasts will not lead meaning some contrasts will not lead to very powerful teststo very powerful tests

• When there is high (but not perfect) When there is high (but not perfect) correlation between regressors, correlation between regressors, parameters can be estimated…parameters can be estimated…

• ……but the estimates will be but the estimates will be inefficientinefficient estimated (ie highly variable)estimated (ie highly variable)

• ……meaning some contrasts will not lead meaning some contrasts will not lead to very powerful teststo very powerful tests

cm = [1 0 0]

cd = [1 -1 0]

A B A+B

cm = [1 0 0]

cd = [1 -1 0]

()

A B A+B

convolved with HRF!

• SPM shows pairwise correlation SPM shows pairwise correlation between regressors, but this will NOT between regressors, but this will NOT tell you that, eg, Xtell you that, eg, X11+X+X22 is highly is highly correlated with Xcorrelated with X33……

• … … so some contrasts can still be so some contrasts can still be inefficient, even though pairwise inefficient, even though pairwise correlations are lowcorrelations are low

Page 19: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A word on orthogonalisationA word on orthogonalisationA word on orthogonalisationA word on orthogonalisation

• To remove correlation between two regressors, To remove correlation between two regressors, you can explicitly orthogonalise one (Xyou can explicitly orthogonalise one (X11) with ) with respect to the other (Xrespect to the other (X22):):

XX11 = X = X11 – (X – (X22XX22

++)X)X11 (Gram-Schmidt)(Gram-Schmidt)

• Paradoxically, this will NOT change the Paradoxically, this will NOT change the parameter estimate for Xparameter estimate for X11, but will for X, but will for X22

• In other words, the parameter estimate for the In other words, the parameter estimate for the orthogonalised regressor is unchanged!orthogonalised regressor is unchanged!

• This reflects fact that parameter estimates This reflects fact that parameter estimates automatically reflect orthogonal component of automatically reflect orthogonal component of each regressor…each regressor…

• ……so no need to orthogonalise, UNLESS you so no need to orthogonalise, UNLESS you have a priori reason for assigning have a priori reason for assigning common common variancevariance to the other regressor to the other regressor

• To remove correlation between two regressors, To remove correlation between two regressors, you can explicitly orthogonalise one (Xyou can explicitly orthogonalise one (X11) with ) with respect to the other (Xrespect to the other (X22):):

XX11 = X = X11 – (X – (X22XX22

++)X)X11 (Gram-Schmidt)(Gram-Schmidt)

• Paradoxically, this will NOT change the Paradoxically, this will NOT change the parameter estimate for Xparameter estimate for X11, but will for X, but will for X22

• In other words, the parameter estimate for the In other words, the parameter estimate for the orthogonalised regressor is unchanged!orthogonalised regressor is unchanged!

• This reflects fact that parameter estimates This reflects fact that parameter estimates automatically reflect orthogonal component of automatically reflect orthogonal component of each regressor…each regressor…

• ……so no need to orthogonalise, UNLESS you so no need to orthogonalise, UNLESS you have a priori reason for assigning have a priori reason for assigning common common variancevariance to the other regressor to the other regressor

YY

XX22XX11

2

1

XX11

2

Page 20: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A word on orthogonalisationA word on orthogonalisationA word on orthogonalisationA word on orthogonalisation

1(M1) = 1.6

2(M1) = 0.7

1 = 0.9

2 = 0.7

X1 X2

1(M2) = 0.9

2 = 1.15

X1 X2

Orthogonalise X2 (Model M1)

X1 X2

Orthogonalise X1 (Model M2)

= 1– 2

= ( 1+ 2

T = 0.5 1 -0.5 1

Page 21: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 22: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

fMRI AnalysisfMRI AnalysisfMRI AnalysisfMRI Analysis

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

Page 23: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

fMRI AnalysisfMRI AnalysisfMRI AnalysisfMRI Analysis

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

Page 24: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

(Epoch) fMRI example…(Epoch) fMRI example…(Epoch) fMRI example…(Epoch) fMRI example…

box-car function

= 1 + (t)

voxel timeseries

2+

baseline (mean)

(box-car unconvolved)

Page 25: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

(Epoch) fMRI example…(Epoch) fMRI example…(Epoch) fMRI example…(Epoch) fMRI example…

y

data v

ecto

r

(v

oxel

time s

eries

)

=

= X

design

mat

rix

1

2

param

eters

+

+

erro

r vec

tor

Page 26: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

power spectrum

Low frequency noiseLow frequency noiseLow frequency noiseLow frequency noise

noise

signal(eg infinite 30s on-off)

• Low frequency noise:– Physical (scanner drifts)– Physiological (aliased)

• cardiac (~1 Hz)• respiratory (~0.25 Hz)

• Low frequency noise:– Physical (scanner drifts)– Physiological (aliased)

• cardiac (~1 Hz)• respiratory (~0.25 Hz)

aliasing

highpassfilter

power spectrum

Page 27: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

(Epoch) fMRI example…(Epoch) fMRI example…...with highpass filter...with highpass filter

(Epoch) fMRI example…(Epoch) fMRI example…...with highpass filter...with highpass filter

+

erro

r vec

tor

+

data v

ecto

r

y

=

design

mat

rix

= X

param

eters

Page 28: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

(Epoch) fMRI example…(Epoch) fMRI example……fitted and adjusted data…fitted and adjusted data(Epoch) fMRI example…(Epoch) fMRI example……fitted and adjusted data…fitted and adjusted data

Raw fMRI timeseries

Residuals highpass filtered (and scaled)

fitted high-pass filter

Adjusted data

fitted box-car

Page 29: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

fMRI AnalysisfMRI AnalysisfMRI AnalysisfMRI Analysis

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

Page 30: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Convolution with HRFConvolution with HRFConvolution with HRFConvolution with HRF

Boxcar function convolved with HRF

=

hæmodynamic response

Residuals Unconvolved fit

Convolved fit Residuals (less structure)

Page 31: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

fMRI AnalysisfMRI AnalysisfMRI AnalysisfMRI Analysis

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

• Scans are treated as a timeseries…

… and can be filtered to remove low-frequency (1/f) noise

• Effects of interest are convolved with haemodynamic (BOLD) response function (HRF), to capture sluggish nature of response

• Scans can no longer be treated as independent observations…

… they are typically temporally autocorrelated (for TRs<8s)

Page 32: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Temporal autocorrelation…Temporal autocorrelation…Temporal autocorrelation…Temporal autocorrelation…

• Because the data are typically correlated from one scan to the next, one cannot Because the data are typically correlated from one scan to the next, one cannot assume the degrees of freedom (dfs) are simply the number of scans minus the assume the degrees of freedom (dfs) are simply the number of scans minus the dfs used in the modeldfs used in the model – need “ – need “effective degrees of freedomeffective degrees of freedom””

• In other words, the residual errors are not independent:In other words, the residual errors are not independent:

YY = = XX + + ~ ~ NN(0,(0,22VV)) VV I, V=AAI, V=AA''

wwhere A is the intrinsic autocorrelationhere A is the intrinsic autocorrelation

• Generalised least squares:Generalised least squares:

KKYY = = KKXX + + KKKK ~ ~ NN(0, (0, 22VV) ) VV= = KAAKAA''KK''

(autocorrelation is a special case of (autocorrelation is a special case of “nonsphericity”“nonsphericity”…)…)

• Because the data are typically correlated from one scan to the next, one cannot Because the data are typically correlated from one scan to the next, one cannot assume the degrees of freedom (dfs) are simply the number of scans minus the assume the degrees of freedom (dfs) are simply the number of scans minus the dfs used in the modeldfs used in the model – need “ – need “effective degrees of freedomeffective degrees of freedom””

• In other words, the residual errors are not independent:In other words, the residual errors are not independent:

YY = = XX + + ~ ~ NN(0,(0,22VV)) VV I, V=AAI, V=AA''

wwhere A is the intrinsic autocorrelationhere A is the intrinsic autocorrelation

• Generalised least squares:Generalised least squares:

KKYY = = KKXX + + KKKK ~ ~ NN(0, (0, 22VV) ) VV= = KAAKAA''KK''

(autocorrelation is a special case of (autocorrelation is a special case of “nonsphericity”“nonsphericity”…)…)

Page 33: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Temporal autocorrelation (History)Temporal autocorrelation (History)Temporal autocorrelation (History)Temporal autocorrelation (History)

• One method is to estimate One method is to estimate AA, using, for example, an AR(p) model, then:, using, for example, an AR(p) model, then:

KK = = AA-1-1 VV= = II (allows OLS)(allows OLS)

This “pre-whitening” is sensitive, but can be biased if K mis-estimatedThis “pre-whitening” is sensitive, but can be biased if K mis-estimated

• Another method Another method (SPM99)(SPM99) is to smooth the data with a known autocorrelation is to smooth the data with a known autocorrelation that swamps any intrinsic autocorrelation:that swamps any intrinsic autocorrelation:

KK = = SS VV= SAA= SAA''SS’ ~ ’ ~ SSSS'' (use GLS)(use GLS)

Effective degrees of freedom calculated with Satterthwaite approximation ( Effective degrees of freedom calculated with Satterthwaite approximation ( df = trace(RV)df = trace(RV)22/trace(RVRV)/trace(RVRV) ) )

This is more robust (providing the temporal smoothing is sufficient, eg 4s This is more robust (providing the temporal smoothing is sufficient, eg 4s FWHM Gaussian), but less sensitiveFWHM Gaussian), but less sensitive

• Most recent method Most recent method (SPM2)(SPM2) is to restrict K to highpass filter, and estimate is to restrict K to highpass filter, and estimate

residual autocorrelation residual autocorrelation AA using voxel-wide, one-step ReML… using voxel-wide, one-step ReML…

• One method is to estimate One method is to estimate AA, using, for example, an AR(p) model, then:, using, for example, an AR(p) model, then:

KK = = AA-1-1 VV= = II (allows OLS)(allows OLS)

This “pre-whitening” is sensitive, but can be biased if K mis-estimatedThis “pre-whitening” is sensitive, but can be biased if K mis-estimated

• Another method Another method (SPM99)(SPM99) is to smooth the data with a known autocorrelation is to smooth the data with a known autocorrelation that swamps any intrinsic autocorrelation:that swamps any intrinsic autocorrelation:

KK = = SS VV= SAA= SAA''SS’ ~ ’ ~ SSSS'' (use GLS)(use GLS)

Effective degrees of freedom calculated with Satterthwaite approximation ( Effective degrees of freedom calculated with Satterthwaite approximation ( df = trace(RV)df = trace(RV)22/trace(RVRV)/trace(RVRV) ) )

This is more robust (providing the temporal smoothing is sufficient, eg 4s This is more robust (providing the temporal smoothing is sufficient, eg 4s FWHM Gaussian), but less sensitiveFWHM Gaussian), but less sensitive

• Most recent method Most recent method (SPM2)(SPM2) is to restrict K to highpass filter, and estimate is to restrict K to highpass filter, and estimate

residual autocorrelation residual autocorrelation AA using voxel-wide, one-step ReML… using voxel-wide, one-step ReML…

KYKY = = KXKX + + KK KK ~ ~ NN(0, (0, 22VV) ) VV = = KAA'K'KAA'K'

Page 34: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)

• Nonsphericity means (kind of) that: Nonsphericity means (kind of) that:

CC = = cov(cov(22II

• Nonsphericity can be modelled by set Nonsphericity can be modelled by set of of variance components:variance components:

CC QQ11 + + QQ22 + + QQ3 3 ......

((iiare are hyper-parametershyper-parameters))

- Non-identical (inhomogeneous):- Non-identical (inhomogeneous):(e.g, two groups of subjects)(e.g, two groups of subjects)

1Q 2Q

Scans

Sca

ns

cov())spherical

- Non-independent (autocorrelated):- Non-independent (autocorrelated):(e.g, white noise + AR(1))(e.g, white noise + AR(1)) 1Q 2Q

New inSPM2

Page 35: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)

^

^

OLS = (XTX)-1XTy (= X+y)

ML = (XTC-1X)-1XTC

-1y

C = ReML( yyT, X, Q )

V = ReML( yjyjT, X,

Q )

• Joint estimation of parameters and hyper-Joint estimation of parameters and hyper-parameters requires parameters requires ReMLReML

• ReML gives (Restricted) ReML gives (Restricted) Maximum LikelihoodMaximum Likelihood (ML)(ML) estimates of (hyper)parameters, rather estimates of (hyper)parameters, rather than than Ordinary Least Square (OLS)Ordinary Least Square (OLS) estimates estimates

• ML estimates are more efficient, entail exact ML estimates are more efficient, entail exact dfs (no Satterthwaite approx)…dfs (no Satterthwaite approx)…

• ……but computationally expensive: ReML is but computationally expensive: ReML is iterative (unless only one hyper-parameter)iterative (unless only one hyper-parameter)

• To speed up:To speed up:– Correlation of errors (V) estimated by Correlation of errors (V) estimated by poolingpooling

over voxelsover voxels– Covariance of errors (Covariance of errors (22V) V) estimated by estimated by

single, voxel-specific scaling hyperparameter single, voxel-specific scaling hyperparameter

New inSPM2

2211ˆˆ QQ

T

voxel

yy

Page 36: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)Nonsphericity and ReML (SPM2)

1.1. Voxels to be pooled collected by first-pass Voxels to be pooled collected by first-pass through data (OLS)through data (OLS)

(biased if correlation structure (biased if correlation structure not stationary across voxels?)not stationary across voxels?)

2.2. Correlation structure V estimated iteratively Correlation structure V estimated iteratively using ReML once, pooling over all voxelsusing ReML once, pooling over all voxels

3.3. Remaining hyper-parameter estimated using Remaining hyper-parameter estimated using V and ReML noniteratively, for each voxelV and ReML noniteratively, for each voxel

• Estimation of nonsphericity is used to pre-Estimation of nonsphericity is used to pre-whiten the data and design matrix,whiten the data and design matrix, W=V W=V-1/2 -1/2 (or (or by KW, if highpass filter K present) by KW, if highpass filter K present)

• (which is why design matrices in SPM2 can (which is why design matrices in SPM2 can differ from those in SPM99 after estimation)differ from those in SPM99 after estimation)

B

New inSPM2

X

W

WX

Page 37: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Xy

c = +1 0 0 0 0 0 0 0 0 0 0c = +1 0 0 0 0 0 0 0 0 0 0

)ˆ(ˆ

ˆ

T

T

cdtS

ct

)ˆ(ˆ

ˆ

T

T

cdtS

ct

cWXWXccdtSTTT )()(ˆ)ˆ(ˆ 2

V

X

)(

ˆˆ

2

2

Rtrace

WXWY

ReML-estimation

ReML-estimation

WyWX )()cov(2

2/1

V

VW

)(WXWXIR

The Full-Monty T-test (SPM2)The Full-Monty T-test (SPM2)The Full-Monty T-test (SPM2)The Full-Monty T-test (SPM2)New inSPM2

Page 38: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 39: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

• If If n=100,000n=100,000 voxels tested with voxels tested with ppuu=0.05=0.05 of falsely rejecting of falsely rejecting HHoo... ...

… …then approx then approx n n p puu (eg 5,000) (eg 5,000)

will do so by chance (false will do so by chance (false positives, or “type I” errors)positives, or “type I” errors)

• Therefore need to “correct” p-Therefore need to “correct” p-values for number of comparisonsvalues for number of comparisons

• A severe correction would be a A severe correction would be a Bonferroni, where Bonferroni, where ppcc = p = pu u /n/n……

… …but this is only appropriate when but this is only appropriate when the the nn tests independent… tests independent…

… … SPMs are smooth, meaning that SPMs are smooth, meaning that nearby voxels are correlatednearby voxels are correlated

=> Gaussian Field Theory...=> Gaussian Field Theory...

• If If n=100,000n=100,000 voxels tested with voxels tested with ppuu=0.05=0.05 of falsely rejecting of falsely rejecting HHoo... ...

… …then approx then approx n n p puu (eg 5,000) (eg 5,000)

will do so by chance (false will do so by chance (false positives, or “type I” errors)positives, or “type I” errors)

• Therefore need to “correct” p-Therefore need to “correct” p-values for number of comparisonsvalues for number of comparisons

• A severe correction would be a A severe correction would be a Bonferroni, where Bonferroni, where ppcc = p = pu u /n/n……

… …but this is only appropriate when but this is only appropriate when the the nn tests independent… tests independent…

… … SPMs are smooth, meaning that SPMs are smooth, meaning that nearby voxels are correlatednearby voxels are correlated

=> Gaussian Field Theory...=> Gaussian Field Theory...

Multiple comparisons…Multiple comparisons…Multiple comparisons…Multiple comparisons…

Gaussian10mm FWHM(2mm pixels)

pu = 0.05

SPM{t} Eg random noise

Page 40: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Gaussian Field TheoryGaussian Field TheoryGaussian Field TheoryGaussian Field Theory

• Consider SPM as lattice representation Consider SPM as lattice representation

of continuous random fieldof continuous random field

• ““Euler characteristic” - topological Euler characteristic” - topological

measure of “excursion set” (e.g,measure of “excursion set” (e.g, # components - # “holes”)# components - # “holes”)

• Smoothness estimated by covariance of Smoothness estimated by covariance of partial derivatives of residuals partial derivatives of residuals (expressed as “resels” or (expressed as “resels” or FWHMFWHM))

• Assumes:Assumes:1) residuals are multivariate normal1) residuals are multivariate normal2) smoothness » voxel size2) smoothness » voxel size(practically, (practically, FWHMFWHM 3 3

VoxDim)VoxDim)

• Not necessarily stationary: smoothness Not necessarily stationary: smoothness estimated locally as resels-per-voxelestimated locally as resels-per-voxel

• Consider SPM as lattice representation Consider SPM as lattice representation

of continuous random fieldof continuous random field

• ““Euler characteristic” - topological Euler characteristic” - topological

measure of “excursion set” (e.g,measure of “excursion set” (e.g, # components - # “holes”)# components - # “holes”)

• Smoothness estimated by covariance of Smoothness estimated by covariance of partial derivatives of residuals partial derivatives of residuals (expressed as “resels” or (expressed as “resels” or FWHMFWHM))

• Assumes:Assumes:1) residuals are multivariate normal1) residuals are multivariate normal2) smoothness » voxel size2) smoothness » voxel size(practically, (practically, FWHMFWHM 3 3

VoxDim)VoxDim)

• Not necessarily stationary: smoothness Not necessarily stationary: smoothness estimated locally as resels-per-voxelestimated locally as resels-per-voxel

Page 41: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

• General form for expected Euler characteristic for General form for expected Euler characteristic for DD dimensions: dimensions:

EE[[(AAuu)] = ] = R Rd d (()) d d ((uu))

• General form for expected Euler characteristic for General form for expected Euler characteristic for DD dimensions: dimensions:

EE[[(AAuu)] = ] = R Rd d (()) d d ((uu))

Rd (): d-dimensional Minkowski

– function of dimension, d, space and smoothness:

R0() = () Euler characteristic of

R1() = resel diameter

R2() = resel surface area

R3() = resel volume

d (): d-dimensional EC density of Z(x)

– function of dimension, d, threshold, u, and statistic, e.g. Z-statistic:

0(u) = 1- (u)

1(u) = (4 ln2)1/2 exp(-u2/2) / (2)

2(u) = (4 ln2) exp(-u2/2) / (2)3/2

3(u) = (4 ln2)3/2 (u2 -1) exp(-u2/2) / (2)2

4(u) = (4 ln2)2 (u3 -3u) exp(-u2/2) / (2)5/2

Generalised FormGeneralised FormGeneralised FormGeneralised Form

Page 42: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Levels of InferenceLevels of InferenceLevels of InferenceLevels of Inference

• Three levels of inference:Three levels of inference:

– extreme voxel valuesextreme voxel values voxel-levelvoxel-level inferenceinference

– big suprathreshold clustersbig suprathreshold clusters cluster-levelcluster-level inferenceinference

– many suprathreshold clustersmany suprathreshold clusters set-levelset-level inferenceinference

• Three levels of inference:Three levels of inference:

– extreme voxel valuesextreme voxel values voxel-levelvoxel-level inferenceinference

– big suprathreshold clustersbig suprathreshold clusters cluster-levelcluster-level inferenceinference

– many suprathreshold clustersmany suprathreshold clusters set-levelset-level inferenceinference

n=82n=82

n=32n=32

n=1n=122

Parameters:Parameters:

““Height” threshold, Height” threshold, uu - t > 3.09- t > 3.09““Extent” threshold, Extent” threshold, kk - 12 voxels- 12 voxels

Dimension, Dimension, DD - 3 - 3 Volume, Volume, SS - 32- 323 3 voxelsvoxelsSmoothnessSmoothness, FWHM, FWHM - 4.7 voxels- 4.7 voxels

Omnibus:Omnibus: P(c P(c 7, t 7, t u) = 0.031 u) = 0.031

voxel-level: voxel-level: P(t P(t 4.37) = .048 4.37) = .048

set-level:set-level: P(c P(c 3, n 3, n k, t k, t u) = 0.019 u) = 0.019

cluster-level:cluster-level: P(n P(n 82, t 82, t u) = 0.029 u) = 0.029

Page 43: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

(Spatial) Specificity vs. Sensitivity(Spatial) Specificity vs. Sensitivity(Spatial) Specificity vs. Sensitivity(Spatial) Specificity vs. Sensitivity

Page 44: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Small-volume correctionSmall-volume correctionSmall-volume correctionSmall-volume correction

• If have an a priori region of interest, no need to correct for If have an a priori region of interest, no need to correct for whole-brain!whole-brain!

• But can use GFT to correct for a Small Volume (But can use GFT to correct for a Small Volume (SVCSVC))

• Volume can be based on:Volume can be based on:

– An anatomically-defined regionAn anatomically-defined region

– A geometric approximation to the above (eg rhomboid/sphere)A geometric approximation to the above (eg rhomboid/sphere)

– A functionally-defined mask (based on an ORTHOGONAL contrast!)A functionally-defined mask (based on an ORTHOGONAL contrast!)

• Extent of correction can be APPROXIMATED by a Bonferonni Extent of correction can be APPROXIMATED by a Bonferonni correction for the number of resels…correction for the number of resels…

• ..but correction also depends on shape (surface area) as well as ..but correction also depends on shape (surface area) as well as size (volume) of region (may want to smooth volume if rough)size (volume) of region (may want to smooth volume if rough)

• If have an a priori region of interest, no need to correct for If have an a priori region of interest, no need to correct for whole-brain!whole-brain!

• But can use GFT to correct for a Small Volume (But can use GFT to correct for a Small Volume (SVCSVC))

• Volume can be based on:Volume can be based on:

– An anatomically-defined regionAn anatomically-defined region

– A geometric approximation to the above (eg rhomboid/sphere)A geometric approximation to the above (eg rhomboid/sphere)

– A functionally-defined mask (based on an ORTHOGONAL contrast!)A functionally-defined mask (based on an ORTHOGONAL contrast!)

• Extent of correction can be APPROXIMATED by a Bonferonni Extent of correction can be APPROXIMATED by a Bonferonni correction for the number of resels…correction for the number of resels…

• ..but correction also depends on shape (surface area) as well as ..but correction also depends on shape (surface area) as well as size (volume) of region (may want to smooth volume if rough)size (volume) of region (may want to smooth volume if rough)

Page 45: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Example SPM window

Page 46: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 47: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Fixed vs. Random EffectsFixed vs. Random EffectsFixed vs. Random EffectsFixed vs. Random Effects

Subject 1

• Subjects can be Subjects can be FixedFixed or or RandomRandom variables variables

• If subjects are a Fixed variable in a single design If subjects are a Fixed variable in a single design matrix (SPM “sessions”), the error term conflates matrix (SPM “sessions”), the error term conflates within- and between-subject variancewithin- and between-subject variance

– In PET, this is not such a problem because the In PET, this is not such a problem because the within-subject (between-scan) variance can be as within-subject (between-scan) variance can be as great as the between-subject variance; but in fMRI great as the between-subject variance; but in fMRI the between-scan variance is normally much the between-scan variance is normally much smaller than the between-subject variancesmaller than the between-subject variance

• If one wishes to make an inference from a subject If one wishes to make an inference from a subject sample to the population, one needs to treat sample to the population, one needs to treat subjects as a Random variablesubjects as a Random variable, and needs a proper , and needs a proper mixture of within- and between-subject variancemixture of within- and between-subject variance

• In SPM, this is achieved by a two-stage procedure:In SPM, this is achieved by a two-stage procedure:1)1) (Contrasts of) parameters are estimated from a (Contrasts of) parameters are estimated from a

(Fixed Effect) model for each subject(Fixed Effect) model for each subject2)2) Images of these contrasts become the data for a Images of these contrasts become the data for a

second design matrix (usually simple t-test or ANOVA) second design matrix (usually simple t-test or ANOVA)

• Subjects can be Subjects can be FixedFixed or or RandomRandom variables variables

• If subjects are a Fixed variable in a single design If subjects are a Fixed variable in a single design matrix (SPM “sessions”), the error term conflates matrix (SPM “sessions”), the error term conflates within- and between-subject variancewithin- and between-subject variance

– In PET, this is not such a problem because the In PET, this is not such a problem because the within-subject (between-scan) variance can be as within-subject (between-scan) variance can be as great as the between-subject variance; but in fMRI great as the between-subject variance; but in fMRI the between-scan variance is normally much the between-scan variance is normally much smaller than the between-subject variancesmaller than the between-subject variance

• If one wishes to make an inference from a subject If one wishes to make an inference from a subject sample to the population, one needs to treat sample to the population, one needs to treat subjects as a Random variablesubjects as a Random variable, and needs a proper , and needs a proper mixture of within- and between-subject variancemixture of within- and between-subject variance

• In SPM, this is achieved by a two-stage procedure:In SPM, this is achieved by a two-stage procedure:1)1) (Contrasts of) parameters are estimated from a (Contrasts of) parameters are estimated from a

(Fixed Effect) model for each subject(Fixed Effect) model for each subject2)2) Images of these contrasts become the data for a Images of these contrasts become the data for a

second design matrix (usually simple t-test or ANOVA) second design matrix (usually simple t-test or ANOVA)

Subject 2

Subject 3

Subject 4

Subject 6

Multi-subject Fixed Effect model

error df ~ 300

Subject 5

Page 48: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

WHEN special case of n independent observations per subject:

var(pop) = 2b + 2

w / Nn

Two-stage “Summary Statistic” approachTwo-stage “Summary Statistic” approachTwo-stage “Summary Statistic” approachTwo-stage “Summary Statistic” approach

p < 0.001 (uncorrected)

SPM{t}

1st-level (within-subject) 2nd-level (between-subject)

cont

rast

imag

es o

f c

i

1^

2^

3^

4^

5^

6^

1^

^

^

^

^

^

wwithin-subject error^

N=6 subjects(error df =5)

One-sample t-test

pop

Page 49: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Limitations of 2-stage approachLimitations of 2-stage approachLimitations of 2-stage approachLimitations of 2-stage approach

• Summary statistic approach is a special case, valid Summary statistic approach is a special case, valid only when each subject’s design matrix is identical only when each subject’s design matrix is identical (“(“balanced designsbalanced designs”)”)

• In practice, the approach is reasonably robust to In practice, the approach is reasonably robust to unbalanced designs (Penny, 2004)unbalanced designs (Penny, 2004)

• More generally, exact solutions to any hierarchical More generally, exact solutions to any hierarchical GLM can be obtained using ReMLGLM can be obtained using ReML

• This is computationally expensive to perform at This is computationally expensive to perform at every voxel (so not implemented in SPM2)every voxel (so not implemented in SPM2)

• Plus modelling of nonsphericity at 2Plus modelling of nonsphericity at 2ndnd-level can -level can minimise potential bias of unbalanced designs… minimise potential bias of unbalanced designs…

• Summary statistic approach is a special case, valid Summary statistic approach is a special case, valid only when each subject’s design matrix is identical only when each subject’s design matrix is identical (“(“balanced designsbalanced designs”)”)

• In practice, the approach is reasonably robust to In practice, the approach is reasonably robust to unbalanced designs (Penny, 2004)unbalanced designs (Penny, 2004)

• More generally, exact solutions to any hierarchical More generally, exact solutions to any hierarchical GLM can be obtained using ReMLGLM can be obtained using ReML

• This is computationally expensive to perform at This is computationally expensive to perform at every voxel (so not implemented in SPM2)every voxel (so not implemented in SPM2)

• Plus modelling of nonsphericity at 2Plus modelling of nonsphericity at 2ndnd-level can -level can minimise potential bias of unbalanced designs… minimise potential bias of unbalanced designs…

New inSPM2

Page 50: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Nonsphericity again!Nonsphericity again!Nonsphericity again!Nonsphericity again!

• When tests at 2When tests at 2ndnd-level are more complicated than -level are more complicated than 1/2-sample t-tests, errors can be non i.i.d1/2-sample t-tests, errors can be non i.i.d

• For example, two groups (e.g, patients and controls) For example, two groups (e.g, patients and controls) may have different variances (non-identically may have different variances (non-identically distributed; distributed; inhomogeniety of varianceinhomogeniety of variance))

• Or when taking more than one parameter per subject Or when taking more than one parameter per subject ((repeated measuresrepeated measures, e.g, multiple basis functions in , e.g, multiple basis functions in event-related fMRI), errors may be non-independent event-related fMRI), errors may be non-independent

(If nonsphericity correction selected, inhomogeniety (If nonsphericity correction selected, inhomogeniety assumed, and further option for repeated measures)assumed, and further option for repeated measures)

• Same method of variance component estimation with Same method of variance component estimation with ReML (that used for autocorrelation) is usedReML (that used for autocorrelation) is used

(Greenhouse-Geisser correction for repeated-(Greenhouse-Geisser correction for repeated-measures ANOVAs is a special case approximation)measures ANOVAs is a special case approximation)

• When tests at 2When tests at 2ndnd-level are more complicated than -level are more complicated than 1/2-sample t-tests, errors can be non i.i.d1/2-sample t-tests, errors can be non i.i.d

• For example, two groups (e.g, patients and controls) For example, two groups (e.g, patients and controls) may have different variances (non-identically may have different variances (non-identically distributed; distributed; inhomogeniety of varianceinhomogeniety of variance))

• Or when taking more than one parameter per subject Or when taking more than one parameter per subject ((repeated measuresrepeated measures, e.g, multiple basis functions in , e.g, multiple basis functions in event-related fMRI), errors may be non-independent event-related fMRI), errors may be non-independent

(If nonsphericity correction selected, inhomogeniety (If nonsphericity correction selected, inhomogeniety assumed, and further option for repeated measures)assumed, and further option for repeated measures)

• Same method of variance component estimation with Same method of variance component estimation with ReML (that used for autocorrelation) is usedReML (that used for autocorrelation) is used

(Greenhouse-Geisser correction for repeated-(Greenhouse-Geisser correction for repeated-measures ANOVAs is a special case approximation)measures ANOVAs is a special case approximation)

New inSPM2

Inhomogeneous variance(3 groups of 4 subjects)

Q

Repeated measures(3 groups of 4 subjects)

Q

1

2

3

Page 51: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Hierarchical ModelsHierarchical ModelsHierarchical ModelsHierarchical Models

• Two-stage approach is special case of Two-stage approach is special case of Hierarchical GLMHierarchical GLM

• In a Bayesian framework, parameters of one In a Bayesian framework, parameters of one level can be made priors on distribution of level can be made priors on distribution of parameters at lower level: “parameters at lower level: “Parametric Parametric Empirical Bayes” Empirical Bayes” (Friston et al, 2002)(Friston et al, 2002)

• The parameters and hyperparameters at each The parameters and hyperparameters at each level can be estimated using level can be estimated using EM algorithmEM algorithm (generalisation of ReML)(generalisation of ReML)

• Note parameters and hyperparameters at final Note parameters and hyperparameters at final level do not differ from classical frameworklevel do not differ from classical framework

• Second-level could be subjects; it could also Second-level could be subjects; it could also be voxels…be voxels…

• Two-stage approach is special case of Two-stage approach is special case of Hierarchical GLMHierarchical GLM

• In a Bayesian framework, parameters of one In a Bayesian framework, parameters of one level can be made priors on distribution of level can be made priors on distribution of parameters at lower level: “parameters at lower level: “Parametric Parametric Empirical Bayes” Empirical Bayes” (Friston et al, 2002)(Friston et al, 2002)

• The parameters and hyperparameters at each The parameters and hyperparameters at each level can be estimated using level can be estimated using EM algorithmEM algorithm (generalisation of ReML)(generalisation of ReML)

• Note parameters and hyperparameters at final Note parameters and hyperparameters at final level do not differ from classical frameworklevel do not differ from classical framework

• Second-level could be subjects; it could also Second-level could be subjects; it could also be voxels…be voxels…

y = X(1) (1) + (1)

(1)= X(2) (2) + (2)

(n-1)= X(n) (n) + (n)

C(i) = k

(i) Qk

(i)

New inSPM2

Page 52: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Parametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMs

• Bayes rule:Bayes rule:

p(p(|y) = p(y|) p()• Bayes rule:Bayes rule:

p(p(|y) = p(y|) p()

New inSPM2

Posterior Likelihood Prior

(PPM) (SPM)

• What are the priors?

– In “classical” SPM, no (flat) priors

– In “full” Bayes, priors might be from theoretical arguments, or from independent data

– In “empirical” Bayes, priors derive from same data, assuming a hierarchical model for generation of that data

Page 53: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Parametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMs

• Bayes rule:Bayes rule:

p(p(|y) = p(y|) p()• Bayes rule:Bayes rule:

p(p(|y) = p(y|) p()

New inSPM2

Posterior Likelihood Prior

(PPM) (SPM)

• For PPMs in SPM2, priors come from distribution over voxels

• If remove mean over voxels, prior mean can be set to zero (a “shrinkage” prior)

• One can threshold posteriors for a given probability of a parameter estimate greater than some value …

• …to give a posterior probability map (PPM)

u

)(yft

)0|( tp

Classical T-test

)|( yp

Bayesian test

Page 54: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Parametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsParametric Empirical Bayes & PPMsNew inSPM2

SP

Mm

ip[0

, 0, 0

]

<

< <

PPM2.06

rest [2.06]

SPMresults:C:\home\spm\analysis_PET

Height threshold P = 0.95

Extent threshold k = 0 voxels

Design matrix1 4 7 10 13 16 19 22

147

1013161922252831343740434649525560

contrast(s)

4

SP

Mm

ip[0

, 0, 0

]

<

< <

SPM{T39.0

}

rest

SPMresults:C:\home\spm\analysis_PET

Height threshold T = 5.50

Extent threshold k = 0 voxels

Design matrix1 4 7 10 13 16 19 22

147

1013161922252831343740434649525560

contrast(s)

3

• Activations greater than certain amount Voxels with non-zero activations

• Can infer no responses Cannot “prove the null hypothesis”

• No fallacy of inference Fallacy of inference (large df)

• Inference independent of search volume Correct for search volume

• Computationally expensive Computationally faster

• Activations greater than certain amount Voxels with non-zero activations

• Can infer no responses Cannot “prove the null hypothesis”

• No fallacy of inference Fallacy of inference (large df)

• Inference independent of search volume Correct for search volume

• Computationally expensive Computationally faster

Page 55: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 56: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 57: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 58: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A categorical analysisA categorical analysisA categorical analysisA categorical analysis

Experimental designExperimental design

Word generationWord generation GGWord repetitionWord repetition RR

R G R G R G R G R G R GR G R G R G R G R G R G

G - R = Intrinsic word generationG - R = Intrinsic word generation

……under assumption of under assumption of pure insertionpure insertion,,ie, that G and R do not differ in other waysie, that G and R do not differ in other ways

Page 59: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 60: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Cognitive ConjunctionsCognitive ConjunctionsCognitive ConjunctionsCognitive Conjunctions

• One way to minimise problem of pure insertion is to One way to minimise problem of pure insertion is to isolate same process in several different ways (ie, isolate same process in several different ways (ie, multiple subtractions of different conditions)multiple subtractions of different conditions)

A1A1 A2A2

B2B2B1B1

Task (1/2)Task (1/2)ViewingViewing Naming Naming

Stim

uli (

A/B

)S

timul

i (A

/B)

Obj

ects

C

olou

rsO

bjec

ts

Col

ours

Visual ProcessingVisual Processing V V Object Recognition Object Recognition RRPhonological RetrievalPhonological Retrieval PP

Object viewingObject viewing R,VR,VColour viewingColour viewing VVObject namingObject naming P,R,VP,R,VColour namingColour naming P,VP,V

(Object - Colour viewing) [1 -1 0 0] (Object - Colour viewing) [1 -1 0 0] &&

(Object - Colour naming) [0 0 1 -1](Object - Colour naming) [0 0 1 -1]

[ R,V - V ] & [ P,R,V - P,V ] = R & R = R[ R,V - V ] & [ P,R,V - P,V ] = R & R = R

(assuming (assuming RxPRxP = 0; see later) = 0; see later)

Price et al, 1997Common object Common object

recognition response (R)recognition response (R)

Page 61: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Cognitive ConjunctionsCognitive ConjunctionsCognitive ConjunctionsCognitive Conjunctions

• Original (SPM97) definition of conjunctions Original (SPM97) definition of conjunctions entailed sum of two simple effects (A1-A2 + entailed sum of two simple effects (A1-A2 + B1-B2) plus B1-B2) plus exclusiveexclusive masking with masking with interaction (A1-A2) - (B1-B2)interaction (A1-A2) - (B1-B2)

• Ie, “effects significant and of similar size”Ie, “effects significant and of similar size”

• (Difference between conjunctions and (Difference between conjunctions and masking is that conjunction p-values reflect masking is that conjunction p-values reflect the conjoint probabilities of the contrasts)the conjoint probabilities of the contrasts)

• SPM2 defintion of conjunctions uses SPM2 defintion of conjunctions uses advances in Gaussian Field Theory (e.g, advances in Gaussian Field Theory (e.g, TT22 fields), allowing corrected p-values fields), allowing corrected p-values

• However, the logic has changed slightly, in However, the logic has changed slightly, in that voxels can survive a conjunction even that voxels can survive a conjunction even though they show an interactionthough they show an interaction

New inSPM2

A1-A2

B

1-B

2

A1-A2

B

1-B

2 p(A1=A2)<p

p((A1-A2)= (B1-B2))>P2

p(A1=A2+B1=B2)<P1

+

+

p(B1=B2)<p

Page 62: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 63: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Nonlinear parametric responsesNonlinear parametric responsesNonlinear parametric responsesNonlinear parametric responses

Inverted ‘U’ response toInverted ‘U’ response toincreasing word presentationincreasing word presentation

rate in the DLPFCrate in the DLPFCSPM{F}SPM{F}

Polynomial expansion:Polynomial expansion:f(x) ~ f(x) ~ x + x2 + ...

…(N-1)th order for N levels

Lin

ear

Qua

drat

ic

E.g, F-contrast [0 1 0] on E.g, F-contrast [0 1 0] on Quadratic Parameter =>Quadratic Parameter =>

Page 64: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 65: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Interactions and pure insertionInteractions and pure insertionInteractions and pure insertionInteractions and pure insertion

• Presence of an interaction can show a failure of Presence of an interaction can show a failure of pure insertion (using earlier example)…pure insertion (using earlier example)…

A1A1 A2A2

B2B2B1B1

Task (1/2)Task (1/2)ViewingViewing Naming Naming

Stim

uli (

A/B

)S

timul

i (A

/B)

Obj

ects

C

olou

rsO

bjec

ts

Col

ours

Visual ProcessingVisual Processing V V Object Recognition Object Recognition RRPhonological RetrievalPhonological Retrieval PP

Object viewingObject viewing R,VR,VColour viewingColour viewing VVObject namingObject naming P,R,V,P,R,V,RxPRxPColour namingColour naming P,VP,V Naming-specific Naming-specific

object recognitionobject recognition

viewing namingviewing naming

Obj

ect -

Col

our

Obj

ect -

Col

our

(Object – Colour) x (Viewing – Naming) (Object – Colour) x (Viewing – Naming)

[1 -1 0 0] - [0 0 1 -1] = [1 -1] [1 -1 0 0] - [0 0 1 -1] = [1 -1] [1 -1] = [1 -1 -1 1] [1 -1] = [1 -1 -1 1]

[ R,V - V ] - [ P,R,V,[ R,V - V ] - [ P,R,V,RxPRxP - P,V ] = R – R, - P,V ] = R – R,RxPRxP = = RxPRxP

Page 66: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

A taxonomy of designA taxonomy of designA taxonomy of designA taxonomy of design

• Categorical designsCategorical designsSubtraction Subtraction - Additive factors and pure insertion- Additive factors and pure insertion

Conjunction Conjunction - Testing multiple hypotheses- Testing multiple hypotheses

• Parametric designsParametric designsLinear Linear - Cognitive components and dimensions- Cognitive components and dimensions

NonlinearNonlinear - Polynomial expansions - Polynomial expansions

• Factorial designsFactorial designsCategoricalCategorical - Interactions and pure insertion- Interactions and pure insertion

- Adaptation, modulation and dual-task inference- Adaptation, modulation and dual-task inference

ParametricParametric - Linear and nonlinear interactions- Linear and nonlinear interactions

- Psychophysiological Interactions- Psychophysiological Interactions

Page 67: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

SPM{Z}SPM{Z}

Attentional modulation ofAttentional modulation ofV1 - V5 contributionV1 - V5 contribution

AttentionAttention

V1V1

V5V5

attention

no attention

V1 activityV1 activity

V5

acti

vity

timetime

V1

acti

vity

Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)

Parametric, factorial design, in which one factor is psychological (eg attention)

...and other is physiological (viz. activity extracted from a brain region of interest)

Page 68: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)Psycho-physiological Interaction (PPI)

• PPIs tested by a GLM with form:PPIs tested by a GLM with form:

yy = ( = (V1V1AA).).11 + + V1V1..22 + + AA..33 + + c = [1 0 0]c = [1 0 0]

• However, the interaction term of interest, However, the interaction term of interest, V1V1AA, is the product of V1 , is the product of V1 activity and Attention block AFTER convolution with HRFactivity and Attention block AFTER convolution with HRF

• We are really interested in interaction at neural level, but:We are really interested in interaction at neural level, but:

(HRF (HRF V1) V1) (HRF (HRF A) A) HRF HRF (V1 (V1 A A) )

(unless A low frequency, eg, blocked; so problem for event-related PPIs)(unless A low frequency, eg, blocked; so problem for event-related PPIs)

• SPM2 can effect a SPM2 can effect a deconvolutiondeconvolution of physiological regressors (V1), before of physiological regressors (V1), before calculating interaction term and reconvolving with the HRFcalculating interaction term and reconvolving with the HRF

• Deconvolution is ill-constrained, so regularised using smoothness priorsDeconvolution is ill-constrained, so regularised using smoothness priors(using ReML)(using ReML)

New inSPM2

Page 69: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

OverviewOverviewOverviewOverview

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

1. General Linear Model1. General Linear ModelDesign MatrixDesign MatrixGlobal normalisationGlobal normalisation

2. fMRI timeseries2. fMRI timeseriesHighpass filteringHighpass filteringHRF convolutionHRF convolutionTemporal autocorrelationTemporal autocorrelation

3. Statistical Inference3. Statistical InferenceGaussian Field TheoryGaussian Field Theory

4. Random Effects4. Random Effects

5. Experimental Designs5. Experimental Designs

6. Effective Connectivity6. Effective Connectivity

Page 70: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Effective vs. functional connectivityEffective vs. functional connectivityEffective vs. functional connectivityEffective vs. functional connectivity

No connection between B and C,yet B and C correlated because of common input from A, eg:

A = V1 fMRI time-seriesB = 0.5 * A + e1C = 0.3 * A + e2

Correlations:

A B C10.49 10.30 0.12 1

A

B

C

0.49

0.31

-0.02

2=0.5, ns.Functional Functional connectivityconnectivity

Effective connectivityEffective connectivity

Page 71: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Dynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingDynamic Causal Modelling

• PPIs allow a simple (restricted) test of effective connectivityPPIs allow a simple (restricted) test of effective connectivity

• Structural Equation Modelling is more powerful (Buchel & Friston, 1997)Structural Equation Modelling is more powerful (Buchel & Friston, 1997)

• However in SPM2, Dynamic Causal Modelling (However in SPM2, Dynamic Causal Modelling (DCMDCM) is preferred) is preferred

• DCMs are dynamic models specified at the neural levelDCMs are dynamic models specified at the neural level

• The neural dynamics are transformed into predicted BOLD signals using a The neural dynamics are transformed into predicted BOLD signals using a realistic biological haemodynamic forward model (realistic biological haemodynamic forward model (HDMHDM))

• The neural dynamics comprise a deterministic state-space model and a The neural dynamics comprise a deterministic state-space model and a bilinear approximation to model interactions between variablesbilinear approximation to model interactions between variables

New inSPM2

Page 72: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Dynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingDynamic Causal Modelling

• The variables consist of:The variables consist of:connections between regionsconnections between regionsself-connections self-connections direct inputs (eg, visual stimulations)direct inputs (eg, visual stimulations)contextual inputs (eg, attention)contextual inputs (eg, attention)

• Connections can be bidirectionalConnections can be bidirectional

• Variables estimated using EM algorithmVariables estimated using EM algorithm

• Priors are:Priors are:empirical (for haemodynamic model)empirical (for haemodynamic model)principled (dynamics to be convergent)principled (dynamics to be convergent)shrinkage (zero-mean, for connections)shrinkage (zero-mean, for connections)

• Inference using posterior probabilitiesInference using posterior probabilities

• Methods for Bayesian model comparisonMethods for Bayesian model comparison

New inSPM2

direct inputs - u1

(e.g. visual stimuli)

z2 V5z1 V1

y1

z3 SPC

contextual inputs - u2

(e.g. attention)

y2 y3

z = f(z,u,z) Az + uBz + Cu

y = h(z,h) +

z = state vectoru = inputs = parameters (connection/haemodynamic)

.

Page 73: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Dynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingNew inSPM2

Z2

stimuliu1

contextu2

Z1

+

+

-

-

+

u1

u2

z2

z1

-

-

Page 74: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Dynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingDynamic Causal ModellingNew inSPM2

V1 IFG

V5

SPC

Motion

Photic

Attention

.82(100%)

.42(100%)

.37(90%)

.69 (100%).47

(100%)

.65 (100%)

.52 (98%)

.56(99%)

Friston et al. (2003)

Büchel & Friston (1997)

EffectsEffects

Photic – dots vs fixationPhotic – dots vs fixationMotion – moving vs staticMotion – moving vs staticAttenton – detect changesAttenton – detect changes

• Attention modulates the backward-Attention modulates the backward-connections IFGconnections IFG→SPC and SPC→V5→SPC and SPC→V5

• The intrinsic connection V1→V5 is The intrinsic connection V1→V5 is insignificant in the absence of motioninsignificant in the absence of motion

Page 75: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:
Page 76: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Friston KJ, Holmes AP, Worsley KJ, Poline J-B, Frith CD, Frackowiak RSJ (1995) Statistical parametric maps in functional imaging: A general linear approach” Human Brain Mapping 2:189-210

Worsley KJ & Friston KJ (1995) Analysis of fMRI time series revisited — again” NeuroImage 2:173-181

Friston KJ, Josephs O, Zarahn E, Holmes AP, Poline J-B (2000) “To smooth or not to smooth” NeuroImage

Zarahn E, Aguirre GK, D'Esposito M (1997) “Empirical Analyses of BOLD fMRI Statistics” NeuroImage 5:179-197

Holmes AP, Friston KJ (1998) “Generalisability, Random Effects & Population Inference” NeuroImage 7(4-2/3):S754

Worsley KJ, Marrett S, Neelin P, Evans AC (1992) “A three-dimensional statistical analysis for CBF activation studies in human brain”Journal of Cerebral Blood Flow and Metabolism 12:900-918

Worsley KJ, Marrett S, Neelin P, Vandal AC, Friston KJ, Evans AC (1995) “A unified statistical approach for determining significant signals in images of cerebral activation” Human Brain Mapping 4:58-73

Friston KJ, Worsley KJ, Frackowiak RSJ, Mazziotta JC, Evans AC (1994) Assessing the Significance of Focal Activations Using their Spatial Extent” Human Brain Mapping 1:214-220

Cao J (1999) The size of the connected components of excursion sets of 2, t and F fields” Advances in Applied Probability (in press)

Worsley KJ, Marrett S, Neelin P, Evans AC (1995) Searching scale space for activation in PET images” Human Brain Mapping 4:74-90

Worsley KJ, Poline J-B, Vandal AC, Friston KJ (1995) Tests for distributed, non-focal brain activations” NeuroImage 2:183-194

Friston KJ, Holmes AP, Poline J-B, Price CJ, Frith CD (1996) Detecting Activations in PET and fMRI: Levels of Inference and Power” Neuroimage 4:223-235

Friston KJ, Holmes AP, Worsley KJ, Poline J-B, Frith CD, Frackowiak RSJ (1995) Statistical parametric maps in functional imaging: A general linear approach” Human Brain Mapping 2:189-210

Worsley KJ & Friston KJ (1995) Analysis of fMRI time series revisited — again” NeuroImage 2:173-181

Friston KJ, Josephs O, Zarahn E, Holmes AP, Poline J-B (2000) “To smooth or not to smooth” NeuroImage

Zarahn E, Aguirre GK, D'Esposito M (1997) “Empirical Analyses of BOLD fMRI Statistics” NeuroImage 5:179-197

Holmes AP, Friston KJ (1998) “Generalisability, Random Effects & Population Inference” NeuroImage 7(4-2/3):S754

Worsley KJ, Marrett S, Neelin P, Evans AC (1992) “A three-dimensional statistical analysis for CBF activation studies in human brain”Journal of Cerebral Blood Flow and Metabolism 12:900-918

Worsley KJ, Marrett S, Neelin P, Vandal AC, Friston KJ, Evans AC (1995) “A unified statistical approach for determining significant signals in images of cerebral activation” Human Brain Mapping 4:58-73

Friston KJ, Worsley KJ, Frackowiak RSJ, Mazziotta JC, Evans AC (1994) Assessing the Significance of Focal Activations Using their Spatial Extent” Human Brain Mapping 1:214-220

Cao J (1999) The size of the connected components of excursion sets of 2, t and F fields” Advances in Applied Probability (in press)

Worsley KJ, Marrett S, Neelin P, Evans AC (1995) Searching scale space for activation in PET images” Human Brain Mapping 4:74-90

Worsley KJ, Poline J-B, Vandal AC, Friston KJ (1995) Tests for distributed, non-focal brain activations” NeuroImage 2:183-194

Friston KJ, Holmes AP, Poline J-B, Price CJ, Frith CD (1996) Detecting Activations in PET and fMRI: Levels of Inference and Power” Neuroimage 4:223-235

Some ReferencesSome ReferencesSome ReferencesSome References

Page 77: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

PCA/SVD and EigenimagesPCA/SVD and EigenimagesPCA/SVD and EigenimagesPCA/SVD and Eigenimages

A time-series of 1D imagesA time-series of 1D images128 scans of 32 “voxels”128 scans of 32 “voxels”

Expression of 1st 3 “eigenimages”Expression of 1st 3 “eigenimages”

Eigenvalues and spatial “modes”Eigenvalues and spatial “modes”

The time-series ‘reconstituted’The time-series ‘reconstituted’

Page 78: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

PCA/SVD and EigenimagesPCA/SVD and EigenimagesPCA/SVD and EigenimagesPCA/SVD and Eigenimages

Y Y (DATA)(DATA)

timetime

voxelsvoxels

Y = USVY = USVTT = = ss11UU11VV11TT + + ss22UU22VV22

T T + ... + ...

+ ...+ ...APPROX. APPROX. OF YOF Y

UU11

== ss11

VV11

APPROX. APPROX. OF YOF Y

+ + ss22

UU22

VV22

+ + ss33 APPROX. APPROX. OF YOF Y

UU33

VV33

Page 79: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Time x Condition interactionTime x Condition interactionTime x Condition interactionTime x Condition interaction

Time x condition interactions (Time x condition interactions (i.e. adaptationi.e. adaptation))assessed with the SPM{T}assessed with the SPM{T}

Page 80: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Minimise the difference between the observed (Minimise the difference between the observed (SS) and implied () and implied () covariances by adjusting the ) covariances by adjusting the path coefficients (path coefficients (BB) )

The implied covariance structure: The implied covariance structure: xx = x.B + z= x.B + zxx = z.(I - B)= z.(I - B)-1-1

x : matrix of time-series of Regions 1-3x : matrix of time-series of Regions 1-3B: matrix of unidirectional path coefficientsB: matrix of unidirectional path coefficients

Variance-covariance structure:Variance-covariance structure:xxT T . x = . x = = (I-B)= (I-B)-T-T. C.(I-B). C.(I-B)-1-1

where Cwhere C = z= zTT z z

xxTT.x is the implied variance covariance structure .x is the implied variance covariance structure C contains the residual variances (u,v,w) and covariancesC contains the residual variances (u,v,w) and covariances

The free parameters are estimated by minimising a [maximum likelihood] function of The free parameters are estimated by minimising a [maximum likelihood] function of SS and and

Structural Equation Modelling (SEM)Structural Equation Modelling (SEM)Structural Equation Modelling (SEM)Structural Equation Modelling (SEM)

1

3

2

zz zz

zz

12B

23B13B

Page 81: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

Attention - No attentionAttention - No attentionAttention - No attentionAttention - No attention

AttentionNo attention

0.760.47

0.750.43

Changes in “effective connectivity”

Page 82: Statistical Analysis Rik Henson With thanks to: Karl Friston, Andrew Holmes, Stefan Kiebel, Will Penny Statistical Analysis Rik Henson With thanks to:

PPPP

==

Second-order InteractionsSecond-order InteractionsSecond-order InteractionsSecond-order Interactions

V5V5

V1V1

V1xPPV1xPP

V5V5

2 =11, p<0.01

0.14

Modulatory influence of parietal cortex on V1 to V5Modulatory influence of parietal cortex on V1 to V5