fyst17 lecture 8 statistics: fitting and hypothesis testing

42
FYST17 Lecture 8 Statistics and hypothesis testing Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons Suggested reading: statistics book chap 3 &11

Upload: others

Post on 11-Apr-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

FYST17 Lecture 8Statistics and hypothesis testing

Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

Suggested reading: statistics book chap 3 &11

Page 2: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Plan for today:

• Introduction to concepts

– The Gaussian distribution

• Likelihood functions

• Hypothesis testing

– Including p-values and significance

• More examples

2

Page 3: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Interpretation of probability

3

(Bayesian approach)

Page 4: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

PDF = probability density function

4

Page 5: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Definitions

5

-1 xy 1

Page 6: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

PDF examples

Binomial: N trials with p chance of success, probability for n successes:

If N→ and p → 0 but Np→ then we have Poisson: (already N>50, p<0.1

works) f 𝑛; 𝜆 =𝜆𝑛

𝑛!𝑒−𝜆

And for large s can

use Gaussian

f

6

Page 7: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

The Gaussian distribution

The Gaussian pdf is defined by

E[x] = V[x] = ²

”standard Gaussian”

If y is a Gaussian with , ², then x =

y − μ

σfollows (x)

7

Page 8: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

The Gaussian distribution

It is useful to know the most common Gaussian integrals:

@Wikipedia

8

Page 9: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

ATLAS examples of Gaussian distributions

Distribution of vertex z coordinate for tracks

Invariant mass for K0s peak

fitted with a Gaussian (!)

9

Page 10: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Central limit theorem

What we used already in MC studies

Try for yourselves!

Example: sum of 10 uniform

numbers = Gaussian!

Gaussian functions play

important role in applied statistics

Uncertainties tend to be Gaussian!

Central Limit theorem:The sum on N independent continuous random variables xi with means i and variances i² becomes a Gaussian random variable with mean = i i and variance ² = i i² in the limit that N approaches infinity

10

Page 11: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Quick exercise

Measurement of transverse momentum of a track from a fit

– Radius of helix given by R=0.3BpT

– Track fit returns a Gaussian uncertainty in the curvature, e.g. the pdf is Gaussian in 1/pT

– What is the error on pT?

𝜎𝑝𝑇𝑝𝑇

= 𝑝𝑇 ⋅ 𝜎1/𝑝𝑇

11

Page 12: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Error propagation in two variables

13

Page 13: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Parameter estimation

14

Page 14: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Estimators

15

Page 15: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Likelihood functions

Given a PDF f(x) with parameter(s) , what is the probability that with N observations, xi falls in the intervals [xi ; xi+ dxi ]?Described by the likelihood function:

16

Page 16: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Given a set of measurements xi and parameter(s) , the likelihood function is defined as:

The principle of maximum likelihood for parameter estimation consists of maximizing the likelihood of parameter(s) (here ) given some data (here x)

The likelihood function plays a central role in statistics, as it can shown to be: Consistent (converges to the right value!) Asymptotically normal (converges with Gaussian

errors)Efficient and ”optimal” if it can be applied in practice

Computational: often easier to minimize log likelihood:

In problems with Gaussian errors boils down to a ²

Two versions, in practice:- Binned likelihood- Unbinned likelihood

Likelihood functions

f

17

Page 17: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Binned likelihoodSum over bins in a histogram:

18

Page 18: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Unbinned likelihoodSum over single measurements:

19

Page 19: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Hypothesis testing

20

Page 20: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Hypotheses and acceptance/rejection regions

21

Page 21: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Test statistics

1. State hypothesis (null and alternative)

2. Set criteria for decision, select test statistics, select a significance level

3. Compute the value of the test statistics and from that the probability of observation under null-hypothesis (p-value)

4. Make the decision! Reject null hypothesis if p-value is below significance level

Null hypothesisAlternative hypothesis

22

Page 22: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Test statistics

23

Page 23: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Example of hypothesis testThe spin of the newly discovered Higgs-like particle (spin 0 or 2?)

PDF of spin 2hypothesis

PDF of spin 0hypothesis

Test statistics (Likelihood ratio[decay angles] ) 24

Page 24: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Selection

25

Page 25: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Other selection options

26

Page 26: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

ALICE example

27

Page 27: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Type I / Type II errors:

29

Page 28: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Signal/background efficiency

30

Page 29: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance tests/goodness of fit

32

Page 30: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

p-values

33

Page 31: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance of an observed signal

!

34

Page 32: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance of an observed signal

!

35

Page 33: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance vs p-value

36

Small p = unexpected

Z or

Page 34: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance of a peak

37

Page 35: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Significance of a peak

38

Page 36: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

How many ’s?

39

Page 37: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Look-elsewhere-effect (LLE)

Example from CDF: Is there a bump at 7.2 GeV ? (and even 7.75 GeV?!)

Excess has significance but when we take into account that the bump(s) could have been anywhere in the spectrum (the look-elsewhere-effect) significance is reduced:

p-value(corr) = p-value × (number of places it might have been spotted in spectrum)

In this case ~ mass interval / width of bump

Results in low significanceNever saw these again

40

Page 38: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Remember the penta-quark …

41

Page 39: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

The ATLAS/CMS diphoton bump

42

Page 40: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Limit settingCMS 2012 Higgs

result

(example 3.2)

43

Page 41: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Confidence intervalsUpper limt on signal cross section/ SM Higgs cross section in H WW channel

(example 3.2)

44

Page 42: FYST17 Lecture 8 Statistics: fitting and hypothesis testing

Summary/ outlook

• Gaussian distribution very useful

– Errors tend to be gaussian

• To check a New Physics hypothesis against the Standard Model

– Define test statistics

– Define level of significance

– Remember the look elsewhere effect

• P-values gives P(data|null hypothesis)

– It does not say whether the hypothesis is true!

45