on some assumptions of null hypothesis statistical testing

73
On some assumptions of Null Hypothesis Statistical Testing (NHST) Alexandre Galv~ao Patriota Departamento de Estat stica Instituto de Matematica e Estat stica Universidade de S~ao Paulo Alexandre G. Patriota ([email protected]) On some assumptions of NHST 1 / 40

Upload: others

Post on 29-May-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: On some assumptions of Null Hypothesis Statistical Testing

On some assumptions of Null HypothesisStatistical Testing (NHST)

Alexandre Galvao Patriota

Departamento de EstatısticaInstituto de Matematica e Estatıstica

Universidade de Sao Paulo

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 1 / 40

Page 2: On some assumptions of Null Hypothesis Statistical Testing

Outline

1 Main Goals

2 The classical statistical model

3 Hypothesis testing

4 P-value definition and its limitations

5 An alternative measure of evidence and some of its properties

6 Numerical illustration

7 Final remarks

8 References

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 2 / 40

Page 3: On some assumptions of Null Hypothesis Statistical Testing

Main Goals

Main goals

The main goals of this presentation are

to discuss the classical statistical model and statisticalhypotheses,

to present some limitations of the classical p-value withnumerical examples,

to introduce an alternative measure of evidence, calleds-value, that overcomes some limitations of the p-value.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 3 / 40

Page 4: On some assumptions of Null Hypothesis Statistical Testing

Main Goals

Main goals

The main goals of this presentation are

to discuss the classical statistical model and statisticalhypotheses,

to present some limitations of the classical p-value withnumerical examples,

to introduce an alternative measure of evidence, calleds-value, that overcomes some limitations of the p-value.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 3 / 40

Page 5: On some assumptions of Null Hypothesis Statistical Testing

Main Goals

Main goals

The main goals of this presentation are

to discuss the classical statistical model and statisticalhypotheses,

to present some limitations of the classical p-value withnumerical examples,

to introduce an alternative measure of evidence, calleds-value, that overcomes some limitations of the p-value.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 3 / 40

Page 6: On some assumptions of Null Hypothesis Statistical Testing

Main Goals

Main goals

The main goals of this presentation are

to discuss the classical statistical model and statisticalhypotheses,

to present some limitations of the classical p-value withnumerical examples,

to introduce an alternative measure of evidence, calleds-value, that overcomes some limitations of the p-value.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 3 / 40

Page 7: On some assumptions of Null Hypothesis Statistical Testing

The classical statistical model

The classical statistical modelThe classical statistical model is:

(Ω,F ,P),

where:

Ω is the space of possible experimentoutcomes,

F is a σ-field of Ω,

P is a family of non-random probabilitymeasures that possibly explain theexperiment outcomes.

P

P1 : F → [0, 1]

P2 : F → [0, 1]

.

.

.

Pj : F → [0, 1]

.

.

.

The quantity of interest is g(P).For instance:

g(P) = EP (Z),

g(P) = P(Z1 ∈ B|Z2 ∈ A),

etc.

Remark: a random vector Z is a measurable function from (Ω,F)to (Z,B)

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 4 / 40

Page 8: On some assumptions of Null Hypothesis Statistical Testing

The classical statistical model

The classical statistical modelThe classical statistical model is:

(Ω,F ,P),

where:

Ω is the space of possible experimentoutcomes,

F is a σ-field of Ω,

P is a family of non-random probabilitymeasures that possibly explain theexperiment outcomes.

P

P1 : F → [0, 1]

P2 : F → [0, 1]

.

.

.

Pj : F → [0, 1]

.

.

.

The quantity of interest is g(P).For instance:

g(P) = EP (Z),

g(P) = P(Z1 ∈ B|Z2 ∈ A),

etc.

Remark: a random vector Z is a measurable function from (Ω,F)to (Z,B)

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 4 / 40

Page 9: On some assumptions of Null Hypothesis Statistical Testing

The classical statistical model

A particular model

Take Z = (X , γ), where X is the observable random vectorand γ is the unobservable one.

Conditional, marginal and joint distributions can be used tomake inferences about γ.

Take P = P0 and build your joint probabilityP0 from:

γ ∼ f0(·) (with no unknown constants),

X |γ ∼ f1(·|γ)

Now, you are ready to be a hard coreBayesian!

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 5 / 40

Page 10: On some assumptions of Null Hypothesis Statistical Testing

The classical statistical model

A particular model

Take Z = (X , γ), where X is the observable random vectorand γ is the unobservable one.

Conditional, marginal and joint distributions can be used tomake inferences about γ.

Take P = P0 and build your joint probabilityP0 from:

γ ∼ f0(·) (with no unknown constants),

X |γ ∼ f1(·|γ)

Now, you are ready to be a hard coreBayesian!

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 5 / 40

Page 11: On some assumptions of Null Hypothesis Statistical Testing

The classical statistical model

A particular model

Take Z = (X , γ), where X is the observable random vectorand γ is the unobservable one.

Conditional, marginal and joint distributions can be used tomake inferences about γ.

Take P = P0 and build your joint probabilityP0 from:

γ ∼ f0(·) (with no unknown constants),

X |γ ∼ f1(·|γ)

Now, you are ready to be a hard coreBayesian!

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 5 / 40

Page 12: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Classical null hypotheses

Can we reduce the family P to a subfamily P0, where P0 ⊂ P?

The positive claim can be written by means of a null hypothesis:

H0: “at least one measure in P0 could generate theobserved data”

(or simply H0 : “P ∈ P0” )

Under a parametric model, there exists a finite dimensional setΘ such that:

P ≡ Pθ : θ ∈ Θ , where Θ ⊆ Rp , p <∞,

H 0 : θ ∈ Θ0 , where Θ0 ⊂ Θ and P0 ≡ Pθ : θ ∈ Θ0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 6 / 40

Page 13: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Classical null hypotheses

Can we reduce the family P to a subfamily P0, where P0 ⊂ P?

The positive claim can be written by means of a null hypothesis:

H0: “at least one measure in P0 could generate theobserved data”

(or simply H0 : “P ∈ P0” )

Under a parametric model, there exists a finite dimensional setΘ such that:

P ≡ Pθ : θ ∈ Θ , where Θ ⊆ Rp , p <∞,

H 0 : θ ∈ Θ0 , where Θ0 ⊂ Θ and P0 ≡ Pθ : θ ∈ Θ0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 6 / 40

Page 14: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Classical null hypotheses

Can we reduce the family P to a subfamily P0, where P0 ⊂ P?

The positive claim can be written by means of a null hypothesis:

H0: “at least one measure in P0 could generate theobserved data”

(or simply H0 : “P ∈ P0” )

Under a parametric model, there exists a finite dimensional setΘ such that:

P ≡ Pθ : θ ∈ Θ , where Θ ⊆ Rp , p <∞,

H 0 : θ ∈ Θ0 , where Θ0 ⊂ Θ and P0 ≡ Pθ : θ ∈ Θ0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 6 / 40

Page 15: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Alternative hypotheses

According to Fisher, the negation of H0 cannot be expressedin terms of probability measures.

The alternative hypothesis H1 makes sense if we are certain

about the family P : H1 : P ∈ (P − P0) .

In the last context, we can choose1 between H0 and H1 —Neyman and Pearson approach.

1since they would be mutually exclusive and exhaustiveAlexandre G. Patriota ([email protected]) On some assumptions of NHST 7 / 40

Page 16: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Alternative hypotheses

According to Fisher, the negation of H0 cannot be expressedin terms of probability measures.

The alternative hypothesis H1 makes sense if we are certain

about the family P : H1 : P ∈ (P − P0) .

In the last context, we can choose1 between H0 and H1 —Neyman and Pearson approach.

1since they would be mutually exclusive and exhaustiveAlexandre G. Patriota ([email protected]) On some assumptions of NHST 7 / 40

Page 17: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Alternative hypotheses

According to Fisher, the negation of H0 cannot be expressedin terms of probability measures.

The alternative hypothesis H1 makes sense if we are certain

about the family P : H1 : P ∈ (P − P0) .

In the last context, we can choose1 between H0 and H1 —Neyman and Pearson approach.

1since they would be mutually exclusive and exhaustiveAlexandre G. Patriota ([email protected]) On some assumptions of NHST 7 / 40

Page 18: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Bayesian hypotheses

Recall the hard core Bayesian approach, where P = P0 andZ = (X , γ).

The general null and alternative hypotheses are:

H0 : “γ ∈ Γ0” and H1 : “γ ∈ (Γ− Γ0)” .

The focus is not on the family of probability measures P , sinceP0 is given.

A classical statistician may also test Bayesian hypotheses. Ratherthan p-values, they would use estimated conditional probabilities.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 8 / 40

Page 19: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Bayesian hypotheses

Recall the hard core Bayesian approach, where P = P0 andZ = (X , γ).

The general null and alternative hypotheses are:

H0 : “γ ∈ Γ0” and H1 : “γ ∈ (Γ− Γ0)” .

The focus is not on the family of probability measures P , sinceP0 is given.

A classical statistician may also test Bayesian hypotheses. Ratherthan p-values, they would use estimated conditional probabilities.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 8 / 40

Page 20: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Bayesian hypotheses

Recall the hard core Bayesian approach, where P = P0 andZ = (X , γ).

The general null and alternative hypotheses are:

H0 : “γ ∈ Γ0” and H1 : “γ ∈ (Γ− Γ0)” .

The focus is not on the family of probability measures P , sinceP0 is given.

A classical statistician may also test Bayesian hypotheses. Ratherthan p-values, they would use estimated conditional probabilities.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 8 / 40

Page 21: On some assumptions of Null Hypothesis Statistical Testing

Hypothesis testing

Bayesian hypotheses

Recall the hard core Bayesian approach, where P = P0 andZ = (X , γ).

The general null and alternative hypotheses are:

H0 : “γ ∈ Γ0” and H1 : “γ ∈ (Γ− Γ0)” .

The focus is not on the family of probability measures P , sinceP0 is given.

A classical statistician may also test Bayesian hypotheses. Ratherthan p-values, they would use estimated conditional probabilities.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 8 / 40

Page 22: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-value definition

The p-value for testing the classical null hypothesis H0 is definedas follows

p(P0, x ) = supP∈P0

P(TH0(X ) > TH0(x )

)where TH0 is a statistic such that the more discrepant is H0

from x , the larger is its observed value.2

p(P0, x ) ≈ 0 indicates that the best case in H0 provides a smallprobability to more “extreme events” than the observed one.

2i.e., TH0 could be −2 log of the likelihood-ratio statistic.Alexandre G. Patriota ([email protected]) On some assumptions of NHST 9 / 40

Page 23: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-value definition

The p-value for testing the classical null hypothesis H0 is definedas follows

p(P0, x ) = supP∈P0

P(TH0(X ) > TH0(x )

)where TH0 is a statistic such that the more discrepant is H0

from x , the larger is its observed value.2

p(P0, x ) ≈ 0 indicates that the best case in H0 provides a smallprobability to more “extreme events” than the observed one.

2i.e., TH0 could be −2 log of the likelihood-ratio statistic.Alexandre G. Patriota ([email protected]) On some assumptions of NHST 9 / 40

Page 24: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-value limitations

Consider two null hypotheses H0 : “P ∈ P0” and H ′0 : “P ∈ P ′0”such that H0 =⇒ H ′0. Then, we would expect that:

P ′0

P0

| |p(P0, x ) p(P ′0, x )

But it is not always the case!

The previous p-value is not monotone over the set of nullhypotheses/Sets.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 10 / 40

Page 25: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-value limitations

Consider two null hypotheses H0 : “P ∈ P0” and H ′0 : “P ∈ P ′0”such that H0 =⇒ H ′0. Then, we would expect that:

P ′0

P0

| |p(P0, x ) p(P ′0, x )

But it is not always the case!

The previous p-value is not monotone over the set of nullhypotheses/Sets.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 10 / 40

Page 26: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-value limitations

Consider two null hypotheses H0 : “P ∈ P0” and H ′0 : “P ∈ P ′0”such that H0 =⇒ H ′0. Then, we would expect that:

P ′0

P0

| |p(P0, x ) p(P ′0, x )

But it is not always the case!

The previous p-value is not monotone over the set of nullhypotheses/Sets.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 10 / 40

Page 27: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statisticunder H0 : µ = 0 is

TH0(X ) = nX>X ∼ χ22,

under H ′0 : µ1 = µ2 is

TH ′0(X ) =

n

2(X1 − X2)2 ∼ χ2

1,

where X = (X1, X2)> is the maximum likelihood estimator forµ.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 11 / 40

Page 28: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statisticunder H0 : µ = 0 is

TH0(X ) = nX>X ∼ χ22,

under H ′0 : µ1 = µ2 is

TH ′0(X ) =

n

2(X1 − X2)2 ∼ χ2

1,

where X = (X1, X2)> is the maximum likelihood estimator forµ.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 11 / 40

Page 29: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statisticunder H0 : µ = 0 is

TH0(X ) = nX>X ∼ χ22,

under H ′0 : µ1 = µ2 is

TH ′0(X ) =

n

2(X1 − X2)2 ∼ χ2

1,

where X = (X1, X2)> is the maximum likelihood estimator forµ.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 11 / 40

Page 30: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

P-values do not respect monotonicity

Observed sample H0 : µ = 0 H ′0 : µ1 = µ2

(x1, x2) x1 − x2 p-value p-value(0.05,-0.05) 0.1 0.9753 0.8231(0.09,-0.11) 0.2 0.9039 0.6547(0.14,-0.16) 0.3 0.7977 0.5023(0.19,-0.21) 0.4 0.6697 0.3711(0.23,-0.27) 0.5 0.5331 0.2636(0.28,-0.32) 0.6 0.4049 0.1797(0.33,-0.37) 0.7 0.2926 0.1175(0.37,-0.43) 0.8 0.2001 0.0736(0.42,-0.48) 0.9 0.1308 0.0442(0.47,-0.53) 1.0 0.0813 0.0253

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 12 / 40

Page 31: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Level curves (contour curves)

−1.0 −0.5 0.0 0.5 1.0

−1.

0−

0.5

0.0

0.5

1.0

Significance level 10%

µ1

µ 2

Region of rejection

Region of rejection

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 13 / 40

Page 32: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Level curves (contour curves)

−1.0 −0.5 0.0 0.5 1.0

−1.

0−

0.5

0.0

0.5

1.0

Significance level 10%

µ1

µ 2

Region of rejection

Region of rejection

H ′0 : µ1 = µ2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 14 / 40

Page 33: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Level curves (contour curves)

−1.0 −0.5 0.0 0.5 1.0

−1.

0−

0.5

0.0

0.5

1.0

Significance level 10%

µ1

µ 2

Region of rejection

Region of rejection

H ′0 : µ1 = µ2

H0 : µ = 0

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 15 / 40

Page 34: On some assumptions of Null Hypothesis Statistical Testing

P-value definition and its limitations

Level curves (contour curves)

−1.0 −0.5 0.0 0.5 1.0

−1.

0−

0.5

0.0

0.5

1.0

Significance level 10%

µ1

µ 2

Region of rejection

Region of rejection

H ′0 : µ1 = µ2

H0 : µ = 0

Rejection of H ′0 butno rejection of H0

and

⇓non-coherent conclusion

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 16 / 40

Page 35: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

An alternative measure of evidence (parametric

case)

In what follows, we present an alternative measure called s-valueto overcome the previous issue (Patriota, 2013, FSS, 233).

The s-value is a function s : 2Θ ×X → [0, 1] such that

s(Θ0, x ) =

supα ∈ (0, 1) : Λα(x ) ∩Θ0 6= ∅, if Θ0 6= ∅,0, if Θ0 = ∅.

where Λα is a confidence set for θ with confidence level 1− αwith some “nice” properties.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 17 / 40

Page 36: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

An alternative measure of evidence (parametric

case)

In what follows, we present an alternative measure called s-valueto overcome the previous issue (Patriota, 2013, FSS, 233).

The s-value is a function s : 2Θ ×X → [0, 1] such that

s(Θ0, x ) =

supα ∈ (0, 1) : Λα(x ) ∩Θ0 6= ∅, if Θ0 6= ∅,0, if Θ0 = ∅.

where Λα is a confidence set for θ with confidence level 1− αwith some “nice” properties.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 17 / 40

Page 37: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Interpretation

Interpretation: s = s(Θ0, x ) is the largest significance level α(or 1− s is the smallest confidence level 1− α) for which theconfidence set and the set Θ0 have at least one element incommon.

Large values of s indicate that there exists at least oneelement in Θ0 close to the center of Λα (e.g., close to the MLestimate).

Small values of s indicate that ALL elements of Θ0 are far awayfrom the center of Λα.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 18 / 40

Page 38: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Interpretation

Interpretation: s = s(Θ0, x ) is the largest significance level α(or 1− s is the smallest confidence level 1− α) for which theconfidence set and the set Θ0 have at least one element incommon.

Large values of s indicate that there exists at least oneelement in Θ0 close to the center of Λα (e.g., close to the MLestimate).

Small values of s indicate that ALL elements of Θ0 are far awayfrom the center of Λα.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 18 / 40

Page 39: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Interpretation

Interpretation: s = s(Θ0, x ) is the largest significance level α(or 1− s is the smallest confidence level 1− α) for which theconfidence set and the set Θ0 have at least one element incommon.

Large values of s indicate that there exists at least oneelement in Θ0 close to the center of Λα (e.g., close to the MLestimate).

Small values of s indicate that ALL elements of Θ0 are far awayfrom the center of Λα.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 18 / 40

Page 40: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Graphical illustration: s1 = s(Θ1, x )

−1.5 −1.0 −0.5 0.0 0.5 1.0 1.5

−1.

5−

1.0

−0.

50.

00.

51.

01.

5

θ1

θ 2

Θ1

Θ2

Λs1

θ

Θ

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 19 / 40

Page 41: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Graphical illustration: s2 = s(Θ2, x )

−1.5 −1.0 −0.5 0.0 0.5 1.0 1.5

−1.

5−

1.0

−0.

50.

00.

51.

01.

5

θ1

θ 2

Θ1

Θ2

Λs1

θ

Θ

Λs2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 20 / 40

Page 42: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Properties: the s-value is a possibility measure

1 s(∅, x ) = 0 and s(Θ, x ) = 1,

2 If Θ1 ⊆ Θ2, then s(Θ1, x ) ≤ s(Θ2, x ),

3 For any Θ1,Θ2 ⊆ Θ, s(Θ1 ∪Θ2, x ) = maxs(Θ1, x ), s(Θ2, x ),

4 If Θ1 ⊆ Θ, then s(Θ1, x ) = supθ∈Θ1s(θ, x ),

5 s(Θ1, x ) = 1 or s(Θc1, x ) = 1:

if θ ∈ Θ1 (closure of Θ1), then s(Θ1, x ) = 1,

if θ ∈ Θc1, then s(Θc

1, x ) = 1.

where θ is an element of the center of Λα, i.e., θ ∈⋂

α Λα(x ).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 21 / 40

Page 43: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Properties: the s-value is a possibility measure

1 s(∅, x ) = 0 and s(Θ, x ) = 1,

2 If Θ1 ⊆ Θ2, then s(Θ1, x ) ≤ s(Θ2, x ),

3 For any Θ1,Θ2 ⊆ Θ, s(Θ1 ∪Θ2, x ) = maxs(Θ1, x ), s(Θ2, x ),

4 If Θ1 ⊆ Θ, then s(Θ1, x ) = supθ∈Θ1s(θ, x ),

5 s(Θ1, x ) = 1 or s(Θc1, x ) = 1:

if θ ∈ Θ1 (closure of Θ1), then s(Θ1, x ) = 1,

if θ ∈ Θc1, then s(Θc

1, x ) = 1.

where θ is an element of the center of Λα, i.e., θ ∈⋂

α Λα(x ).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 21 / 40

Page 44: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Properties: the s-value is a possibility measure

1 s(∅, x ) = 0 and s(Θ, x ) = 1,

2 If Θ1 ⊆ Θ2, then s(Θ1, x ) ≤ s(Θ2, x ),

3 For any Θ1,Θ2 ⊆ Θ, s(Θ1 ∪Θ2, x ) = maxs(Θ1, x ), s(Θ2, x ),

4 If Θ1 ⊆ Θ, then s(Θ1, x ) = supθ∈Θ1s(θ, x ),

5 s(Θ1, x ) = 1 or s(Θc1, x ) = 1:

if θ ∈ Θ1 (closure of Θ1), then s(Θ1, x ) = 1,

if θ ∈ Θc1, then s(Θc

1, x ) = 1.

where θ is an element of the center of Λα, i.e., θ ∈⋂

α Λα(x ).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 21 / 40

Page 45: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Properties: the s-value is a possibility measure

1 s(∅, x ) = 0 and s(Θ, x ) = 1,

2 If Θ1 ⊆ Θ2, then s(Θ1, x ) ≤ s(Θ2, x ),

3 For any Θ1,Θ2 ⊆ Θ, s(Θ1 ∪Θ2, x ) = maxs(Θ1, x ), s(Θ2, x ),

4 If Θ1 ⊆ Θ, then s(Θ1, x ) = supθ∈Θ1s(θ, x ),

5 s(Θ1, x ) = 1 or s(Θc1, x ) = 1:

if θ ∈ Θ1 (closure of Θ1), then s(Θ1, x ) = 1,

if θ ∈ Θc1, then s(Θc

1, x ) = 1.

where θ is an element of the center of Λα, i.e., θ ∈⋂

α Λα(x ).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 21 / 40

Page 46: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Properties: the s-value is a possibility measure

1 s(∅, x ) = 0 and s(Θ, x ) = 1,

2 If Θ1 ⊆ Θ2, then s(Θ1, x ) ≤ s(Θ2, x ),

3 For any Θ1,Θ2 ⊆ Θ, s(Θ1 ∪Θ2, x ) = maxs(Θ1, x ), s(Θ2, x ),

4 If Θ1 ⊆ Θ, then s(Θ1, x ) = supθ∈Θ1s(θ, x ),

5 s(Θ1, x ) = 1 or s(Θc1, x ) = 1:

if θ ∈ Θ1 (closure of Θ1), then s(Θ1, x ) = 1,

if θ ∈ Θc1, then s(Θc

1, x ) = 1.

where θ is an element of the center of Λα, i.e., θ ∈⋂

α Λα(x ).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 21 / 40

Page 47: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Decisions about H0

Let Φ be a function such that:

Φ(Θ0) = 〈s(Θ0), s(Θc0)〉.

Then,

Φ(Θ0) = 〈a, 1〉 =⇒ rejection of H0 if a is “small” enough,

Φ(Θ0) = 〈1, b〉 =⇒ acceptance of H0 if b is“small”enough.

Φ(Θ0) = 〈1, 1〉 =⇒ total ignorance about H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 22 / 40

Page 48: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Decisions about H0

Let Φ be a function such that:

Φ(Θ0) = 〈s(Θ0), s(Θc0)〉.

Then,

Φ(Θ0) = 〈a, 1〉 =⇒ rejection of H0 if a is “small” enough,

Φ(Θ0) = 〈1, b〉 =⇒ acceptance of H0 if b is“small”enough.

Φ(Θ0) = 〈1, 1〉 =⇒ total ignorance about H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 22 / 40

Page 49: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

Decisions about H0

Let Φ be a function such that:

Φ(Θ0) = 〈s(Θ0), s(Θc0)〉.

Then,

Φ(Θ0) = 〈a, 1〉 =⇒ rejection of H0 if a is “small” enough,

Φ(Θ0) = 〈1, b〉 =⇒ acceptance of H0 if b is“small”enough.

Φ(Θ0) = 〈1, 1〉 =⇒ total ignorance about H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 22 / 40

Page 50: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

How to find the thresholds for a and b to decide

about H0?

This is still an open problem.

We could try to find those thresholds via loss functions.

or via frequentist criteria by employing the following asymptoticproperty:

Property: If the statistical model is regular and the confidenceregion is built from a statistics Tθ(X ) that converges indistribution to χ2

k , then:

sa = 1− Fk(F−1H0

(1− pa)),

where pa = 1− FH0(t) is the asymptotic p-value to test H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 23 / 40

Page 51: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

How to find the thresholds for a and b to decide

about H0?

This is still an open problem.

We could try to find those thresholds via loss functions.

or via frequentist criteria by employing the following asymptoticproperty:

Property: If the statistical model is regular and the confidenceregion is built from a statistics Tθ(X ) that converges indistribution to χ2

k , then:

sa = 1− Fk(F−1H0

(1− pa)),

where pa = 1− FH0(t) is the asymptotic p-value to test H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 23 / 40

Page 52: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

How to find the thresholds for a and b to decide

about H0?

This is still an open problem.

We could try to find those thresholds via loss functions.

or via frequentist criteria by employing the following asymptoticproperty:

Property: If the statistical model is regular and the confidenceregion is built from a statistics Tθ(X ) that converges indistribution to χ2

k , then:

sa = 1− Fk(F−1H0

(1− pa)),

where pa = 1− FH0(t) is the asymptotic p-value to test H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 23 / 40

Page 53: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

How to find the thresholds for a and b to decide

about H0?

This is still an open problem.

We could try to find those thresholds via loss functions.

or via frequentist criteria by employing the following asymptoticproperty:

Property: If the statistical model is regular and the confidenceregion is built from a statistics Tθ(X ) that converges indistribution to χ2

k , then:

sa = 1− Fk(F−1H0

(1− pa)),

where pa = 1− FH0(t) is the asymptotic p-value to test H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 23 / 40

Page 54: On some assumptions of Null Hypothesis Statistical Testing

An alternative measure of evidence and some of its properties

How to find the thresholds for a and b to decide

about H0?

This is still an open problem.

We could try to find those thresholds via loss functions.

or via frequentist criteria by employing the following asymptoticproperty:

Property: If the statistical model is regular and the confidenceregion is built from a statistics Tθ(X ) that converges indistribution to χ2

k , then:

sa = 1− Fk(F−1H0

(1− pa)),

where pa = 1− FH0(t) is the asymptotic p-value to test H0.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 23 / 40

Page 55: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statistic is

Tµ(x ) = n(X − µ)>(X − µ) ∼ χ22,

The confidence set Λα is given by

Λα(x ) = µ ∈ R2 : Tµ(x ) ≤ F−12 (1− α),

where F2 is the cumulative chi-squared distribution with twodegrees of freedom.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 24 / 40

Page 56: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statistic is

Tµ(x ) = n(X − µ)>(X − µ) ∼ χ22,

The confidence set Λα is given by

Λα(x ) = µ ∈ R2 : Tµ(x ) ≤ F−12 (1− α),

where F2 is the cumulative chi-squared distribution with twodegrees of freedom.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 24 / 40

Page 57: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Example: Bivariate Normal distribution

Let X = (X1, . . . ,Xn) be a sample from a bivariate normaldistribution with mean vector µ = (µ1, µ2)> and identityvariance matrix.

Notice that (−2 log of) the likelihood-ratio statistic is

Tµ(x ) = n(X − µ)>(X − µ) ∼ χ22,

The confidence set Λα is given by

Λα(x ) = µ ∈ R2 : Tµ(x ) ≤ F−12 (1− α),

where F2 is the cumulative chi-squared distribution with twodegrees of freedom.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 24 / 40

Page 58: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Numerical illustration

Observed sample H0 : µ = 0 H ′0 : µ1 = µ2

(x1, x2) x1 − x2 p/s-value p-value s-value

(0.05,-0.05) 0.1 0.9753 0.8231 0.9753(0.09,-0.11) 0.2 0.9039 0.6547 0.9048(0.14,-0.16) 0.3 0.7977 0.5023 0.7985(0.19,-0.21) 0.4 0.6697 0.3711 0.6703(0.23,-0.27) 0.5 0.5331 0.2636 0.5353(0.28,-0.32) 0.6 0.4049 0.1797 0.4066(0.33,-0.37) 0.7 0.2926 0.1175 0.2938(0.37,-0.43) 0.8 0.2001 0.0736 0.2019(0.42,-0.48) 0.9 0.1308 0.0442 0.1320(0.47,-0.53) 1.0 0.0813 0.0253 0.0821

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 25 / 40

Page 59: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Numerical illustration

Observed sample H0 : µ = 0 H ′0 : µ1 = µ2

(x1, x2) x1 − x2 p/s-value p-value s-value(0.05,-0.05) 0.1 0.9753 0.8231 0.9753

(0.09,-0.11) 0.2 0.9039 0.6547 0.9048(0.14,-0.16) 0.3 0.7977 0.5023 0.7985(0.19,-0.21) 0.4 0.6697 0.3711 0.6703(0.23,-0.27) 0.5 0.5331 0.2636 0.5353(0.28,-0.32) 0.6 0.4049 0.1797 0.4066(0.33,-0.37) 0.7 0.2926 0.1175 0.2938(0.37,-0.43) 0.8 0.2001 0.0736 0.2019(0.42,-0.48) 0.9 0.1308 0.0442 0.1320(0.47,-0.53) 1.0 0.0813 0.0253 0.0821

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 26 / 40

Page 60: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Numerical illustration

Observed sample H0 : µ = 0 H ′0 : µ1 = µ2

(x1, x2) x1 − x2 p/s-value p-value s-value(0.05,-0.05) 0.1 0.9753 0.8231 0.9753(0.09,-0.11) 0.2 0.9039 0.6547 0.9048

(0.14,-0.16) 0.3 0.7977 0.5023 0.7985(0.19,-0.21) 0.4 0.6697 0.3711 0.6703(0.23,-0.27) 0.5 0.5331 0.2636 0.5353(0.28,-0.32) 0.6 0.4049 0.1797 0.4066(0.33,-0.37) 0.7 0.2926 0.1175 0.2938(0.37,-0.43) 0.8 0.2001 0.0736 0.2019(0.42,-0.48) 0.9 0.1308 0.0442 0.1320(0.47,-0.53) 1.0 0.0813 0.0253 0.0821

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 27 / 40

Page 61: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Numerical illustration

Observed sample H0 : µ = 0 H ′0 : µ1 = µ2

(x1, x2) x1 − x2 p/s-value p-value s-value(0.05,-0.05) 0.1 0.9753 0.8231 0.9753(0.09,-0.11) 0.2 0.9039 0.6547 0.9048(0.14,-0.16) 0.3 0.7977 0.5023 0.7985(0.19,-0.21) 0.4 0.6697 0.3711 0.6703(0.23,-0.27) 0.5 0.5331 0.2636 0.5353(0.28,-0.32) 0.6 0.4049 0.1797 0.4066(0.33,-0.37) 0.7 0.2926 0.1175 0.2938(0.37,-0.43) 0.8 0.2001 0.0736 0.2019(0.42,-0.48) 0.9 0.1308 0.0442 0.1320

(0.47,-0.53) 1.0 0.0813 0.0253 0.0821

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 28 / 40

Page 62: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x1) = 0.9753

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 29 / 40

Page 63: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x2) = 0.9048

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 30 / 40

Page 64: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x3) = 0.7985

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 31 / 40

Page 65: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x4) = 0.6703

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 32 / 40

Page 66: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x5) = 0.5353

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 33 / 40

Page 67: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x6) = 0.4066

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 34 / 40

Page 68: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x7) = 0.2938

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 35 / 40

Page 69: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x8) = 0.2019

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 36 / 40

Page 70: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x9) = 0.1320

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 37 / 40

Page 71: On some assumptions of Null Hypothesis Statistical Testing

Numerical illustration

Graphical illustration: s(µ1 = µ2, x10) = 0.0821

−0.2 0.0 0.2 0.4 0.6 0.8

−0.

8−

0.6

−0.

4−

0.2

0.0

0.2

µ1

µ 2

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 38 / 40

Page 72: On some assumptions of Null Hypothesis Statistical Testing

Final remarks

Final remarks

The s-value:

can be applied directly whenever the log-likelihood function isconcave by the formula s = 1− F (FH0(1− p))

is a possibilistic measure and can be studied by means of theAbstract belief Calculus ABC (Darwiche, Ginsberg, 1992).

can be justified by desiderata (more basic axioms).

avoids the p-value problem of non-monotonicity.

is a classic alternative to the FBST (Pereira, Stern, 1998).

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 39 / 40

Page 73: On some assumptions of Null Hypothesis Statistical Testing

References

References:

Darwiche, A.Y., Ginsberg, M.L. (1992). A symbolic generalization ofprobability theory, AAAI-92, Tenth National Conference onArtificial Intelligence.

Patriota, AG (2017). On some assumptions of Null HypothesisStatistical Testing, Educational and Psychological Measurement,77, 507–524.

Patriota, AG (2013). A classical measure of evidence for general nullhypotheses, Fuzzy Sets and Systems, 233, 74-88.

Pereira, C.A.B., Stern, J.M. (1999). Evidence and credibility: FullBayesian significance test for precise hypotheses, Entropy, 1,99–110.

Alexandre G. Patriota ([email protected]) On some assumptions of NHST 40 / 40