categorical data

77
Categorical Data Ziad Taib Biostatistics AstraZeneca February 2014

Upload: wenda

Post on 22-Feb-2016

115 views

Category:

Documents


4 download

DESCRIPTION

Categorical Data. Ziad Taib Biostatistics AstraZeneca February 2014. Types of Data. Continuous Blood pressure Time to event. Categorical sex. quantitative. qualitative. Discrete No of relapses. Ordered Categorical Pain level. Types of data analysis (Inference). Parametric Vs - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Categorical  Data

Categorical Data

Ziad TaibBiostatisticsAstraZeneca

February 2014

Page 2: Categorical  Data

Types of Data

ContinuousBlood pressureTime to event

Ordered CategoricalPain level

DiscreteNo of relapses

Categoricalsex

quantitative qualitative

Page 3: Categorical  Data

Types of data analysis (Inference)

ParametricVs

Non parametric

FrequentistVs

Bayesian

Model basedVs

Data driven

Page 4: Categorical  Data

Categorical data

In a RCT, endpoints and surrogate endpoints can be categorical or ordered categorical variables. In the simplest cases we have binary responses (e.g. responders non-responders). In Outcomes research it is common to use many ordered categories (no improvement, moderate improvement, high improvement).

Page 5: Categorical  Data

Binary variables

• Sex• Mortality• Presence/absence of an AE• Responder/non-responder according to

some pre-defined criteria• Success/Failure

Page 6: Categorical  Data

Inference problems

1. Binary data (proportions)• One sample• Paired data

2. Ordered categorical data3. Combining categorical data4. Logistic regression5. A Bayesian alternative6. ODDS and NNT

Page 7: Categorical  Data

Categorical data

In a RCT, endpoints and surrogate endpoints can be categorical or ordered categorical variables. In the simplest cases we have binary responses (e.g. responders non-responders). In Outcomes research it is common to use many ordered categories (no improvement, moderate improvement, high improvement).

Page 8: Categorical  Data

Bernoulli experiment

Randomexperience

Failure0

Success1

Hole in one?

With probability p

With probability1-p

Page 9: Categorical  Data

Binary variables

• Sex• Mortality• Presence/absence of an AE• Responder/non-responder according to

some pre-defined criteria• Success/Failure

Page 10: Categorical  Data

Estimation• Assume that a treatment has been

applied to n patients and that at the end of the trial they were classified according to how they responded to the treatment: 0 meaning not cured and 1 meaning cured. The data at hand is thus a sample of n independent binary variables

• The probability of being cured by this treatment can be estimated by

satisfying

Page 11: Categorical  Data

Hypothesis testing• We can test the null hypothesis

• Using the test statistic

• When n is large, Z follows, under the null hypothesis, the standard normal distribution (obs! Not when p very small or very large).

Page 12: Categorical  Data

Hypothesis testing• For moderate values of n we can use the

exact Bernoulli distribution of leading to the sum being Binomially distributed i.e.

• As with continuous variables, tests can be used to build confidence intervals.

Page 13: Categorical  Data

Example 1: Hypothesis test based on binomial distr.

Consider testing H0: P=0.5against Ha: P>0.5

and where: n=10 and y=number of successes=8

p-value=(probability of obtaining a result at least as extreme as the one observed)=Prob(8 or more responders)=P8+ P9+ P10=={using the binomial formula}=0.0547

Page 14: Categorical  Data

Example 2

RCT of two analgesic drugs A and B given in a random order to each of 100 patients. After both treatment periods, each patient states a preference for one of the drugs.

Result: 65 patients preferred A and 35 B

Page 15: Categorical  Data

Example (cont’d)

Hypotheses: H0: P=0.5 against H1: P0.5

Observed test-statistic: z=2.90

p-value: p=0.0037(exact p-value using the binomial distr. = 0.0035)

95% CI for P: (0.56 ; 0.74)

Page 16: Categorical  Data

Two proportions• Sometimes we want to compare the proportion of

successes in two separate groups. For this purpose we take two samples of sizes n1 and n2. We let yi1 and pi1 be the observed number of subjects and the proportion of successes in the ith group. The difference in population proportions of successes and its large sample variance can be estimated by

Page 17: Categorical  Data

Two proportions (continued)• Assume we want to test the null hypothesis that

there is no difference between the proportions of success in the two groups (p11=p12). Under the null hypothesis, we can estimate the common proportion by

• Its large sample variance is estimated by

• Leading to the test statistic

Page 18: Categorical  Data

Example

In a trial for acute ischemic stroke

Treatment n responders*rt-PA 312 147 (47.1%)placebo 312 122 (39.1%)

*early improvement defined on a neurological scale

Point estimate: 0.080 (s.e.=0.0397)

95% CI: (0.003 ; 0.158)

p-value: 0.043

Page 19: Categorical  Data

Two proportions (Chi square)• The problem of comparing two proportions

can sometimes be formulated as a problem of independence! Assume we have two groups as above (treatment and placebo). Assume further that the subjects were randomized to these groups. We can then test for independence between belonging to a certain group and the clinical endpoint (success or failure). The data can be organized in the form of a contingency table in which the marginal totals and the total number of subjects are considered as fixed.

Page 20: Categorical  Data

Failure Success Total

Drug 165 147 312

Placebo 190 122 312

Total 355 462 N=624

R E S P O N S E

TREATMENT

2 x 2 Contingency table

Page 21: Categorical  Data

Failure Success Total

Drug Y10 Y11 Y1.

Placebo Y20 Y21 Y2.

Total Y.0 Y.1 N=Y..

R E S P O N S E

TREATMENT

2 x 2 Contingency table

Page 22: Categorical  Data

Hyper geometric distribution

Urn containing W white balls and R red balls: N=W+R

• n balls are drawn at random without replacement.

• Y is the number of white balls (successes)

• Y follows the Hyper geometric Distribution with parameters (N, W, n)

Page 23: Categorical  Data

Contingency tables

• N subjects in total• y.1 of these are special (success)• y1. are drawn at random • Y11 no of successes among these y1. • Y11 is HG(N,y.1,y 1.)

in general

Page 24: Categorical  Data

Contingency tables

• The null hypothesis of independence is tested using the chi square statistic

• Which, under the null hypothesis, is chi square distributed with one degree of freedom provided the sample sizes in the two groups are large (over 30) and the expected frequency in each cell is non negligible (over 5)

Page 25: Categorical  Data

Contingency tables• For moderate sample sizes we use Fisher’s exact

test. According to this calculate the desired probabilities using the exact Hyper-geometric distribution. The variance can then be calculated. To illustrate consider:

• Using this and expectation m11 we have the randomization chi square statistic. With fixed margins only one cell is allowed to vary. Randomization is crucial for this approach.

Page 26: Categorical  Data

The (Pearson) Chi-square test

35 contingency table

The Chi-square test is used for testing the independence between the two factors

Other factorA B C D E

i niA niB niC niD niE ni

One Factor ii niiA niiB niiC niiD niiE nii

iii niiiA niiiB niiiC niiiD niiiE niii

nA nB nC nD nE niA

Page 27: Categorical  Data

The (Pearson) Chi-square test

The test-statistic is:

i j ij

2ijij2

E)E(O

where Oij = observed frequencies

and Eij = expected frequencies (under independence)

the test-statistic approximately follows a chi-square distribution

p

Page 28: Categorical  Data

Example 5Chi-square test for a 22 table

Examining the independence between two treatments and a classification into responder/non-responder is equivalent to comparing the proportion of responders in the two groups

NINDS again non-resp responderrt-PA 165 147 312placebo 190 122 312

355 269Observed frequencies

non-resp responderrt-PA 177.5 134.5 312placebo 177.5 134.5 312

355 269

Expected frequencies

Page 29: Categorical  Data

• p0=(122+147)/(324)=0.43• v(p0)=0.00157

which gives a p-value of 0.043 in all these cases. This implies the drug is better than placebo. However when using Fisher’s exact test or using a continuity correction the chi square test the p-value is 0.052.

Page 30: Categorical  Data

TABLE OF GRP BY Y

Frequency‚ Row Pct ‚nonresp ‚resp ‚ Total ƒƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆ placebo ‚ 190 ‚ 122 ‚ 312 ‚ 60.90 ‚ 39.10 ‚ ƒƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆ rt-PA ‚ 165 ‚ 147 ‚ 312 ‚ 52.88 ‚ 47.12 ‚ ƒƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆƒƒƒƒƒƒƒƒˆ Total 355 269 624

STATISTICS FOR TABLE OF GRP BY Y

Statistic DF Value Prob ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ Chi-Square 1 4.084 0.043 Likelihood Ratio Chi-Square 1 4.089 0.043 Continuity Adj. Chi-Square 1 3.764 0.052 Mantel-Haenszel Chi-Square 1 4.077 0.043 Fisher's Exact Test (Left) 0.982 (Right) 0.026 (2-Tail) 0.052 Phi Coefficient 0.081 Contingency Coefficient 0.081 Cramer's V 0.081

Sample Size = 624

SAS| output

Page 31: Categorical  Data

Odds, Odds Ratios and relative RisksThe odds of success in group i is estimated by

The odds ratio of success between the two groups i is estimated by

Define risk for success in the ith group as the proportion of cases with success. The relative risk between the two groups is estimated by

Page 32: Categorical  Data

Categorical data• Nominal

– E.g. patient residence at end of follow-up (hospital, nursing home, own home, etc.)

• Ordinal (ordered)– E.g. some global rating

• Normal, not at all ill• Borderline mentally ill• Mildly ill• Moderately ill• Markedly ill• Severely ill• Among the most extremely ill patients

Page 33: Categorical  Data

Categorical data & Chi-square testOther factor

A B C D Ei niA niB niC niD niE ni

One Factor ii niiA niiB niiC niiD niiE nii

iii niiiA niiiB niiiC niiiD niiiE niii

nA nB nC nD nE niA

The chi-square test is useful for detection of a general association between treatment and categorical response (in either the nominal or ordinal scale), but it cannot identify a particular relationship, e.g. a location shift.

Page 34: Categorical  Data

Nominal categorical data Disease category dip snip fup bop other treatment A 33 15 34 26 8 116 group B 28 18 34 20 14 114 61 33 68 46 22 230

Chi-square test: 2 = 3.084 , df=4 , p = 0.544

Page 35: Categorical  Data

Ordered categorical data• Here we assume two groups one receiving the

drug and one placebo. The response is assumed to be ordered categorical with J categories.

• The null hypothesis is that the distribution of subjects in response categories is the same for both groups.

• Again the randomization and the HG distribution lead to the same chi square test statistic but this time with (J-1) df. Moreover the same relationship exists between the two versions of the chi square statistic.

Page 36: Categorical  Data

The Mantel-Haensel statistic The aim here is to combine data from

several (H) strata for comparing two groups drug and placebo. The expected frequency and the variance for each stratum are used to define the Mantel-Haensel statistic

which is chi square distributed with one df.

Page 37: Categorical  Data

• Consider again the Bernoulli situation, where Y is a binary r.v. (success or failure) with p being the success probability. Sometimes Y can depend on some other factors or covariates. Since Y is binary we cannot use usual regression.

Logistic regression

Page 38: Categorical  Data

Logistic regression• Logistic regression is part of a category of statistical

models called generalized linear models (GLM). This broad class of models includes ordinary regression and ANOVA, as well as multivariate statistics such as ANCOVA and loglinear regression. An excellent treatment of generalized linear models is presented in Agresti (1996).

• Logistic regression allows one to predict a discrete outcome, such as group membership, from a set of variables that may be continuous, discrete, dichotomous, or a mix of any of these. Generally, the dependent or response variable is dichotomous, such as presence/absence or success/failure.

Page 39: Categorical  Data

Simple linear regression

Age SBP Age SBP Age SBP 22 131 41 139 52 128 23 128 41 171 54 105 24 116 46 137 56 145 27 106 47 111 57 141 28 114 48 115 58 153 29 123 49 133 59 157 30 117 49 128 63 155 32 122 50 183 67 176 33 99 51 130 71 172 35 121 51 133 77 178 40 147 51 144 81 217

Table 1 Age and systolic blood pressure (SBP) among 33 adult women

Page 40: Categorical  Data

80

100

120

140

160

180

200

220

20 30 40 50 60 70 80 90

SBP (mm Hg)

Age (years)

adapted from Colton T. Statistics in Medicine. Boston: Little Brown, 1974

Age1.22281.54SBP

Page 41: Categorical  Data

Simple linear regression• Relation between 2 continuous variables (SBP

and age)

• Regression coefficient b1– Measures association between y and x– Amount by which y changes on average when x

changes by one unit– Least squares method

y

x

xβαy 11Slope

Page 42: Categorical  Data

Multiple linear regression• Relation between a continuous variable and a

set of i continuous variables

• Partial regression coefficients bi

– Amount by which y changes on average when xi changes by one unit and all the other xis remain constant

– Measures association between xi and y adjusted for all other xi

• Example– SBP versus age, weight, height, etc

xβ ... xβ xβαy ii2211

Page 43: Categorical  Data

Multiple linear regression

Predicted Predictor variables

Response variable Explanatory variables

Outcome variable CovariablesDependent Independent

variables

xβ ... xβ xβα y ii2211

Page 44: Categorical  Data

Logistic regression

Age CD Age CD Age CD

22 0 40 0 54 0 23 0 41 1 55 1 24 0 46 0 58 1 27 0 47 0 60 1 28 0 48 0 60 0 30 0 49 1 62 1 30 0 49 0 65 1 32 0 50 1 67 1 33 0 51 0 71 1 35 1 51 1 77 1 38 0 52 0 81 1

Table 2 Age and signs of coronary heart disease (CD)

Page 45: Categorical  Data

How can we analyse these data?

• Compare mean age of diseased and non-diseased

– Non-diseased: 38.6 years– Diseased: 58.7 years (p<0.0001)

• Linear regression?

Page 46: Categorical  Data

Dot-plot: Data from Table 2

AGE (years)

Sig

ns o

f cor

onar

y di

seas

e

No

Yes

0 20 40 60 80 100

Page 47: Categorical  Data

Logistic regression (2)Table 3 Prevalence (%) of signs of CD according

to age group Diseased

Age group # in group # %

20 - 29 5 0 0

30 - 39 6 1 17

40 - 49 7 2 29

50 - 59 7 4 57

60 - 69 5 4 80

70 - 79 2 2 100

80 - 89 1 1 100

Page 48: Categorical  Data

Dot-plot: Data from Table 3

0

20

40

60

80

100

0 2 4 6 8

Diseased %

Age group

Page 49: Categorical  Data

Logistic function (1)

0,0

0,2

0,4

0,6

0,8

1,0Probability of disease

x

P y x ee

x

x( )

b

b1

Page 50: Categorical  Data

ln( )

( )P y xP y x

x1

b

Transformation

)()(xyP

xyP1

logit of P(y|x)

{P y x e

e

x

x( )

b

b1

0,0

0,2

0,4

0,6

0,8

1,0

Page 51: Categorical  Data

Fitting equation to the data

• Linear regression: Least squares or Maximum likelihood

• Logistic regression: Maximum likelihood• Likelihood function

– Estimates parameters and b – Practically easier to work with log-likelihood

n

iiiii xyxylL

1

)(1ln)1()(ln)(ln)(

Page 52: Categorical  Data

Maximum likelihood• Iterative computing (Newton-Raphson)

– Choice of an arbitrary value for the coefficients (usually 0)

– Computing of log-likelihood– Variation of coefficients’ values– Reiteration until maximisation (plateau)

• Results– Maximum Likelihood Estimates (MLE) for

and b– Estimates of P(y) for a given value of x

Page 53: Categorical  Data

Multiple logistic regression• More than one independent variable

– Dichotomous, ordinal, nominal, continuous …

• Interpretation of bi – Increase in log-odds for a one unit increase in

xi with all the other xis constant– Measures association between xi and log-

odds adjusted for all other xi

ii2211 xβ ... xβ xβαP-1

P ln

Page 54: Categorical  Data

Statistical testing

• Question– Does model including given independent

variable provide more information about dependent variable than model without this variable?

• Three tests– Likelihood ratio statistic (LRS)– Wald test– Score test

Page 55: Categorical  Data

Likelihood ratio statistic

• Compares two nested models Log(odds) = + b1x1 + b2x2 + b3x3 (model 1) Log(odds) = + b1x1 + b2x2 (model 2)

• LR statistic-2 log (likelihood model 2 / likelihood model 1) =-2 log (likelihood model 2) minus -2log (likelihood model 1)LR statistic is a 2 with DF = number of extra parameters in model

Page 56: Categorical  Data

Example 6

Fitting a Logistic regression model to the NINDS data, using only one covariate (treatment group).

NINDS again non-resp responderrt-PA 165 147 312placebo 190 122 312

355 269Observed frequencies

Page 57: Categorical  Data

SAS| output

The LOGISTIC Procedure Response Profile Ordered Binary Value Outcome Count 1 EVENT 269 2 NO EVENT 355 Model Fitting Information and Testing Global Null Hypothesis BETA=0 Intercept Intercept and Criterion Only Covariates Chi-Square for Covariates AIC 855.157 853.069 . SC 859.593 861.941 . -2 LOG L 853.157 849.069 4.089 with 1 DF (p=0.0432) Score . . 4.084 with 1 DF (p=0.0433) Analysis of Maximum Likelihood Estimates Parameter Standard Wald Pr > Standardized Odds Variable DF Estimate Error Chi-Square Chi-Square Estimate Ratio INTERCPT 1 -0.4430 0.1160 14.5805 0.0001 . . GRP 1 0.3275 0.1622 4.0743 0.0435 0.090350 1.387

Page 58: Categorical  Data

Logistic regression example

• AZ trial (CLASS) in acute stroke comparing clomethiazole (n=678) with placebo (n=675)

• Response defined as a Barthel Index score 60 at 90 days

• Covariates:– STRATUM (time to start of trmt: 0-6, 6-12)– AGE– SEVERITY (baseline SSS score)– TRT (treatment group)

Page 59: Categorical  Data

SAS| output

Response Profile

Ordered Value BI_60 Count

1 1 750 2 0 603

Analysis of Maximum Likelihood Estimates

Parameter Standard Wald Pr > Standardized Odds Variable DF Estimate Error Chi-Square Chi-Square Estimate Ratio

INTERCPT 1 2.4244 0.5116 22.4603 0.0001 TRT 1 0.1299 0.1310 0.9838 0.3213 0.035826 1.139 STRATUM 1 0.1079 0.1323 0.6648 0.4149 0.029751 1.114 AGE 1 -0.0673 0.00671 100.6676 0.0001 -0.409641 0.935 SEVERITY 1 0.0942 0.00642 215.0990 0.0001 0.621293 1.099

Conditional Odds Ratios and 95% Confidence Intervals

Wald Confidence Limits Odds Variable Unit Ratio Lower Upper

TRT 1.0000 1.139 0.881 1.472 STRATUM 1.0000 1.114 0.859 1.444 AGE 1.0000 0.935 0.923 0.947 SEVERITY 1.0000 1.099 1.085 1.113

Page 60: Categorical  Data

Risk

Odds and NNT

Page 61: Categorical  Data

David Brennan (former) CEO of AstraZeneca

Page 62: Categorical  Data

?

Page 63: Categorical  Data
Page 65: Categorical  Data

Risk, Odds and probability

• Risk or chance can be measures by:– p= probabilities– O=Odds = p/(1-p)

• Probability and odds contain the same information and are equally valid as measures of chance. In case of infrequent events the distinction is unimportant. For frequent events they can be quite different.

Page 66: Categorical  Data

4 measures of association (effect)

– Quite often we are interested in risk and probabilty only as a way to measure association or effect:

Risk is lower in the drug group= cure is associated with drug = the drug has an effect

– This can be done in different ways1. Absolute Risk (Prospective Studies)2. Relative Risk (Prospective Studies) 3. Odds Ratio (Prospective or Retrospective)4. (Number Needed to Treat) (Prospective Studies)

Page 67: Categorical  Data

Here the 20% represents the relative risk, derived by

• 1.6% = risk of breast cancer reported for women who took aspirin. • 2.0% = risk of breast cancer in women not taking aspirin.

Thus, the absolute risk reduction was 2.0% – 1.6%% = 0.4%, which translates to 4 fewer cases per 1,000 women.

But the relative risk reduction = 1 – 1.6%/2.0% = 0.20 so the impressive 20% figure was advertised.

Computations of relative risk reduction will always appear impressively large when actual event rates are low.

• An article in a leading medical journal reporting on a population-based case-control study of breast cancer stated that “women who used aspirin over a five-year period had a 20% reduction in breast cancer”.

Page 68: Categorical  Data

• Man använder relativa tal för gynnsamma effekter, så att de ser stora ut, och absoluta tal för skadliga effekter, så att de ser små ut. Detta knep använder sig Törnberg och Nyström också av när de säger att den positiva effekten är 30 procent, medan den skadliga effekten drabbar endast 0,1 procent av kvinnorna …… Detta är manipulation när den är som värst …

• »Om man undersöker 2 000 kvinnor med mammografi regelbundet i 10 år, kommer 1 av dem att ha nytta av mammografin eftersom hon kommer att undgå att dö av bröstcancer. Samtidigt kommer 10 friska kvinnor av de 2 000 deltagarna på grund av mammografin att få diagnosen bröstcancer och bli behandlade i onödan«.

Page 69: Categorical  Data

The Relative Risk Versus p-value

• A big relative risk does not necessarily mean that the p-value is small

• The big relative risk could have occurred by chance if the sample size were small

• The p-value depends both on the magnitude of the relative risk as well as the sample size.

Page 70: Categorical  Data

Relative risk: Fractions are funny• Assume we computed 2.5 as the relative risk. In this calculation we divided

the probability for group 1 by its counterpart for group 2. • If we had divided the group 2 probability by the group 1 probability, we would

have obtained a relative risk of 0.4. This is fine because 0.4 (2/5) and 2.5 (5/2) are reciprocal fractions.

0.8 (4/5) 1.25 (5/4)

0.75 (3/4)

0.67 (2/3) 1.50 (3/2)

0.50 (1/2)

1.33 (4/3)

2.00 (2/1)

Reciprocal fractions

Page 71: Categorical  Data

The scale of the odds ratios:• The variability of the odds ratio is typically

based in the logarithmic scale due to assymetric character of odds ratio.

• This is especially important when deriving confidence intervals

Log(OR)

OR1

0

20.50

0.693-0.693

Page 72: Categorical  Data

On Estonia: • There were 485 female passengers: 14 survived and 471 died. • There were 504 male passengers: 80 survived and 424 died

Alive Dead Total

Female 14 471 485Male 80 424 504Total 94 895 989

For females, the odds were 33 to 1 against surviving (471/14=33.65). For males, the odds were almost 5 to 1 in favor of death (424/80=5.3). The odds ratio is 0.15 (5.3/33.5). There is a 6.6 fold greater odds of death for females than for males.

Odds ratio vs. relative risk

Page 73: Categorical  Data

• The relative risk: For females, the probability of death is 97% (471/485=0.97). For males, the probability is 84% (424/504=0.84). The relative risk of death is 0.87 (0.84/0.97 =1/1.15 ). There is a 1.15 greater probability of death for females than for males.

• There is quite a difference. Both measurements show that women were more likely to die. But the odds ratio (6.6) implies that women are much worse off than the relative risk 1.15. Which number is a fairer comparison?

There are three issues here:

– (I) The relative risk measures events in a way that is interpretable and consistent with the way “people” really think.

– (II) The relative risk, though, cannot always be computed in a research design.

– (III) Also, the relative risk can sometimes lead to ambiguous and confusing conclusions.

Page 74: Categorical  Data

RELATIVE RISK AND ODDS RATIO

• Suppose there are two groups, one with a 25% chance of mortality and the other with a 50% chance of mortality. Most people would say that the latter group has it twice as bad. But the odds ratio is 3, which seems too big. 0.25/(1-0.25) / 0.50/(1-0.50) = 1/3

• Even more extreme examples are possible. A change from 25% to 75% mortality represents a relative risk of 3, but an odds ratio of 9.

Page 75: Categorical  Data
Page 76: Categorical  Data

RELATIVE RISK vs. ODDS RATIO

• Why would we even bother calculating the odds ratio when we can calculate relative risk?

• The odds ratio turns out to be important because you can calculate it either in cohort studies or case-control studies

• The relative risk can only be calculated from cohort studies

• OR can be close to the relative risk: If the disease is very rare, it is almost exactly the relative risk

• How rare? Rough rule of thumb: If proportion with disease is less than 1%. If disease is not rare, OR still follows same direction as RR, but may not be a very accurate estimate of RR

Page 77: Categorical  Data

Number needed to treat: NNT

• NNT should be completed with follow up period and unfavourable event avoided.

• NNT presupposes that there is statistically significant difference (*).

• How much NNT is good? No magic figure: (10-500) risky surgerey – standard inexpensive drug with no side effect – active treatment – preventive treatment etc.

• Statistical properties? Confidence intervals?

• When AR = 0, NNT becomes infinite!

• The distribution of NNT is complicated because its behaviour around AR = 0;

• The moments of NNT do not exist;

• Simple calculations with NNT like addition can give nonsensical results.

2

^

1 ˆˆ

11

ARNNT