chapter 24 single-factor (one-way) analysis of variance

Post on 18-Dec-2021

10 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

3/18/2013

1

Chapter 24

Single-Factor (One-Way)

Analysis of Variance (ANOVA)

and

Analysis of Means (ANOM)

Introduction

โ€ข This chapter describes single-factor analysis of

variance (ANOVA) experiments with 2 or more levels

(or treatments).

โ€ข The method is based on a fixed effects model (as

opposed to a random effects model, or components of

variance model). It tests the ๐ป0 that the different

processes give an equal response.

โ€ข With fixed effects model, the levels are specifically

chosen. The test hypothesis is about the mean

responses due to factor levels. Conclusions apply only

to the factor levels considered.

3/18/2013

2

24.1 S4/IEE Application Examples:

ANOVA and ANOM

โ€ข Transactional 30,000-foot-level metric: DSO reduction was

chosen as an S4/IEE project. A cause-and-effect matrix

ranked company as an important input that could affect the

DSO response (i.e., the team thought that some

companies were more delinquent in payments than other

companies). From randomly sampled data, a statistical

assessment was conducted to test the hypothesis of

equality of means for the DSOs of these companies.

24.1 S4/IEE Application Examples:

ANOVA and ANOM

โ€ข Manufacturing 30,000-foot-level metric (KPOV): An S4/IEE

project was to improve the capability/performance of the

diameter of a manufactured product (i.e., reduce the

number of parts beyond the specification limits). A cause-

and-effect matrix ranked cavity of the four-cavity mold as

an important input that could be yielding different part

diameters. From randomly sampled data, statistical tests

were conducted to test the hypotheses of mean diameter

equality and equality of variances for the cavities.

3/18/2013

3

24.1 S4/IEE Application Examples:

ANOVA and ANOM

โ€ข Transactional and manufacturing 30.000-foot-level cycle

time metric (a lean metric): An S4/IEE project was to

improve the time from order entry to fulfillment. The WIP

at each process step was collected at the end of the day

for a random number of days. Statistical tests were

conducted to test the hypothesis that the mean and

variance of WIP at each step was equal.

24.2 Application Steps

1. Describe the problem using a response variable that

corresponds to the KPOV or measured quality characteristic.

2. Describe the analysis (e.g., determine if there is a difference)

3. State the null and alternative hypotheses.

4. Choose a large enough sample and conduct the experiment

randomly.

5. Generate an ANOVA table.

6. Test the data normality and equality of variance assumptions.

7. Make hypothesis decisions about factors from ANOVA table.

8. Calculate (if desired) epsilon squared (๐œ€2).

9. Conduct an analysis of means (ANOM).

10. Translate conclusions into terms relevant to the problem or

process in question.

3/18/2013

4

24.3 Single-factor ANOVA

Hypothesis Test

โ€ข ANOVA assesses the differences between samples taken

at different factor levels to determine if these differences

are large enough relative to error to conclude that the

factor level causes a statistically significant difference in

response.

โ€ข For a single-factor analysis of variance, a linear statistical

model can describe the observations of a level with ๐‘— observations taken under level ๐‘– (๐‘– = 1,2,โ€ฆ , ๐‘Ž; ๐‘— = 1,2,โ€ฆ , ๐‘›):

๐‘ฆ๐‘–๐‘— = ๐œ‡ + ๐œ๐‘– + ๐œ€๐‘–๐‘— where ๐‘ฆ๐‘–๐‘— is the (๐‘–๐‘—)th observation, ๐œ‡ is the overall mean, ๐œ๐‘– is

the ๐‘–th level effect, and ๐œ€๐‘–๐‘— is random error.

24.3 Single-factor ANOVA

Hypothesis Test

โ€ข In ANOVA test, model errors are assumed to be normally

and independently distributed random variables with mean

0 and variance ๐œŽ2. The variance is assumed constant for

all factor levels.

โ€ข An expression for the hypothesis test of means is:

๐ป0: ๐œ‡1 = ๐œ‡2 = โ‹ฏ = ๐œ‡๐‘Ž

๐ป๐‘Ž: ๐œ‡๐‘– โ‰  ๐œ‡๐‘— for at least one pair (๐‘–, ๐‘—)

โ€ข When ๐ป0 is true, all levels have a common mean ๐œ‡, which

leads to an equivalent expression in terms of ๐œ๐‘–.

๐ป0: ๐œ1 = ๐œ2 = โ‹ฏ = ๐œ๐‘Ž = 0

๐ป๐‘Ž: ๐œ๐‘– โ‰  0 (for at least one ๐‘–)

3/18/2013

5

24.4 Single-factor ANOVA Table

Calculations

โ€ข The total sum of squares of deviations about the grand

average ๐‘ฆ (also referred as the total corrected sum of

squares) represents the overall variability of the data:

๐‘†๐‘†total = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ )2๐‘›

๐‘—=1

๐‘Ž

๐‘–=1

โ€ข A division of ๐‘†๐‘†total by the number of degrees of freedom

would yield a sample variance of ๐‘ฆโ€™s. For this situation, the

overall number of degrees of freedom is ๐‘Ž๐‘› โˆ’ 1 = ๐‘ โˆ’ 1.

24.4 Single-factor ANOVA Table

Calculations

โ€ข Total variability in data, as measured by ๐‘†๐‘†total (the total

corrected sum of squares) can be partitioned into a sum of

two elements. The first element is the sum of squares for

differences between factor level averages and the grand

average. The second element is the sum of squares of the

differences of observations within factor levels from the

average of factorial levels. The first element is a measure

of the difference between the means of the levels, whereas

the second one is due to random error.

๐‘†๐‘†total = ๐‘†๐‘†factor levels + ๐‘†๐‘†error

3/18/2013

6

24.4 Single-factor ANOVA Table

Calculations

โ€ข ๐‘†๐‘†factor levels is called the sum of squares due to factor

levels (i.e., between factor levels or treatments)

๐‘†๐‘†factor levels = ๐‘› (๐‘ฆ ๐‘– โˆ’ ๐‘ฆ )2๐‘Ž

๐‘–=1

โ€ข ๐‘†๐‘†error is called the sum of squares due to error (i.e., within

factor levels or treatments)

๐‘†๐‘†error = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ ๐‘–)2

๐‘›

๐‘—=1

๐‘Ž

๐‘–=1

โ€ข When divided by the appropriate number of degrees of

freedom, these SSs give good estimates

24.4 Single-factor ANOVA Table

Calculations

โ€ข When divided by the appropriate number of degrees of

freedom, these SSs give good estimates of the total

variability, the variability between factor levels, and the

variability within factor levels (or error).

โ€ข Expressions for the mean square are

๐‘€๐‘†factor levels =๐‘€๐‘†factor levels

๐‘Ž โˆ’ 1

๐‘€๐‘†error =๐‘†๐‘†error

๐‘Ž(๐‘› โˆ’ 1)=

๐‘†๐‘†error

๐‘ โˆ’ ๐‘Ž

3/18/2013

7

24.4 Single-factor ANOVA Table

Calculations

โ€ข If there is no difference in treatment means, the two

estimates are presumed to be similar. If there is a

difference, we suspect that the observed difference is

caused by differences in the treatment (factor) levels.

โ€ข F-test statistic tests ๐ป0: there is no difference in factor

levels.

๐น0 =๐‘€๐‘†factor levels

๐‘€๐‘†error

โ€ข ๐ป0 should be rejected if ๐น0 > ๐น๐›ผ,๐‘Žโˆ’1,๐‘โˆ’๐‘Ž.

โ€ข Alternatively, a ๐‘-value could be calculated for ๐น0. ๐ป0 should

be rejected if ๐‘ โˆ’ value > ๐›ผ.

Source of

Variation

Sum of

Squares

Degrees of

Freedom

Mean

Square ๐‘ญ๐ŸŽ

Between-

factor levels ๐‘†๐‘†factor levels ๐‘Ž โˆ’ 1 ๐‘€๐‘†factor levels ๐น0 =

๐‘€๐‘†factor levels

๐‘€๐‘†error

Error

(within-

factor levels)

๐‘†๐‘†error ๐‘ โˆ’ ๐‘Ž ๐‘€๐‘†error

Total ๐‘†๐‘†total ๐‘ โˆ’ 1

24.4 Single-factor ANOVA Table

Calculations

3/18/2013

8

24.5 Estimation of Model

Parameters

โ€ข In addition to factor-level significance, it is useful to

estimate the parameters of the single-factor model and the

CI on the factor-level means.

โ€ข For the single-factor model

๐‘ฆ๐‘–๐‘— = ๐œ‡ + ๐œ๐‘– + ๐œ€๐‘–๐‘— Estimates for the overall mean and factor-level effects are

๐œ‡ = ๐‘ฆ

๐œ ๐‘– = ๐‘ฆ ๐‘– โˆ’ ๐‘ฆ , ๐‘– = 1,2,โ€ฆ , ๐‘Ž

โ€ข A 100(1-)% CI on the ๐‘–th factor level is

๐‘ฆ ๐‘– ยฑ ๐‘ก๐›ผ,๐‘โˆ’๐‘Ž

๐‘€๐‘†error

๐‘›

24.6 Unbalanced Data

โ€ข A design is considered unbalanced when the number of

observations in the factor level is different. For this

situation, ANOVA equations need slight modification

๐‘†๐‘†factor levels = ๐‘›๐‘–(๐‘ฆ ๐‘– โˆ’ ๐‘ฆ )2๐‘Ž

๐‘–=1

โ€ข A balanced design is preferable to an unbalanced design.

With a balanced design, the power of the test in maximized

and the test statistic is robust to small departure from the

assumption of equal variances.

3/18/2013

9

24.7 Model Adequacy

โ€ข As in the regression model, valid ANOVA requires that

certain assumptions be satisfied.

โ€ข One typical assumption is that errors are normally and

independently distributed with mean 0 and constant but

unknown variance ๐‘๐ผ๐ท(0, ๐œŽ2). โ€ข To help with meeting the independence and normal

distribution requirement, an experimenter needs to select

an adequate sample size and randomly conduct the trials.

โ€ข After data are collected, computer programs offer routines

to test the assumptions.

โ€ข Generally, in a fixed effects ANOVA, moderate departures

from normality of the residuals are of little concern.

24.7 Model Adequacy

โ€ข In addition to an analysis of residuals, there is also a direct

statistical test for equality of variance.

๐ป0: ๐œŽ12 = ๐œŽ2

2 = โ‹ฏ = ๐œŽ๐‘Ž2

๐ป๐‘Ž: above not true for at least one ๐œŽ๐‘–2

โ€ข Bartlettโ€™s test is frequently used to test this hypothesis

when the normality assumption is valid. Leveneโ€™s test can

be used when the normality assumption is questionable.

3/18/2013

10

24.8 Analysis of Residuals: Fitted Value Plots and Data Transformations

โ€ข Residual plots should show no structure relative to any

factor included in the fitted response; however, trends in

the data may occur for various reasons.

โ€ข One phenomenon that may occur is inconsistent variance.

Fortunately, a balanced fixed effects model is robust to

variance not being homogeneous.

โ€ข A data transformation may then be used to reduce this

phenomenon in the residuals, which would yield a more

precise significance test.

24.8 Analysis of Residuals: Fitted Value Plots and Data Transformations

โ€ข Another situation occurs when the output is count data,

where a square root transformation may be appropriate,

while a lognormal transformation is often appropriate if the

trial outputs are standard deviation values and a logit might

be helpful when there are upper and lower limits.(Table 24.2)

โ€ข As an alternative to the transformations included in the table,

Box (I988) describes a method for eliminating unnecessary

coupling of dispersion effects and location effects by

determining an approximate transformation using a lambda

plot.

โ€ข With transformations, the conclusions of the analysis apply to

the transformed populations.

3/18/2013

11

24.8 Analysis of Residuals: Fitted Value Plots and Data Transformations

Data Characteristics Data (๐’™๐’Š ๐’๐’“ ๐’‘๐’Š) Transformation

๐œŽ โˆ ๐‘๐‘œ๐‘›๐‘ ๐‘ก๐‘Ž๐‘›๐‘ก None

๐œŽ โˆ ๐œ‡2 1 ๐‘ฅ๐‘–

๐œŽ โˆ ๐œ‡3/2 1 ๐‘ฅ๐‘–

๐œŽ โˆ ๐œ‡ log ๐‘ฅ๐‘–

๐œŽ โˆ ๐œ‡, Poisson (count) data ๐‘ฅ๐‘– or ๐‘ฅ๐‘– + 1

Binomial proportions ๐‘ ๐‘–๐‘›โˆ’1( ๐‘๐‘–)

Upper- and lower-bound data

(e.g., 0~1 probability of failure)

Logit transformation:

๐‘™๐‘œ๐‘”๐‘ฅ๐‘– โˆ’ ๐‘™๐‘œ๐‘ค๐‘’๐‘Ÿ ๐‘™๐‘–๐‘š๐‘–๐‘ก

๐‘ข๐‘๐‘๐‘’๐‘Ÿ ๐‘™๐‘–๐‘š๐‘–๐‘ก โˆ’ ๐‘ฅ๐‘–

24.9 Comparing Pairs of

Treatment Means

โ€ข The rejection of the null hypothesis in an ANOVA indicates

that there is a difference between the factor levels

(treatments). However, no information is given to determine

which means are different.

โ€ข Sometimes it is useful to make further comparisons and

analysis among groups of factor level means. Multiple

comparison methods assess differences between treatment

means in either the factor level totals or the factor level

averages.

โ€ข Methods include those of Tukey and Fisher. Montgomery

(I997) describes several methods of making comparisons.

โ€ข The analysis of means (ANOM) approach is to compare

individual means to a grand mean.

3/18/2013

12

24.10 Example 24.1:

Single-Factor ANOVA

โ€ข The bursting strengths of diaphragms were determined in an

experiment. Use analysis of variance techniques to determine

if there is a statistically significant difference at a level of 0.05.

Type 1 Type 2 Type 3 Type 4 Type 5 Type 6 Type 7

59.0 65.7 65.3 67.9 60.6 73.1 59.4

62.3 62.8 63.7 67.4 65.0 71.9 61.6

65.2 59.1 68.9 62.9 68.2 67.8 56.3

65.5 60.2 70.0 61.7 66.0 67.4 62.7

24.10 Example 24.1:

Single-Factor ANOVA

Minitab:

Stat

ANOVA

One-Way

Graphs

Boxplot

3/18/2013

13

24.10 Example 24.1:

Single-Factor ANOVA

Minitab:

Stat

ANOVA

One-Way

Graphs

Ind. values

24.10 Example 24.1:

Single-Factor ANOVA

Minitab:

Stat

ANOVA

One-Way

One-way ANOVA: Strength versus Type

Source DF SS MS F P

Type 6 265.34 44.22 4.92 0.003

Error 21 188.71 8.99

Total 27 454.05

S = 2.998 R-Sq = 58.44% R-Sq(adj) = 46.57%

Individual 95% CIs For Mean Based on Pooled StDev

Level N Mean StDev ------+---------+---------+---------+---

1 4 63.000 3.032 (-----*-----)

2 4 61.950 2.942 (-----*-----)

3 4 66.975 2.966 (-----*-----)

4 4 64.975 3.134 (-----*-----)

5 4 64.950 3.193 (-----*-----)

6 4 70.050 2.876 (-----*-----)

7 4 60.000 2.823 (-----*-----)

------+---------+---------+---------+---

60.0 65.0 70.0 75.0

Pooled StDev = 2.998

3/18/2013

14

24.10 Example 24.1:

Single-Factor ANOVA

Minitab:

ANOVA

Test of equal variances

24.10 Example 24.1:

Single-Factor ANOVA

Minitab:

Stat

ANOVA

One-Way

Graphs

Four in One

3/18/2013

15

24.11 Analysis of Means (ANOM)

โ€ข Analysis of means (ANOM) is a statistical test procedure in

a graphical format.

Group 1 2 3 โ‹ฏ ๐’Œ

๐‘ฅ11 ๐‘ฅ21 ๐‘ฅ31 โ‹ฏ ๐‘ฅ๐‘˜1

๐‘ฅ12 ๐‘ฅ22 ๐‘ฅ32 โ‹ฏ ๐‘ฅ๐‘˜2

๐‘ฅ13 ๐‘ฅ23 ๐‘ฅ33 โ‹ฏ ๐‘ฅ๐‘˜3

โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ

๐‘ฅ1๐‘— ๐‘ฅ2๐‘— ๐‘ฅ3๐‘— โ‹ฏ ๐‘ฅ๐‘˜๐‘—

๐‘ฅ 1 ๐‘ฅ 2 ๐‘ฅ 3 โ‹ฏ ๐‘ฅ ๐‘˜

๐‘ 1 ๐‘ 2 ๐‘ 3 โ‹ฏ ๐‘ ๐‘˜

24.11 Analysis of Means (ANOM)

โ€ข The grand mean ๐‘ฅ is simply the average of the group

means (๐‘ฅ ๐‘–).

๐‘ฅ = ๐‘ฅ ๐‘–

๐‘˜๐‘–=1

๐‘˜

โ€ข The pooled estimate for the standard deviation is as follows

๐‘  = ๐‘ ๐‘–

2๐‘˜๐‘–=1

๐‘˜

โ€ข The lower and upper decision lines (LDL and UDL) are

๐ฟ๐ท๐ฟ = ๐‘ฅ โˆ’ โ„Ž๐›ผ๐‘ ๐‘˜ โˆ’ 1

๐‘˜๐‘› ๐‘ˆ๐ท๐ฟ = ๐‘ฅ + โ„Ž๐›ผ๐‘ 

๐‘˜ โˆ’ 1

๐‘˜๐‘›

3/18/2013

16

24.11 Analysis of Means (ANOM)

โ€ข The lower and upper decision lines (LDL and UDL) are

๐ฟ๐ท๐ฟ = ๐‘ฅ โˆ’ โ„Ž๐›ผ๐‘ ๐‘˜ โˆ’ 1

๐‘˜๐‘› ๐‘ˆ๐ท๐ฟ = ๐‘ฅ + โ„Ž๐›ผ๐‘ 

๐‘˜ โˆ’ 1

๐‘˜๐‘›

โ€ข โ„Ž๐›ผ is from Table I for risk level ๐›ผ, number of means ๐‘˜, and

degrees of freedom ๐‘› โˆ’ 1 ๐‘˜ .

โ€ข The means are then plotted against the decision lines. If

any mean fall outside the decision lines, there is a

statistically significant difference for this mean from the

grand mean.

โ€ข If normality can be assumed (๐‘›๐‘ > 5 ๐‘Ž๐‘›๐‘‘ ๐‘› 1 โˆ’ ๐‘ > 5),

ANOM is also directly applicable to attribute data.

24.12 Example 24.2:

Analysis of Means (ANOM)

โ€ข The bursting strengths of diaphragms were determined in an

experiment. Use analysis of variance techniques to determine

if there is a statistically significant difference at a level of 0.05.

Type 1 Type 2 Type 3 Type 4 Type 5 Type 6 Type 7 Sum 59.0 65.7 65.3 67.9 60.6 73.1 59.4

62.3 62.8 63.7 67.4 65.0 71.9 61.6

65.2 59.1 68.9 62.9 68.2 67.8 56.3

65.5 60.2 70.0 61.7 66.0 67.4 62.7

Mean 63.0 62.0 67.0 65.0 65.0 70.1 60.0 451.9

Var. 9.19 8.66 8.80 9.82 10.20 8.27 7.97 62.9

3/18/2013

17

โ€ข The grand mean ๐‘ฅ is

๐‘ฅ = ๐‘ฅ ๐‘–

๐‘˜๐‘–=1

๐‘˜=

451.9

7= 64.6

โ€ข The pooled standard deviation is

๐‘  = ๐‘ ๐‘–2

๐‘˜๐‘–=1

๐‘˜=

62.90

7= 3.00

โ€ข The lower and upper decision lines (LDL and UDL) are

๐ฟ๐ท๐ฟ = ๐‘ฅ โˆ’ โ„Ž๐›ผ๐‘ ๐‘˜ โˆ’ 1

๐‘˜๐‘›= 64.6 โˆ’ 2.94 3.0

7 โˆ’ 1

7 4= 60.49

๐‘ˆ๐ท๐ฟ = 68.65

24.12 Example 24.2:

ANOM of Injection-Molding Data

24.12 Example 24.2:

ANOM of Injection-Molding Data

Minitab:

Stat

ANOVA

Analysis of Means

3/18/2013

18

24.13 Example 24.3:

Analysis of Means (ANOM)

Example 15.1

Example 22.4

โ€ข A comparison of the proportion of total variability of the

factor levels to the error term could be made in % units

using a controversial epsilon square relationship:

๐œ€factor level2 = 100 ร—

๐‘†๐‘†factor

๐‘†๐‘†total

๐œ€error2 = 100 ร—

๐‘†๐‘†error

๐‘†๐‘†total

โ€ข Consider the situation in which a process was randomly

sampled using conventional, rational sampling practices.

Consider also that there were between 25 and 100 sets of

samples taken over time.

24.14 Six Sigma Considerations*

3/18/2013

19

โ€ข For this type of data, the sums of squares from an ANOVA

table can be used to break down total variability into 2 parts.

โ€ข The division of these sums of squares by the correct

number of degrees of freedom yields estimates for the

different source of variation (total, between subgroups, and

within subgroups).

โ€ข The estimator of total variability gives an estimate for LT

capability, while the estimator of within group variability

gives an estimate for ST capability.

โ€ข These concepts of variability can be used to represent the

influence of time on a process.

24.14 Six Sigma Considerations*

โ€ข The ST and LT standard deviation estimates from an

ANOVA table are

๐œŽ ๐ฟ๐‘‡ = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ )2๐‘›

๐‘—=1๐‘Ž๐‘–=1

๐‘›๐‘Ž โˆ’ 1

๐œŽ ๐‘†๐‘‡ = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ ๐‘–)2

๐‘›๐‘—=1

๐‘Ž๐‘–=1

๐‘Ž(๐‘› โˆ’ 1)

โ€ข These two estimators are useful in calculating the ST and

LT capability / performance of the process.

24.14 Six Sigma Considerations*

3/18/2013

20

โ€ข The variable used to measure this capability / performance

is Z. Short-term Z values for the process are

๐‘LSL,ST =๐ฟ๐‘†๐ฟ โˆ’ ๐‘‡

๐œŽ ๐‘†๐‘‡ ๐‘USL,ST =

๐‘ˆ๐‘†๐ฟ โˆ’ ๐‘‡

๐œŽ ๐‘†๐‘‡

โ€ข The nominal specification ๐‘‡ value is used because it

represents the potential capability of the process.

โ€ข Long-term Z values for the process are

๐‘LSL,LT =๐ฟ๐‘†๐ฟ โˆ’ ๐œ‡

๐œŽ ๐ฟ๐‘‡ ๐‘USL,ST =

๐‘ˆ๐‘†๐ฟ โˆ’ ๐œ‡

๐œŽ ๐ฟ๐‘‡

24.14 Six Sigma Considerations*

โ€ข Probability values can then be obtained from the normal

distribution for the different values of ๐‘.

โ€ข These probabilities correspond to the frequency of

occurrence beyond specification limits.

โ€ข Multiplication of these probabilities by one million gives

DPMO.

24.14 Six Sigma Considerations*

3/18/2013

21

24.15 Example 24.4:

Determining Process Capability Using

One-Factor ANOVA # ๐‘ฅ R

1 0.65 0.70 0.65 0.65 0.85 0.70 0.20

2 0.75 0.85 0.75 0.85 0.65 0.77 0.20

3 0.75 0.80 0.80 0.70 0.75 0.76 0.10

4 0.60 0.70 0.70 0.75 0.65 0.68 0.15

5 0.70 0.75 0.65 0.85 0.80 0.75 0.20

6 0.60 0.75 0.75 0.85 0.70 0.73 0.25

7 0.75 0.80 0.65 0.75 0.70 0.73 0.15

8 0.60 0.70 0.80 0.75 0.75 0.72 0.20

9 0.65 0.80 0.85 0.85 0.75 0.78 0.20

10 0.60 0.70 0.60 0.80 0.65 0.67 0.20

11 0.80 0.75 0.90 0.50 0.80 0.75 0.40

12 0.85 0.75 0.85 0.65 0.70 0.76 0.20

13 0.70 0.70 0.75 0.75 0.70 0.72 0.05

14 0.65 0.70 0.85 0.75 0.60 0.71 0.25

15 0.90 0.80 0.80 0.75 0.85 0.82 0.15

16 0.75 0.80 0.75 0.80 0.65 0.75 0.15

โ€ข Example 11.2 Process

capability / performance

metrics

โ€ข Example 22.3 Variance

components

24.15 Example 24.4:

Determining Process Capability Using

One-Factor ANOVA

One-way ANOVA: Data versus Subgroup

Source DF SS MS F P

Subgroup 15 0.10950 0.00730 1.12 0.360

Error 64 0.41800 0.00653

Total 79 0.52750

๐œŽ ๐ฟ๐‘‡ = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ )2๐‘›

๐‘—=1๐‘Ž๐‘–=1

๐‘›๐‘Ž โˆ’ 1=

0.52750

5 16 โˆ’ 1= 0.081714

๐œŽ ๐‘†๐‘‡ = (๐‘ฆ๐‘–๐‘— โˆ’ ๐‘ฆ ๐‘–)2

๐‘›๐‘—=1

๐‘Ž๐‘–=1

๐‘Ž(๐‘› โˆ’ 1)=

0.41800

16(5 โˆ’ 1)= 0.080816

3/18/2013

22

๐‘LSL,ST =๐ฟ๐‘†๐ฟ โˆ’ ๐‘‡

๐œŽ ๐‘†๐‘‡=

.5 โˆ’ .7

.080816 = โˆ’2.4747 ๐‘USL,ST=

๐‘ˆ๐‘†๐ฟ โˆ’ ๐‘‡

๐œŽ ๐‘†๐‘‡= 2.4747

๐‘ƒ ๐‘ > ๐‘USL,ST = ๐‘ƒ ๐‘ > 2.4747 = 0.006667

๐‘ƒ ๐‘ < ๐‘LSL,ST = ๐‘ƒ ๐‘ < โˆ’2.4747 = 0.006667

๐‘๐‘†๐‘‡ = 0.006667 + 0.006667 = 0.013335

๐‘LSL,LT =๐ฟ๐‘†๐ฟ โˆ’ ๐œ‡

๐œŽ ๐ฟ๐‘‡=

.5 โˆ’ .7375

.081714= โˆ’2.9065 ๐‘USL,ST =

๐‘ˆ๐‘†๐ฟ โˆ’ ๐œ‡

๐œŽ ๐ฟ๐‘‡= 1.9886

๐‘ƒ ๐‘ > ๐‘USL,LT = ๐‘ƒ ๐‘ > 1.9886 = 0.023373

๐‘ƒ ๐‘ < ๐‘LSL,LT = ๐‘ƒ ๐‘ < โˆ’2.9065 = 0.001828

๐‘๐ฟ๐‘‡ = 0.023373 + 0.001828 = 0.02520

24.15 Example 24.4:

Determining Process Capability Using

One-Factor ANOVA

โ€ข When conducting the 2-sample t-test to compare the

average of two groups, the data in both groups must be

sampled from a normally distributed population. If that

assumption does not hold, the nonparametric Mann-Whitney

test is a better safeguard against drawing wrong conclusions.

โ€ข The Mann-Whitney test compares the medians from two

populations and works when the Y variable is continuous,

discrete-ordinal or discrete-count, and the X variable is

discrete with two attributes. Of course, the Mann-Whitney

test can also be used for normally distributed data, but in that

case it is less powerful than the 2-sample t-test.

24.16 Non-Parametric Estimate:

Mann-Whitney Test Procedure

http://www.isixsigma.com/tools-templates/hypothesis-

testing/making-sense-mann-whitney-test-median-comparison/

3/18/2013

23

Testing Hypothesis on the Difference

of Medians โ€“ Mann Whitney Test

โ€ข Null Hypothesis: ๐ป0: 1 = 2

โ€ข Test statistic: ๐‘ˆ = ๐‘†1 โˆ’๐‘›1(๐‘›1+1)

2

โ€ข ๐ป๐‘Ž: 1 < 2

โ€ข Critical value ๐‘ˆ๐‘ (from Table, with n1, n2, )

โ€ข Reject ๐ป0 if ๐‘ˆ < ๐‘ˆ๐‘

โ€ข Ha: 1 > 2

โ€ข Find ๐‘ˆโ€™ (from Table, with n1, n2, ), ๐‘ˆ๐‘ = ๐‘›1๐‘›2 โˆ’ ๐‘ˆโ€™

โ€ข Reject ๐ป0 if ๐‘ˆ > ๐‘ˆ๐‘

โ€ข ๐ป๐‘Ž: 1 โ‰  2

โ€ข Find ๐‘ˆ๐‘™๐‘œ๐‘ค๐‘’๐‘Ÿ (from Table, with n1, n2, /2), ๐‘ˆ๐‘ข๐‘๐‘๐‘’๐‘Ÿ = ๐‘›1๐‘›2 โˆ’ ๐‘ˆ๐‘™๐‘œ๐‘ค๐‘’๐‘Ÿ

โ€ข Reject ๐ป0 if ๐‘ˆ < ๐‘ˆ๐‘™๐‘œ๐‘ค๐‘’๐‘Ÿ or ๐‘ˆ > ๐‘ˆ๐‘ข๐‘๐‘๐‘’๐‘Ÿ

Where S1 is the sum of ranks of sample 1

Testing Hypothesis on the Difference

of Medians โ€“ Mann Whitney Test

http://www.lesn.appstate.edu/olson/stat_directory/Statistical%20procedu

res/Mann_Whitney%20U%20Test/Mann-Whitney%20Table.pdf

Two-tailed Test

3/18/2013

24

Testing Hypothesis on the Difference

of Medians โ€“ Mann Whitney Test

http://www.lesn.appstate.edu/olson/stat_directory/Statistical%20procedu

res/Mann_Whitney%20U%20Test/Mann-Whitney%20Table.pdf

One-tailed Test

Testing Hypothesis on the

Difference of Medians โ€“ Example

โ€ข =.05

โ€ข Null Hypothesis: H0: 1 = 2

โ€ข Test statistic:

๐‘ˆ = ๐‘†1 โˆ’๐‘›1 ๐‘›1 + 1

2= 29.5 โˆ’

5 5 + 1

2= 14.5

โ€ข Ha: 1 โ‰  2

โ€ข Ulower= 3 (Table VI, with n1=5, n2=6, .025), Uupper = n1n2 โ€“ Ulower = (5)(6)-3 = 27

โ€ข Reject H0 if U<Ulower or U>Uupper

โ€ข Fail to reject H0

X Rank

4.6 B 1 1

8.9 B 2 2

9.5 A 3 3

9.7 A 4 4

10.8 A 5 5

11.1 B 6 6

12.9 A 7.5 7.5

12.9 B 7.5 7.5

13.0 B 9 9

16.0 A 10 10

16.4 B 11 11

29.5 36.5

Model A 9.5 10.8 12.9 16.0 9.7

Model B 12.9 11.1 13.0 16.4 8.9 4.6

3/18/2013

25

Testing Hypothesis on the

Difference of Medians โ€“ Example

Mann-Whitney Test and CI: Model A, Model B

N Median

Model A 5 10.800

Model B 6 12.000

Point estimate for ETA1-ETA2 is -0.050

96.4 Percent CI for ETA1-ETA2 is (-3.502,6.202)

W = 29.5

Test of ETA1 = ETA2 vs ETA1 not = ETA2 is significant at 1.0000

The test is significant at 1.0000 (adjusted for ties)

Minitab:

Stat

Nonparametrics

Mann-Whitney

โ€ข A Kruskal-Wallis test provides an alternative to a one-way

ANOVA. This test is a generalization of Mann-Whitney test.

โ€ข The null hypothesis is all medians are equal. The alternative

hypothesis is the medians are not all equal.

โ€ข For this test, it is assumed that independent random samples

taken from different populations have a continuous

distribution with the same shape.

โ€ข For many distributions, the Kruskal-Wallis test is more

powerful than Moodโ€™s median test, but it is less robust

against outliers.

24.16 Non-Parametric Estimate:

Kruskal-Wallis Test

3/18/2013

26

โ€ข The yield per acre for 4 methods of growing corn

24.17 Example 24.5:

Nonparametric Kruskal-Wallis Test

Method 1 Method 2 Method 3 Method 4

83 91 101 78

91 90 100 82

94 81 91 81

89 83 93 77

89 84 96 79

96 83 95 81

91 88 94 80

92 91 81

90 89

84

24.17 Example 24.5:

Nonparametric Kruskal-Wallis Test

Kruskal-Wallis Test: Yield versus Mothod

Kruskal-Wallis Test on Yield

Mothod N Median Ave Rank Z

1 9 91.00 21.8 1.52

2 10 86.00 15.3 -0.83

3 7 95.00 29.6 3.60

4 8 80.50 4.8 -4.12

Overall 34 17.5

H = 25.46 DF = 3 P = 0.000

H = 25.63 DF = 3 P = 0.000 (adjusted for ties)

Minitab:

Stat

Nonparametrics

Kruskal-Wallis

3/18/2013

27

โ€ข Like the Kruskal-Wallis test, a Moodโ€™s median test (also

called a median test or sign scores test) is a nonparametric

alternative to ANOVA.

โ€ข In this chi-square test, the null hypothesis is the population

medians are equal. The alternative hypothesis is the

medians are not all equal.

โ€ข For this test, it is assumed that independent random samples

taken from different populations have a continuous

distribution with the same shape.

โ€ข Moodโ€™s median test is more robust to outliers than the

Kruskal-Wallis test.

24.18 Non-Parametric Estimate:

Moodโ€™s Median Test

24.19 Example 24.6:

Nonparametric Moodโ€™s Median Test

Minitab:

Stat

Nonparametrics

Moodโ€™s Median Test

Mood Median Test: Yield versus Mothod

Mood median test for Yield

Chi-Square = 17.54 DF = 3 P = 0.001

Individual 95.0% CIs

Mothod N<= N> Median Q3-Q1 ---------+---------+---------+-----

--

1 3 6 91.0 4.0 (--*---)

2 7 3 86.0 7.3 (---*-----)

3 0 7 95.0 7.0 (---*------)

4 8 0 80.5 2.8 (---*)

---------+---------+---------+-----

--

84.0 91.0 98.0

Overall median = 89.0

3/18/2013

28

โ€ข Variability in an experiment can be caused by nuisance

factors in which we have no interest.

โ€ข These nuisance factors are sometimes unknown and not

controlled. Randomization guards against this type of factor

affecting results.

โ€ข In other situation, the nuisance factor is known but not

controlled. When we observe the value of a factor, it can be

compensated for by using analysis of covariance techniques.

โ€ข In yet another situation, the nuisance factor is both known

and controllable. We can systematically eliminate the effect

on comparisons among factor level considerations (i.e.,

treatments) by using a randomized block design.

24.20 Other Considerations

โ€ข Experiment results can often be improved dramatically

through the wise management of nuisance factors.

โ€ข Statistical software can offer blocking and covariance

analysis options.

24.20 Other Considerations

top related