q8 im10 final

46
CHAPTER 10 Statistical Thinking and Applications Teaching Notes This chapter describes concepts of statistics, statistical thinking, statistical methodology, sampling, experimental design, and process capability. Students should be encouraged to take a “big picture” perspective on this framework, rather than the approach of: “How do I get the right answer?” There is a Bonus Materials file on the Premium website that reviews the basic concepts and techniques of statistics that are relevant to the technical areas of statistical process control (SPC). These topics are typically covered in business or engineering statistics course that students should have had prior to taking a course using this text. Key objectives for this chapter include: To establish the importance of statistics as the "bridge" between quality of design and quality of conformance. The proper use of statistics is highlighted as a quality improvement tool. To help students appreciate the importance of statistical thinking in order to understand inter-related processes, process variation, and the need to reduce it in order to assure quality in operations. 1

Upload: jb-macaroco

Post on 16-Sep-2015

134 views

Category:

Documents


6 download

DESCRIPTION

Solution Manual

TRANSCRIPT

CHAPTER 12

13 Statistical Thinking and ApplicationsChapter 10 - Statistical Thinking and Applications 2

CHAPTER 10Statistical Thinking and ApplicationsTeaching NotesThis chapter describes concepts of statistics, statistical thinking, statistical methodology, sampling, experimental design, and process capability. Students should be encouraged to take a big picture perspective on this framework, rather than the approach of: How do I get the right answer?

There is a Bonus Materials file on the Premium website that reviews the basic concepts and techniques of statistics that are relevant to the technical areas of statistical process control (SPC). These topics are typically covered in business or engineering statistics course that students should have had prior to taking a course using this text. Key objectives for this chapter include:

To establish the importance of statistics as the "bridge" between quality of design and quality of conformance. The proper use of statistics is highlighted as a quality improvement tool.

To help students appreciate the importance of statistical thinking in order to understand inter-related processes, process variation, and the need to reduce it in order to assure quality in operations.

To review definitions and concepts of statistics, and relate them to quality control applications. Spreadsheet techniques for statistical analysis with Excel( software are also emphasized.

To introduce the use of DOE as a tool for drawing conclusions regarding controllable process factors and/or comparing methods for process development or improvement.

To help students to understand the concept of process capability and its effects on quality and conformance to specifications.

To introduce the concept of statistical measurement of service quality

The Instructors Resource folder on the course website has a number of Baldrige video clips which give an inside view of organizations that have received the Baldrige award. A couple of those, that are especially appropriate for this chapter, have scenes that show how statistical thinking and concepts can enhance an organizations quest for world-class quality.ANSWERS TO QUALITY IN PRACTICE QUESTIONSImproving Quality of a Wave Soldering Process Through the Design of Experiments1. The first experimental design at the HP plant did not achieve the true optimum combination of factors, because not all combinations were tested. It is theoretically possible that a better combination of factors exists among those that were not tested. Thus, the ones that were tested could be considered a random sampling of all of the possibilities. It is also likely that some interaction effects were at work, so some of the combinations that produced a higher number of defects had to be eliminated.

2. Experimental design allows the experimenter to systematically evaluate two or more methods to determine which is better, or to determine the levels of controllable factors to optimize process yields or minimize variation of a response variable. Therefore, it is generally faster and more efficient than using one at a time, trial-and-error methods.

Applying Statistical Analysis in a Six Sigma Project at GE-Fanuc1. This case showed how wave soldering technology was applied to electronic circuit boards in a process that is very similar to the Improving Quality of a Wave Soldering Process Through the Design of Experiments case, above. There are many variables that must be taken into account in order to improve the process, some of which are the true root causes of the problems in the process, but many are not. Thus, the critical Xs (independent variables) must be separated from the insignificant many, in order to solve the process problems. Many individuals were involved because this was a cross-functional process problem. No one team member had all of the knowledge, but together, they had all that was needed.

2. The Ni-Au boards could have originally been selected for any number of reasons. Reports on the advantages of using Ni-Au boards could have been read in technical journals, a persuasive salesperson for the vendor may have sold it as a solution to other problems, the process may have been changed to correct problems that existed at that time, or it might simply have been a matter of not having the methodology and expertise to do the comparative analysis needed to find out that the Ni-Au boards were no better than less expensive ones.

3. The F-test shows that the vendors, finishes, and interaction effects are all highly significant for this particular independent variable, Wave Solder Skips. Thus, the team must further analyze the defects to find out why the vendors products vary from their competitors, how the finishes differ, and what the interaction effects are.

ANSWERS TO REVIEW QUESTIONS

1.Statistical thinking is a philosophy of learning and taking action based on the principles of: a) all work occurs in a system of interconnected processes; b) variation exists in these processes; and c) understanding and reducing variation are the keys to success in improving processes. It is important to both managers and workers because without it, consistent, predictable processes cannot be established or improved upon.

2.Common causes of variation occur as a natural part of the process and are difficult to change without making a major change in the system of which they are a part. Special causes of variation arise from sources outside the system and can generally be traced back to a specific change that has occurred and needs correction. For example, a process may be stable and running well until the supplier of a critical material is changed. The new vendor's material causes the process to go out of control (becomes unstable), so the "solution" to the special cause is to have the vendor correct the deficiency, or return to the previous supplier for materials.

3.The two fundamental mistakes which managers can make in attempting to improve a process are: (1) To treat as a special (or outside) cause any fault, complaint, mistake, breakdown, accident, or shortage which is actually due to common causes, and (2) to attribute to common causes any fault, complaint, mistake, breakdown, accident, or shortage which is actually due to a special cause. In the first case, tampering with a stable system can increase the variation in the system. In the second case, the opportunity to reduce variation is missed because the amount of variation is mistakenly assumed to be uncontrollable.

4.The Red Bead experiment emphasizes that little, if anything, can improve quality in a poorly-managed production system. In the experiment, managers control incoming material (white and red beads) and work procedures so rigidly that there is little room for change. It is their mistake that there is an input of red bead production material the workers cannot stop the red beads from coming. Management inspects the beads only after they (and the mistakes involved) have been made. No amount of encouragement, threats, or promises of rewards will improve quality production when it is inevitable, by the nature of the process, that red beads will be produced. Furthermore, the managers have mistakenly believed that the variables in the process are controllable, and therefore that the workers are simply not trying hard enough in their labors. The final point of the Red Bead experiment is that all factors of a process must be examined to locate and correct negative variations.

The Funnel experiment is designed to show how people can and do affect the outcome of a process and create unwanted variation by "tampering" with the process, or indiscriminately trying to remove common causes of variation. The system of dropping the ball through the funnel towards the target is damaged by the variation of each participant moving the funnel around to "get a better aim" at the target. The lesson is that once a plan or process is determined to be correct and is set in motion, no components of the process should be tampered with. The process should be adjusted only if the entire process has been thoroughly examined and found to be in need of change in some way.

5.The methods for the efficient collection, organization, and description of data are called descriptive statistics. Statistical inference is the process of drawing conclusions about unknown characteristics of a population from which the data were taken. Predictive statistics is used to develop predictions of future values based on historical data. The three differ in approach, purpose, and outcomes. Descriptive statistics simply summarize and report on existing conditions, inference helps to make decisions about population characteristics based on sample data. Predictive statistics attempt to look into the future and state what will be the results, if certain assumptions hold. All three of these can be important to a manager who is trying to describe the current characteristics of a process, or make inferences about whether a process is in control, or predict future values of instrument readings in order to determine whether it is properly calibrated.

6. Discrete variables are used to measure whether tangible or intangible output from a process is acceptable or not acceptable (good or bad; defective, or not defective). Discrete variables are often used to classify the quality level of customer service. Was the patient in the hospital satisfied or dissatisfied; customer compliments versus complaints for a tour firm; did the marketing research firm accurately or inaccurately prepare the report? Continuous variables are used to measure quantifiable characteristics, such as time, temperature, weight, dimensions (in inches or millimeters). They are only limited by the accuracy of the measuring instrument being used (for example, atomic clocks can measure time to millionth of a second or finer accuracy). Thus, in quality control applications, continuous variables are used to measure dimensions of parts in an automotive supplier, temperature in drying processes, or times required to service each customer in a bank.

7. A population is a complete set or collection of objects of interest. A sample is a subset of objects taken from a population.

8. The standard error of the mean is the (estimated) standard deviation of the population ( divided by

(( /

). The standard deviation is, of course, a measure of variability within a population, where the standard error of the mean is a measure of the standard deviation within the sampling distribution.

9. The central limit theorem is extremely useful in that it states (approximately) that a sampling distribution can be defined as the distribution obtained by taking a large number of samples of size n from any population with a mean and a standard deviation, (, and calculating their means. The mean of the sample means for this probability distribution will approach , and the standard deviation of the distribution will be ( /

, as larger and larger sample sizes are taken. The CLT is extremely important in any SQC techniques that require sampling.

10.The two factors that influence sampling procedure are the method of selecting the sample and the sample size. Methods include simple random sampling, stratified sampling, systematic sampling and cluster sampling. Sample size is dependent on the amount of variation in the population (measured by () and the amount of error that the decision maker can tolerate at a specified confidence level.

11. The basic questions that must be addressed in a sampling study are:

a. What is the objective of the study?

b. What type of sample should be used?

c. What possible error might result from sampling?

d. What will the study cost?

All of these questions are inter-related, and only the last two are quantifiable. Anyone making a decision regarding these questions needs to have some understanding of the options, relative accuracy, and cost tradeoffs of each potential alternative in sampling studies. The long-run consequences of making a wrong choice in the design of a sampling study can be quite serious. For example, if a pharmaceutical firm makes a substantial error in the design or execution of a sampling study on a new drug, the result could be felt in the direct costs of running a correct study, the likelihood that customers would sue the company for injury suffered in taking the drug, or for the lack of any substantial benefit coming from having taken an ineffective drug. In addition, the firm's reputation could be permanently damaged.

12.Methods of sample selection, listed previously, include: simple random sampling, stratified sampling, systematic sampling and cluster sampling. Simple random sampling is useful where one needs to gather information from a moderately large, homogeneous population of items. For example, if a MBA director wished to find out the attitudes of 300 MBA students toward various policies, procedures, and services provided to the students, s(he) might use a simple random sample to determine whom the survey should be sent to. An automobile insurance company could use a stratified sample to determine accident rates of customers, stratified according to their ages. An auditor might use a systematic sampling to sample accounts receivable records by choosing every 50th record out of a file cabinet. Cluster sampling could be used by management analysts within city government to determine satisfaction levels of residents on a neighborhood by neighborhood (cluster) basis. Judgment sampling should be avoided, except as a way to gather preliminary, impressionistic data before beginning a true sampling study.

13.Systematic errors in sampling can come from bias, non-comparable data, uncritical projection of trends, causation, and improper sampling. They may be avoided by approaches discussed in the chapter. Basically, careful planning of the sampling study, awareness of possible systematic error causes, and careful execution of the study can help to avoid most of the common errors listed above.

14.The purpose of design of experiments is to set up a test or series of tests to enable the analyst to compare two or more methods to determine which is better, or to determine levels of controllable factors to optimize the yield of a process, or minimize variability of a response variable.

15.A factorial experiment is a specific type of experimental design that considers all combinations of levels of each factor. For example, a factorial experiment might be set up to determine the profitability of an ice cream store based on external and broad quality related factors. External factors, which could influence profits, would be days/times of the week and external temperature. Quality factors might be quality of flavors available and perceived quality characteristics (time, timeliness, empathy, etc.) of the workers.

16.Simple factorial experiments often require many iterations of the experiment to be run before reliable results are obtained. ANOVA permits testing multiple combinations of factors, simultaneously. Therefore it is often faster and cheaper to perform ANOVA than using simple factorial experiments. Conclusions regarding interaction effects can also be tested more rigorously using analysis of variance.

17.ANOVA is a methodology for drawing conclusions about equality of means of multiple populations. In its simplest form one-way ANOVA it compares means of observed responses of several different levels of a single factor. ANOVA tests the hypothesis that the means of all populations are equal against the alternative hypothesis that at least one mean differs from the others.

SOLUTIONS TO PROBLEMS CHAPTER 10

Note: Data sets for several problems in this chapter are available in the Excel workbook C10Data.xls on the Premier website accompanying this chapter. Click on the appropriate worksheet tab as noted in the problem (e.g., Prob. 10-1) to access the data.

1.Use the data for Twenty First Century Laundry for the weights of loads of clothes processed through their washing department in a week. (See Prob. 10-1 in C10Data.xls on the Premier website). Apply the Descriptive Statistics and Histogram tools in Excel to compute the mean, standard deviation, and other relevant statistics, as well as a frequency distribution and histogram for the following data. From what type of distribution might you suspect the data are drawn?Answer

1. The following results were obtained from the Twenty First Century Laundry data

Column1

Mean32.920

Standard Error2.590

Median25.500

Mode14.000

Standard Deviation25.899

Sample Variance670.741

Kurtosis0.233

Skewness0.994

Range106.000

Minimum1.000

Maximum107.000

Sum3292.000

Count100.000

Largest(1)107.000

Smallest(1)1.000

Confidence Level(95.0%)5.139

The conclusion that can be reached from looking at the summary statistics and the histogram is that these data are exponentially distributed, with descending frequencies. These data may show that small, low-weight batches are most frequently processed, as represented by the histogram. This has implications for the number and size of washers and dryers that should be installed in the laundry. More detailed data can be found in the Excel( solution spreadsheet coded as P10-1&2.xls.

2.The times for carrying out a blood test at Rivervalley Labs were studied in order to learn about process characteristics. Apply the Descriptive Statistics and Histogram analysis tools in Excel to compute the mean, standard deviation, and other relevant statistics, as well as a frequency distribution and histogram, for the data taken from 100 tests and found in the Prob. 10-2, Excel data set. From what type of distribution might you suspect the data are drawn?Answer

2. One of the advantages of using Excel( spreadsheets is that a great deal of analysis can be done easily. The summary statistics follow. Also shown is the histogram constructed by using Excels( Data Analysis tools (found under the Tools heading on the spreadsheet). For best results in constructing the histogram, it is suggested that students set up their own bins so as to provide 7 to 10 approximately equal sized class intervals for the data. Note that if the program finds that the classes shown in the bins do not extend over the upper or lower range of the data, it will automatically compensate by adding a Less or More category for the outliers.

Descriptive Statistics

Column1

Mean3.578

Standard Error0.081

Median3.600

Mode3.600

Standard Deviation0.812

Sample Variance0.660

Kurtosis-0.267

Skewness-0.223

Range3.600

Minimum1.700

Maximum5.300

Sum357.800

Count100.000

Largest(1)5.300

Smallest(1)1.700

Confidence Level(95.0%)0.161

The conclusion that can be reached from looking at the summary statistics and the histogram is that these data are fairly normally distributed, with some slight skewing to the left. More detailed data can be found in the Excel( solution spreadsheet P10-1&2.xls.

3.The data (Prob. 10-3 in C10Data.xls found on the Premier website for this textbook) represent the weight of castings (in kilograms) from a production line in the Fillmore Metalwork foundry. Based on this sample of 100 castings, compute the mean, standard deviation, and other relevant statistics, as well as a frequency distribution and histogram. What do you conclude from your analysis?Answer

3.Descriptive statistics for the Fillmore Metalwork foundry are shown below.

Descriptive Statistics

Column1

Mean

38.654

Standard Error

0.031

Median

38.600

Mode

38.600

Standard Deviation

0.306

Sample Variance

0.094

Kurtosis

4.112

Skewness

-0.324

Range

2.300

Minimum

37.300

Maximum

39.600

Sum

3865.400

Count

100.000

Largest(1)

39.600

Smallest(1)

37.300

Confidence Level(95.0%)

0.061

The conclusion that can be reached from looking at the summary statistics and the histogram is that these data are not strongly normally distributed, due to significant skewing to the right. More detailed data can be found in the Excel( solution spreadsheet P10-3&4.xls.

4.The data (Prob. 10-4 in C10Data.xls found on the Premier website for this textbook) show the weight of castings (in kilograms) being made in the Fillmore Metalwork foundry and were taken from another production line. Compute the mean, standard deviation, and other relevant statistics, as well as a frequency distribution and histogram. Based on this sample of 100 castings, what do you conclude from your analysis?

Answer

4.Data from the Fillmore Metalwork foundry are shown below.

Descriptive Statistics

Column1

Mean38.6270

Standard Error0.0474

Median38.6000

Mode38.7000

Standard Deviation0.4737

Sample Variance0.2244

Range2.6000

Minimum37.3000

Maximum39.9000

Sum3862.7000

Count100.0000

The conclusion that can be reached from looking at the summary statistics and the histogram is that these data are not normally distributed, due to pronounced skewing to the right. More detailed data can be found in the Excel( solution spreadsheet P10-3&4.xls.

5.San Juan Green Tea is sold in 1/2 liter (500 milliliter) bottles. The standard deviation for the filling process is 10 milliliters. If the process requires a 1 percent, or smaller, probability of over-filling, defined as over 495 milliliters, what must the target mean for the process be?

Answer

5.For San Juan Green Teas bottling process, the values for the 1% cutoff and the standard deviation are:

x = 495 ml; ( = 10 ml

For a total probability of 1% for overfilling:

P(x > upper fill limit) = 0.5000 - P = 0.5000 - .4900 = 0.01

Using the Normal Table, Appendix A, z = 2.33

x - 495 -

z = --------- = --------------- = 2.33

( 10

= 471.7 ml

( The process mean should be 471.7 ml., so that there is only a 1% probability of overfilling.

6.Texas Punch was made by Frutayuda, Inc. and sold in 12-ounce cans to benefit victims of Hurricane Ike. The mean number of ounces placed in a can by an automatic fill pump is 11.8 with a standard deviation of 0.12 ounce. Assuming a normal distribution, what is the probability that the filling pump will cause an overflow in a can, that is, the probability that more than 12 ounces will be released by the pump and overflow the can?Answer

6.For cans of Texas Punch the mean, = 11.8; the standard deviation, ( = 0.12

x - 12 11.8

z = --------- = --------------- = 1.667 = 1.67

( 0.12

P(x > 12) = 0.5000 P ( 0 < z < 1.67 )

P(z > 12) = 0.5000 - 0.4525 = 0.0475

Thus, there is a 4.75% probability of an overflow.

(Results are based on the Standard Normal Table, Appendix A)

7.Kiwi Blend is sold in 900 milliliter (ml) cans. The mean volume of juice placed in a can is 876 ml with a standard deviation of 12 ml. Assuming a normal distribution, what is the probability that the filling machine will cause an overflow in a can, that is, the probability that more than 900 ml will be placed in the can?

Answer

7.The mean, for the Kiwi Blend product is = 876; the standard deviation, ( = 12, x = 900.

x - 900 876

z = --------- = ---------------- = 2.0

( 12

P(x > 900) = 0.5000 - P ( 0 < z < 2.0 )

P(z > 900) = 0.5000 - 0.4772 = 0.0228

Thus, there is a 2.28 % probability of an overflow.

(Results are based on the Standard Normal Table, Appendix A)

8.Wayback Beer bottles have been found to have a standard deviation of 5 ml. If 95 percent of the bottles contain more than 250 ml, what is the average filling volume of the bottles? Answer

8. Given that the standard deviation for Wayback Beer, as ( = 5 ml., x = 250, and P (z > x) = 0.95

Since z = - 1.645 for x > 0.95

250 -

-1.645 = ----------- ; ( = 241.775 ml.

5

(Results are based on the Standard Normal Distribution Table, Appendix A)

9.The mean filling weight of salt containers processed by the Piedra Salt Co. is 15.5 ounces. If 2 percent of the containers contain more than 16 ounces, what is the standard deviation of the filling weight of the containers?

Answer

9.Given that the process mean filling weight is = 15.5 oz. for the Piedra salt containers,

By looking up 0.5000 - 0.0200 = 0.48, we find z = 2.05

16 15.5

z = 2.05 = -------------

(( ( = 0.2439 oz.

(Results are based on the Standard Normal Distribution Table, Appendix A)

10.In filling bottles of E&L Cola, the average amount of over- and under-filling has not yet been precisely defined, but should be kept as low as possible. If the mean fill volume is 11.9 ounces and the standard deviation is 0.05 ounce:

a) What percentage of bottles will have more than 12 ounces (overflow)?

b) Between 11.9 and 11.95?

c) Less than 11.83 ounces?Answer

10.The mean value for E&L Cola, in ounces is: = 11.9; the standard deviation, ( = 0.05

a) P(x > 12.0) = 0.5000 - P ( 11.9 > x > 12.0)

x - (12.0 11.9)

z = --------- = ----------------- = 2.0

( 0.05

P(z > 2.0) = 0.5000 - 0.4772 = 0.0228

Thus, 2.28 % will be filled to more than 12 ounces (overflowing).

b) P (11.9 < x < 11.95)

x - (11.95 11.9)

z = --------- = ----------------- = 1.0

( 0.05

P(0 < z < 1.0) = z = 0.3413

Thus, 34.13 % will have a fill value between 11.9 and 11.95 ounces.

c) For P(x < 11.83) = 0.5000 - P (11.83 < x < 11.9)

x - (11.83 11.9)

z = --------- = ------------------ = -1.4

( 0.05

P(x < 11.83) = 0.5000 - P ( -1.4 < z < 0) = 0.5000 0.4192 = 0.0808

Thus, 8.08 % will be filled to less than 11.83 ounces.

(Results are based on the Standard Normal Table, Appendix A)

11.The frequency table for Prob. 10-11 found in C10Data.xls on the Premier website shows the weight of castings (in kilograms) being made in the Fillmore Metalwork foundry (also see Prob 10-3, in C10Data.xls, for raw data).

a. Based on this sample of 100 castings, find the mean and standard deviation of the sample. (Note: If only the given data are used, it will be necessary to research formulae for calculating the mean and standard deviations using grouped data from a statistics text.)

b. Prepare and use an Excel spreadsheet, if not already done for Problem 10-3, to plot the histogram for the data.

c. Plot the data on normal probability paper to determine whether the distribution of the data is approximately normal. (Note: The Regression tool in Excel has a normal probability plot that may be used here.) Answer

11. An older statistics book from the library may have to be consulted in order to find background information on the formulae used for grouped statistics in solving this problem (see below). The following, use class midpoints estimated from the given cells. (Note that the data have been re-sorted in ascending order to conform to the format in the Excel statistical package.)

Answer

From the Excel statistical package, we get:

Upper CellCumulative

BoundariesFrequencies%

Cell 137.566.0%

Cell 237.81016.0%

Cell 338.13854.0%

Cell 438.42781.0%

Cell 538.71091.0%

Cell 639.0697.0%

Cell 739.3198.0%

Cell 839.62100.0%

For castings made at the Fillmore Metalwork foundry, when calculated using a hand calculator, the following grouped data results can be obtained:Adj. Cell

MidpointsFrequenciesfxfx^2

Cell 138.16228.608709.66

Cell 238.34153.205867.56

Cell 338.520770.0029645.00

Cell 438.7411586.7061405.29

Cell 538.916622.4024211.36

Cell 639.17273.7010701.67

Cell 739.34157.206177.96

Cell 839.5279.003120.50

1003870.80149839.00

a) fx3870.80

= n = 100 = 38.708 (vs. 38.654 from the data in spreadsheet 10-3)

s = = = 0.2856 (versus

0.306 from the spreadsheet data)

Note that the calculated value from the frequency distribution and from the raw data would probably have been much closer had the data been more uniformly distributed and less skewed around the mean.

b) See answer and spreadsheet printouts in Problem 10-3, above.

c) A normal probability plot shows that the data are approximately normally distributed, with an R square value of 0.825. (See spreadsheet Prob10-3&4 for more details). This value is good, but indicates only a moderately good fit because of the skewed distribution, as noted in 10-3.

The conclusions that can be reached are:

The formula for grouped data gives a close approximation of the statistics from Excel, and from the actual data.

The normal probability plot and the histogram show that these data are approximately normally distributed, with an R-squared value of 0.825.

12.The frequency table for Prob. 10-12 found in C10Data.xls on the Premier website shows the weight of another set of castings (in kilograms) being made in the Fillmore Metalwork foundry (also see Prob. 10-4, in C10Data.xls, for raw data).

a. Based on this sample of 100 castings, find the mean and standard deviation of the sample. (Note: If only the given data are used, it will be necessary to research formulae for calculating the mean and standard deviations using grouped data from a statistics text.)

b. Prepare and use an Excel spreadsheet, if not already done for Problem 10-4, to plot the histogram for the data.

c. Plot the data on normal probability paper to determine whether the distribution of the data is approximately normal. (Note: The Regression tool in Excel has a normal probability plot that may be used here.)

Answer

12.An older statistics book from the library may have to be consulted in order to find background information on the formulae used for grouped statistics in solving this problem (see below). The following, use class midpoints estimated from the given cells. (Note that the data have been re-sorted in ascending order to conform to the format in the Excel statistical package.)

From the Excel statistical package, we get:

Upper CellCumulative

BoundariesFrequencies%

Cell 137.511.00%

Cell 237.834.00%

Cell 338.11014.00%

Cell 438.42438.00%

Cell 538.72765.00%

Cell 639.01681.00%

Cell 739.31394.00%

Cell 839.6498.00%

Cell 939.92100.00%

When calculated using a hand calculator, you may obtain the following, using the grouped data formula:

Adj. Cell

MidpointsFrequenciesfxfx^2

Cell 137.35137.351395.02

Cell 237.65137.651417.52

Cell 337.956227.708641.22

Cell 438.2511420.7516093.69

Cell 538.55261002.3038638.67

Cell 638.8524932.4036223.74

Cell 739.1518704.7027589.01

Cell 839.459355.0514006.72

Cell 939.75279.503160.13

Cell 1040.05280.103208.01

1003877.50150373.71

a)

fx3877.5

= n = 100 = 38.775 (vs. 38.627 from the actual data in

spreadsheet 10-4)

s = = = 0.4887 (versus 0.4737

from the data in

spreadsheet 10-4)

b)See answer and spreadsheet printouts in Problem 10-4, above.

c) A normal probability plot shows these data are approximately normally distributed, with an R square value of 0.947. (See spreadsheet Prob10-3&4 for more details).

The conclusions that can be reached are:

The formula for grouped data gives a close approximation of the statistics from Excel, and from the actual data.

The normal probability plot and the histogram show that these data are approximately normally distributed, with an R-squared value of 0.947.

13.In a filling line at A & C Foods, Ltd., the mean fill volume for rice bubbles is 325 grams and the standard deviation is 20 grams. What percentage of containers will have less than 295 grams? More than 345 grams (assuming no overflow)?Answer

13. For the A & C Foods, Ltd. rice bubble filling process, the mean is = 325 grams and the standard deviation, ( = 20

P(x < 450) = 0.5000 - P ( 450 < x < 475)

x - (295 - 325)

z = --------- = ----------------- = - 1.5

( 20

P(z < -1.5) = 0.5000 - 0.4332 = 0.0668

(Results are based on the Standard Normal Table, Appendix A)

Thus, 6.68 % will be filled with less than 295 grams

P(x > 345) = 0.5000 P (325 < x < 345)

x - (345 - 325)

z = --------- = ---------------- = 1.0

(

20

P (z > 1.0) = 0.5000 - 0.3413 = 0.1587

Thus, 15.87 % will contain more than 345 grams

(Results are based on the Standard Normal Table, Appendix A)

14.Tessler Electric utility requires service operators to answer telephone calls from customers in an average time of 0.1 minute or less. A sample of 30 actual operator times was drawn, and the results are given in the following table. In addition, operators are expected to determine customer needs and either respond to them or refer the customer to the proper department within 0.5 minute. Another sample of 30 times was taken for this job component and is also given in the table. If these variables can be considered to be independent, is the average time taken to perform each component statistically different from the standard?

ComponentMean TimeStandard Deviation

Answer0.10230.0183

Service0.52900.0902

Answer

14. Specification for answer time for the Tessler utility is :

H0: Mean response time: ( 1 < 0.10

H1: Mean response time: ( 1 > 0.10

1 = 0.1023, s1 = 0.0183

and the t-test is:

, t29, .05 = 1.699Specification for service time is:

H0: Mean service time: ( 2 < 0.50

H1: Mean service time: ( 2 > 0.50

2 = 0.5290, s2 = 0.0902

and the t-test is:

, t29, .05 = 1.699Because t29, .05 = 1.699, we cannot reject the null hypothesis for t1, but we can reject the hypothesis for t2 . Therefore, there is no statistical evidence that the mean response time exceeds 0.10 for the answer component, but the statistical evidence does support the service component.Note: Problems 1518 address sample size determination and refer to theory covered in the Bonus Material folder for this chapter as contained on the Premier website.

15.You are asked by the owner of the Moonbow Motel to develop a customer satisfaction survey to determine the percentage of customers who are dissatisfied with service. In the past year, 10,000 customers were serviced. She desires a 95 percent level of confidence with an allowable statistical error of 0.01. From past estimates, the manager believes that about 3.5 percent of customers have expressed dissatisfaction. What sample size should you use for this survey?

Answer

15.The size of the population is irrelevant to this customer satisfaction survey, although it is good to know that it is sizable. Therefore, make the following calculations:

n = (z (/2)2 p(1-p) / E2 = (1.96)2 (0.035)(0.965) / (0.01)2 = 1297.5, use 1298

16.Determine the appropriate sample size needed to estimate the proportion of sorting errors at the Puxatawney post office at a 99 percent confidence level. Historically, the sorting error rate is 0.015, and you wish to have an allowable statistical error of 0.02.Answer16.The sample size for the proportion of sorting errors at a post office, using a 99% confidence level is:

n = (z (/2)2 p(1-p) / E2 = (2.33)2 (0.015)(0.985) / (0.02)2 = 200.53, use 201

17.A management engineer at Country Squire Hospital determined that she needs to make a work sampling study to see whether the proportion of idle time in the diagnostic imaging department had changed since being measured in a previous study several years ago. At that time, the percentage of idle time was 8 percent. If the engineer can only take a sample of 850 observations due to cost factors, and can tolerate an allowable error of 0.02, what percent confidence level can be obtained from the study?Answer

17.Using the formula: n = (z (/2)2 p(1-p) / E2 , the engineer at the Country Squire Hospital can solve for z (/2 as follows:

850 = (z (/2 )2 (0.08)(0.92) / (0.02)2

850 = (z (/2 )2 (184)

(z (/2 )2 = 850 / 184

z (/2 = = 2.149, use 2.15

From the Standard Normal Distribution table, Appendix A, we find a probability of 0.4838 for z = 2.15. Because it is only one tail of the distribution, we multiply the area by 2 to get the confidence level of 0.9676. Thus, the management engineer can be almost 97% confident of her results based on this sample size.

18.Localtel, a small telephone company, interviewed 200 customers to determine their satisfaction with service. Nineteen expressed dissatisfaction. Compute the sample size needed to ensure that they can be 90 percent confident of correctly estimating the proportion dissatisfied with an allowable error of 0.04.

Answer

18.First, we must find an estimated p for Localtels sample:

p =__19__ = 0.095

200

Then, using the formula for sample size:

n = (z (/2)2 p (1-p) / E2 , we can solve for n as follows:

n = (1.65)2 (0.095)(0.905) / (0.04)2

n = [(2.723) (0.0860)]/ (0.0016) = 146.36 use 147

Thus, Localtel can be more than 90% confident of their results by using a sample size of 200. Thus, there is no need to take more samples in order to meet the required sample size.

19.Using the Discovery Sampling table found in the Bonus materials on the Premier website, suppose that a population consists of 2,000 units. The critical rate of occurrence is 1 percent, and you wish to be 99 percent confident of finding at least one nonconformity. What sample size should you select?

Answer

19. Using the Discovery Sampling table (in the Sample Size Determination file) in the Bonus materials on the Premier website, for a population of 2,000, and reading the sample size required for a critical 1% rate, with a 99% confidence level yields the solution for the samples size of approximately 400 (use 98.9% confidence, with a critical rate of 1%).

20.The process engineer at Sival Electronics was trying to determine whether three suppliers would be equally capable of supplying the mounting boards for the new gold plated components that she was testing. The table found in Prob. 10-20 on C10Data.xls on the Premier website shows the coded defect levels for the suppliers, according to the finishes that were tested. Lower defect levels are preferable to higher levels. Using one-way ANOVA, analyze these results. What conclusion can be reached, based on these data?

Answer

20. The process engineer at Sival Electronics can develop a one-way ANOVA spreadsheet (see spreadsheet 10-20 for details) that shows:

Supplier 1Supplier 2Supplier 3

Finish 111.96.813.5

Finish 210.35.910.9

Finish 39.58.112.3

Finish 48.77.214.5

Finish 514.27.612.9

SUMMARY

GroupsCountSumAverageVariance

Supplier 1554.610.924.762

Supplier 2535.67.120.697

Supplier 3564.112.821.812

ANOVA

Source of VariationSSdfMSFP-valueF crit

Between Groups84.233242.11717.3770.000293.885

Within Groups29.084122.424

Total113.31714

According to the F-test and probability of P < .05, there is a significant difference between suppliers

21.The process engineer at Sival Electronics is also trying to determine whether a newer, more costly design involving a gold alloy in a computer chip is more effective than the present, less expensive silicon design. She wants to obtain an effective output voltage at both high and low temperatures, when tested with high and low signal strength. She hypothesizes that high signal strength will result in higher voltage output, low temperature will result in higher output, and the gold alloy will result in higher output than the silicon material. She hopes that the main and interaction effects with the expensive gold will be minimal. The data found in Prob. 10-21 on C10Data.xls on the Premier website were gathered in testing of all 2n combinations. What recommendation would you make, based on these data?

Answer

21. Using the following input data:

SignalMaterialTemperatureOutput Voltage

HighGoldLow18

HighGoldHigh12

HighSiliconLow16

HighSiliconHigh10

LowGoldLow8

LowGoldHigh11

LowSiliconLow7

LowSiliconHigh14

the process engineer at Sival Electronics can calculate the main effects as follows:

Signal

High (18 + 12 + 16 + 10)/ 4 = 14

Low ( 8 + 11 + 7 + 14)/ 4 = 10

High - Low = 4

Material

Gold (18 + 12 + 8 + 11)/ 4 = 12.25

Silicon (16 + 10 + 7 + 14)/ 4 = 11.75

Gold - Silicon = 12.25 - 11.75 = 0.5

Temperature

Low (18 + 16 + 8 + 7)/ 4 = 12.25

High (12 + 10 + 11 + 14)/ 4 = 11.75

Low - High = 12.25 - 11.75 = 0.5

The main effects of the signal (high or low) far outweigh the effects of material and temperature, indicating that those factors are insignificant. So, interaction effects will be negligible.

SUGGESTIONS FOR PROJECTS, ETC.1.This project will show that there are different ways to interpret Demings Red Bead experiment, statistically.

2.This experiment is designed to give the student hands on experience in elementary experimental design.

3.This experiment in aerodynamics will give the student some hands on experience in dealing with design problems in which multiple variables can have a significant impact on the operability of the mechanism. ANSWERS TO CASE QUESTIONSCase - The Disciplinary Citation

1.This case study shows Deming's "Red Bead" Experiment in action. Drivers are being blamed for conditions that are not under their control. The problem could be addressed by process measurement, eliminating "special causes" and reducing common causes.

2.A run chart would appear to be a way to begin to understand the process and to determine if it is in control or not. Based on the available data, we have:

Center Line (average) for the chart = 240/40 = 6.0 mistakes

The data show that 12 drivers have exceeded the average. Also, 6 drivers had "no defects". A Pareto chart (see spreadsheet disccase.xls for details) shows that only 8 drivers have more than 10 errors. The characteristics of drivers who are having difficulty should be examined to explain their higher error rates. Are their errors far above normal, or just a little above? Are they well trained? Are they overworked, with more than the average number of difficult orders? Do they have poor equipment? A useful control chart cannot be established, unless "special causes" are dealt with. Analysis should also be done to determine what the good drivers are doing right. Are they more experienced, drive newer cars, have better hearing and vision, etc.?

After corrections, a new chart (called a "c-chart" and discussed in Chapter 12) on the stable process can be set up. Then those who consistently do well can be rewarded and the performance of those who have an unsatisfactory level of errors can be improved.

Case - The Quarterly Sales Report

Obviously, Hagler has got to look beyond the quarterly sales reports to find out what is really going on in his regions. Two simple scatter diagrams (see spreadsheet C10Salecase.xls for details) that plot the entire 5 years of data, divided into large regions and small regions, are very revealing. He might start with those graphs to see what the general trend has been.

The Northeast, Southwest, and Northwest regions show basically flat sales for the 5 year period. The Southwest region had a strange jump in sales in the second quarter of 2001 and again in the third quarter of 2002, possibly due to inventory reduction sales. Otherwise, the three regions showed no consistent trend, while still managing to increase sales in the last quarter of 2003.

Two of the three smaller regions, North Central and Mid-Atlantic, showed upward trends from the third quarter of 2001 to the third quarter of 2003. Unfortunately, Hagler concentrated on the one quarter when sales went down for both of them, while also decreasing for the South Central region, which had been holding its own in a tough market.

To apply concepts of statistical thinking, Hagler must: 1) look at the big picture, graphically, rather than at quarter to quarter variations, 2) find out what each regions unique characteristics are, rather than treating all regions the same, 3) determine what causes variation within and among the regions, 4) determine if variations are due to common causes or special causes (probably the former), and 5) train, help, and support the regional managers if he expects to see different results over the coming year.

See spreadsheet C10SaleCase for further analysis.

Case - The HMO Pharmacy Crisis

It appears that, because of the crisis, top management at Dover is reacting in a typical red Bead manner. The importance of statistical concepts in quality management cannot be overemphasized. Indeed, statistics is essential in implementing a continuous improvement philosophy.

Statistical thinking is a philosophy of learning and action based on the principles that:

1. All work occurs in a system of interconnected processes.

2. Variation exists in all processes.

3. Understanding and reducing variation are keys to success

Understanding processes provides the context for determining the effects of variation and the proper type of managerial action to be taken. By viewing work as a process, we can apply statistical tools to establish consistent, predictable processes, study them, and improve them. While variation exists everywhere, many business decisions do not often account for it, and managers frequently confuse common and special causes of variation. Those inside and outside the pharmacy must understand the nature of variation, before they can focus on reducing it.

The complex interactions of these variations in drugs, equipment, computer systems, professional, clerical, and technical staff, and the environment are not easily understood. Variation due to any of these individual sources could be random; individual sources may not be identified or explainable. However their combined effect in the pharmaceutical system is probably stable and might be predicted statistically. These common causes of variation that are present as a natural part of the process need to be understood before special causes can be separated and eliminated.

To address the problem, Dover should consider using the following steps:

Form a cross-functional group consisting of pharmacists, assistant pharmacists, physicians, nurses, health care insurance experts, and administrative support people who may handle prescriptions or assess pharmaceutical risks.

Develop an understanding of the process by flowcharting all steps, including interconnected processes in various locations.

Do a risk assessment to determine where there are gaps in procedures or potential risk of errors

Develop corrective action teams to work first on the causes with the highest risk if errors in prescription handling occur.

Propose solutions

Pilot test solutions to ensure that problems have been corrected

Disseminate results to all concerned

Track results to determine long-term benefits.

Work on solving problems in priority order

Benchmark best practices for future improvements.

1

_1047654175.unknown

_1047654215.unknown

_1306824951.unknown

_1306829785.unknown

_1122710351.unknown

_1298110149.unknown

_1122710097.unknown

_1047654186.unknown

_1047654214.unknown

_1047654182.unknown

_1047654184.unknown

_1047654181.unknown

_1043318265.unknown

_1043318266.unknown

_1047227205.xlsSheet: Error Data

Sheet: Scatter Diagram

Sheet: Pareto Chart

The Disciplinary Citation Case

Scatter Diagram for Package Errors

Operator

Errors

Avg.

1.0

1.0

2.0

2.0

3.0

3.0

4.0

4.0

5.0

5.0

6.0

6.0

7.0

7.0

8.0

8.0

9.0

9.0

10.0

10.0

11.0

11.0

12.0

12.0

13.0

13.0

14.0

14.0

15.0

15.0

16.0

16.0

17.0

17.0

18.0

18.0

19.0

19.0

20.0

20.0

21.0

21.0

22.0

22.0

23.0

23.0

24.0

24.0

25.0

25.0

26.0

26.0

27.0

27.0

28.0

28.0

29.0

29.0

30.0

30.0

31.0

31.0

32.0

32.0

33.0

33.0

34.0

34.0

35.0

35.0

36.0

36.0

37.0

37.0

38.0

38.0

39.0

39.0

40.0

40.0

6.0

6.0

1.0

6.0

0.0

6.0

14.0

6.0

0.0

6.0

2.0

6.0

18.0

6.0

2.0

6.0

5.0

6.0

13.0

6.0

1.0

6.0

4.0

6.0

6.0

6.0

5.0

6.0

0.0

6.0

0.0

6.0

1.0

6.0

3.0

6.0

15.0

6.0

24.0

6.0

3.0

6.0

4.0

6.0

1.0

6.0

2.0

6.0

3.0

6.0

22.0

6.0

4.0

6.0

8.0

6.0

2.0

6.0

6.0

6.0

8.0

6.0

0.0

6.0

9.0

6.0

20.0

6.0

9.0

6.0

0.0

6.0

3.0

6.0

14.0

6.0

1.0

6.0

1.0

6.0

Operator

Errors

_950803616.unknown