1 slide slide bases of the theory of probability and mathematical statistics
Post on 31-Dec-2015
220 Views
Preview:
TRANSCRIPT
11 11 SlideSlide
SlideSlide
Bases of the theory of probability and mathematical statistics.
22 22 SlideSlide
SlideSlide
An experiment is a situation involving chance or probability that leads to results called outcomes.In the problem above, the experiment is spinning the spinner.
An outcome is the result of a single trial of an experiment.The possible outcomes are landing on yellow, blue, green or red.
An event is one or more outcomes of an experiment.One event of this experiment is landing on blue.
Probability is the measure of how likely an event is.
33 33 SlideSlide
SlideSlide
DefinitionsDefinitionsCertain
Impossible
.5
1
0
50/50
Probability is the numerical measure of the likelihood that the event will occur.
Value is between 0 and 1.
Sum of the probabilities of all eventsis 1.
44 44 SlideSlide
SlideSlide
Experimental vs.Theoretical
Experimental probability:
P(event) = number of times event occurs
total number of trials
Theoretical probability:
P(E) = number of favorable outcomes total number of possible outcomes
55 55 SlideSlide
SlideSlide
Identifying the Type of Probability
Trial Red Blue
1 1
2 1
3 1
4 1
5 1
6 1
Total 2 4
Exp. Prob. 1/3 2/3
You draw a marble out of the bag, record the color, and replace the marble. After 6 draws, you record 2 red marbles
P(red)= 2/6 = 1/3 Experimental(The result is found by
repeating an experiment.)
66 66 SlideSlide
SlideSlide
The complement of A is everything in the sample space S that is NOT in A.
AS
•If the rectangular box is S, and the white circle is A, then everything in the box that’s outside the circle is Ac , which is the complement of A.
77 77 SlideSlide
SlideSlide
TheoremPr (Ac) = 1 - Pr (A)
Example:
If A is the event that a randomly selected student is male, and the probability of A is 0.6, what is Ac and what is its probability?
Ac is the event that a randomly selected student is female, and its probability is 0.4.
88 88 SlideSlide
SlideSlide
The union of A & B (denoted A U B) is everything in the sample space that is in either A or B or both.
AS
•The union of A & B is the whole white area.
B
99 99 SlideSlide
SlideSlide
The intersection of A & B (denoted A∩B) is everything in the sample space that is in both A & B.
S
•The intersection of A & B is the pink overlapping area.
BA
1010 1010 SlideSlide
SlideSlide
Example A family is planning to have 2 children. Suppose boys (B) & girls (G) are equally
likely.
What is the sample space S?
S = {BB, GG, BG, GB}
1111 1111 SlideSlide
SlideSlide
Example continuedIf E is the event that both children are the same sex, what does E look like & what is its probability?
E = {BB, GG}
Since boys & girls are equally likely, each of the four outcomes in the sample space S = {BB, GG, BG, GB} is equally
likely & has a probability of 1/4.
So Pr(E) = 2/4 = 1/2 = 0.5
1212 1212 SlideSlide
SlideSlide
Example cont’d: Recall that E = {BB, GG} & Pr(E)=0.5
What is the complement of E and what is its probability?
Ec = {BG, GB}
Pr (Ec) = 1- Pr(E) = 1 - 0.5 = 0.5
1313 1313 SlideSlide
SlideSlide
Example continued
If F is the event that at least one of the children is a girl, what does F look like & what is its probability?
F = {BG, GB, GG}
Pr(F) = 3/4 = 0.75
1414 1414 SlideSlide
SlideSlide
Recall: E = {BB, GG} & Pr(E)=0.5F = {BG, GB, GG} & Pr(F) = 0.75
What is E∩F? {GG}
What is its probability? 1/4 = 0.25
1515 1515 SlideSlide
SlideSlide
Recall: E = {BB, GG} & Pr(E)=0.5F = {BG, GB, GG} & Pr(F) = 0.75What is the EUF?
{BB, GG, BG, GB} = SWhat is the probability of EUF?
1If you add the separate probabilities of E & F together, do you get Pr(EUF)? Let’s try it.Pr(E) + Pr(F) = 0.5 + 0.75 = 1.25 ≠ 1 = Pr (EUF)Why doesn’t it work?We counted GG (the intersection of E & F) twice.
1616 1616 SlideSlide
SlideSlide
A formula for Pr(EUF)
Pr(EUF) = Pr(E) + Pr(F) - Pr(E∩F)
If E & F do not overlap, then the intersection is the empty set, & the probability of the intersection is zero.
When there is no overlap, Pr(EUF) = Pr(E) + Pr(F) .
1717 1717 SlideSlide
SlideSlide
We can deduce an important result from the conditional law of probability:
( The probability of A does not depend on B. )
or P(A B) P(A) P(B)
If B has no effect on A, then, P(A B) = P(A) and we say the events are independent.
becomes P(A)
P(A B)P(B)
P(A|B)
So, P(A B)P(B)
Independent Independent EventsEvents
1818 1818 SlideSlide
SlideSlide
Tests for independence
P(A B) P(A) P(B)
P(A B) P(A)
or
Independent Independent EventsEvents
P(B A) P(B)
1919 1919 SlideSlide
SlideSlide
The Multiplication The Multiplication RuleRule
If events A and B are independent, then the probability of two events, A and B occurring in a sequence (or simultaneously) is:
This rule can extend to any number of independent events.
)()()()( BPAPBAPBandAP
Two events are independent if the occurrence of the first event does not affect the probability of the occurrence of the second event. More on this later
2020 2020 SlideSlide
SlideSlide
Mutually Mutually ExclusiveExclusive
Two events A and B are mutually exclusive if and only if:
In a Venn diagram this means that event A is disjoint from event B.
A and B are M.E.
0)( BAP
A BA B
A and B are not M.E.
2121 2121 SlideSlide
SlideSlide
The Addition The Addition RuleRule
The probability that at least one of the events A or B will occur, P(A or B), is given by:
If events A and B are mutually exclusive, then the addition rule is simplified to:
This simplified rule can be extended to any number of mutually exclusive events.
)()()()()( BAPBPAPBAPBorAP
)()()()( BPAPBAPBorAP
2222 2222 SlideSlide
SlideSlide
Conditional Conditional ProbabilityProbability
Conditional probability is the probability of an event occurring, given that another event has already occurred. Conditional probability restricts the sample space.The conditional probability of event B occurring, given that event A has occurred, is denoted by P(B|A) and is read as “probability of B, given A.” We use conditional probability when two events occurring in sequence are not independent. In other words, the fact that the first event (event A) has occurred affects the probability that the second event (event B) will occur.
2323 2323 SlideSlide
SlideSlide
Conditional Conditional ProbabilityProbability
Formula for Conditional Probability
Better off to use your brain and work out conditional probabilities from looking at the sample space, otherwise use the formula.
)(
)()|(
)(
)()|(
AP
ABPABPor
BP
BAPBAP
2424 2424 SlideSlide
SlideSlide
Assigning ProbabilitiesAssigning Probabilities Two basic requirements for assigning probabilities
1. The probability assigned to each experimental outcome must be between 0 and 1, inclusively. If we let Ei denote the ith experimental outcome and P(Ei) its probability, then this requirement can be written as
0 P(Ei) 1 for all I
2. The sum of the probabilities for all the experimental outcomes must equal 1.0. For n experimental outcomes, this requirement can be written as
P(E1)+ P(E2)+… + P(En) =1
2525 2525 SlideSlide
SlideSlide
Classical MethodClassical Method If an experiment has n possible outcomes, this method would assign a probability of 1/n to each outcome.
Experiment: Rolling a die
Sample Space: S = {1, 2, 3, 4, 5, 6}
Probabilities: Each sample point has a 1/6 chance of occurring
Example
2626 2626 SlideSlide
SlideSlide
2727 2727 SlideSlide
SlideSlide
THEORETICAL PROBABILITY
I have a quarterMy quarter has a heads side and a tails side
Since my quarter has only 2 sides, there are only 2 possible outcomes when I flip it. It will either land on heads, or tails
HEADS
TAILS
2828 2828 SlideSlide
SlideSlide
THEORETICAL PROBABILITY
When I flip my coin, the probability that my coin will land on heads is 1 in 2
What is the probability that my coin will land on tails??
HEADS
TAILS
2929 2929 SlideSlide
SlideSlide
Theoretical Probability
Right!!! There is a 1 in 2 probability that my coin will land on tails!!!
HEADS
TAILS
A probability of 1 in 2 can be written in three ways:
•As a fraction: ½•As a decimal: .50
•As a percent: 50%
3030 3030 SlideSlide
SlideSlide
I am going to take 1 marble from the bag. What is the probability that I will pick out
a red marble?
Theoretical Probability
I have three marbles in a bag.
1 marble is red
1 marble is blue
1 marble is green
3131 3131 SlideSlide
SlideSlide
Theoretical Probability
Since there are three marbles and only one is red, I have a 1 in 3 chance of picking out a red marble.
I can write this in three ways:
As a fraction: 1/3 As a decimal: .33 As a percent: 33%
3232 3232 SlideSlide
SlideSlide
Experimental Probability
Experimental probability is found by repeating an experiment and observing the outcomes.
3333 3333 SlideSlide
SlideSlide
Experimental Probability
Remember the bag of marbles? The bag has only 1 red, 1 green,
and 1 blue marble in it. There are a total of 3 marbles in
the bag. Theoretical Probability says there
is a 1 in 3 chance of selecting a red, a green or a blue marble.
3434 3434 SlideSlide
SlideSlide
Experimental Probability
Draw 1 marble from the bag.
Marble number red blue green
1 123456
It is a red marble.
Record the outcome on the tally sheet
3535 3535 SlideSlide
SlideSlide
Experimental Probability
Put the red marble back in the bag and draw again.
This time your drew a green marble. Record this outcome on the tally sheet.
Marble number red blue green
1 12 134
3636 3636 SlideSlide
SlideSlide
Experimental Probability
Place the green marble back in the bag. Continue drawing marbles and recording
outcomes until you have drawn 6 times. (remember to place each marble back in the bag before drawing again.)
3737 3737 SlideSlide
SlideSlide
Experimental Probability
After 6 draws your chart will look similar to this.
Look at the red column.
Of our 6 draws, we selected a red marble 2 times.
Marble number red blue green
1 12 13 14 15 16 1
Total 2 1 3
3838 3838 SlideSlide
SlideSlide
Experimental Probability
The experimental probability of drawing a red marble was 2 in 6.
This can be expressed as a fraction: 2/6 or 1/3 a decimal : .33 or a percentage: 33%
Marble number red blue green
1 12 13 14 15 16 1
Total 2 1 3
3939 3939 SlideSlide
SlideSlide
Experimental Probability
Notice the Experimental Probability of drawing a red, blue or green marble.
Marble number red blue green
1 12 13 14 15 16 1
Total 2 1 3
Exp. Prob.
2/6 or 1/3 1/6
3/6 or 1/2
4040 4040 SlideSlide
SlideSlide
Comparing Experimental and Theoretical Probability
Look at the chart at the right.
Is the experimental probability always the same as the theoretical probability?
red blue greenExp. Prob. 1/3 1/6 1/2Theo. Prob. 1/3 1/3 1/3
4141 4141 SlideSlide
SlideSlide
Comparing Experimental and Theoretical Probability
In this experiment, the experimental and theoretical probabilities of selecting a red marble are equal.
red blue greenExp. Prob. 1/3 1/6 1/2Theo. Prob. 1/3 1/3 1/3
4242 4242 SlideSlide
SlideSlide
Comparing Experimental and Theoretical Probability
The experimental probability of selecting a blue marble is less than the theoretical probability.
The experimental probability of selecting a green marble is greater than the theoretical probability.
red blue greenExp. Prob. 1/3 1/6 1/2Theo. Prob. 1/3 1/3 1/3
4343 4343 SlideSlide
SlideSlide
Point and interval estimations of parameters of the normally up-diffused sign. Concept of statistical evaluation.
4444 4444 SlideSlide
SlideSlide
What is statistics?
a branch of mathematics that provides techniques to analyze whether or not your data is significant (meaningful)
Statistical applications are based on probability statements
Nothing is “proved” with statistics Statistics are reported Statistics report the probability that similar
results would occur if you repeated the experiment
4545 4545 SlideSlide
SlideSlide
Statistics deals with numbers
Need to know nature of numbers collectedContinuous variables: type of numbers associated
with measuring or weighing; any value in a continuous interval of measurement. Examples:
Weight of students, height of plants, time to flowering
Discrete variables: type of numbers that are counted or categorical Examples:
Numbers of boys, girls, insects, plants
4646 4646 SlideSlide
SlideSlide
Standard Deviation and Variance Standard deviation and variance are the
most common measures of total risk
They measure the dispersion of a set of observations around the mean observation
4747 4747 SlideSlide
SlideSlide
Standard Deviation and Variance (cont’d) General equation for variance:
If all outcomes are equally likely:
2
2
1
Variance prob( )n
i ii
x x x
2
2
1
1 n
ii
x xn
4848 4848 SlideSlide
SlideSlide
Standard Deviation and Variance (cont’d) Equation for standard deviation:
2
2
1
Standard deviation prob( )n
i ii
x x x
4949 4949 SlideSlide
SlideSlide
1.The Normal distribution – parameters and (or 2)
Comment: If = 0 and = 1 the distribution is called the standard normal distribution
0
0.005
0.01
0.015
0.02
0.025
0.03
0 20 40 60 80 100 120
Normal distribution with = 50 and =15
Normal distribution with = 70 and =20
5050 5050 SlideSlide
SlideSlide
2221
( ) e ,2
x
f x x
The probability density of the normal distribution
If a random variable, X, has a normal distribution with mean and variance 2 then we will write:
2~ ,X N
5151 5151 SlideSlide
SlideSlide
The Chi-square (2) distribution with d.f.
21
2 2
112
2
0
0 0
xx e xf x
x
2
2 2
22
10
2
0 0
x
x e x
x
The Chi-square distribution
5252 5252 SlideSlide
SlideSlide
0
0.1
0.2
0 4 8 12 16
Graph: The 2 distribution
( = 4)
( = 5)
( = 6)
5353 5353 SlideSlide
SlideSlide
1. If z has a Standard Normal distribution then z2
has a 2 distribution with 1 degree of freedom.
Basic Properties of the Chi-Square distribution
2. If z1, z2,…, z are independent random variables each having Standard Normal distribution then
has a 2 distribution with degrees of freedom.
2 2 21 2 ...U z z z
3. Let X and Y be independent random variables having a 2 distribution with 1 and 2 degrees of freedom respectively then X + Y has a 2
distribution with degrees of freedom 1 + 2.
5454 5454 SlideSlide
SlideSlide
continued
4. Let x1, x2,…, xn, be independent random variables having a 2 distribution with 1 , 2 ,…, n degrees of freedom respectively then x1+ x2 +…+ xn has a 2 distribution with degrees of freedom 1 +…+ n.
5. Suppose X and Y are independent random variables with X and X + Y having a 2 distribution with 1
and ( > 1 ) degrees of freedom respectively then Y has a 2 distribution with degrees of freedom - 1.
5555 5555 SlideSlide
SlideSlide
The non-central Chi-squared distribution
If z1, z2,…, z are independent random variables each having a Normal distribution with mean i and variance 2 = 1, then
has a non-central 2 distribution with degrees of freedom and non-centrality parameter
2 2 21 2 ...U z z z
1
221
ii
5656 5656 SlideSlide
SlideSlide
Mean and Variance of non-central 2 distribution
If U has a non-central 2 distribution with degrees of freedom and non-centrality parameter
1
221
ii
Then
1
22i
iUE 42 UVar
If U has a central 2 distribution with degrees of freedom and is zero, thus
UE 2UVar
5757 5757 SlideSlide
SlideSlide
Estimation of Population Parameters
Statistical inference refers to making inferences about a population parameter through the use of sample information
The sample statistics summarize sample information and can be used to make inferences about the population parameters
Two approaches to estimate population parameters Point estimation: Obtain a value estimate for the population parameter Interval estimation: Construct an interval within which the population
parameter will lie with a certain probability
5858 5858 SlideSlide
SlideSlide
Point Estimation
In attempting to obtain point estimates of population parameters, the following questions arise What is a point estimate of the population mean? How good of an estimate do we obtain through the methodology that we
follow?
Example: What is a point estimate of the average yield on ten-year Treasury bonds?
To answer this question, we use a formula that takes sample information and produces a number
5959 5959 SlideSlide
SlideSlide
Point Estimation
A formula that uses sample information to produce an estimate of a population parameter is called an estimator
A specific value of an estimator obtained from information of a specific sample is called an estimate
Example: We said that the sample mean is a good estimate of the population mean The sample mean is an estimator A particular value of the sample mean is an estimate
6060 6060 SlideSlide
SlideSlide
Interval Estimation
In the probabilistic interpretation, we say that
A 95% confidence interval for a population parameter means that, in repeated sampling, 95% of such confidence intervals will include the population parameter
In the practical interpretation, we say that
We are 95% confident that the 95% confidence interval will include the population parameter
6161 6161 SlideSlide
SlideSlide
Constructing Confidence Intervals
Confidence intervals have similar structures
Point Estimate Reliability Factor Standard Error
Reliability factor is a number based on the assumed distribution of the point estimate and the level of confidence
Standard error of the sample statistic providing the point estimate
6262 6262 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
If is the sample mean, then we are interested in the confidence interval, such that the following probability is .9
X
nX
nXP
nX
nP
n
XP
ZP
645.1645.1
645.1645.1
645.1/
645.1
645.1645.19.
6363 6363 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
Following the above expression for the structure of a confidence interval, we rewrite the confidence interval as follows
Note that from the standard normal density
nX
645.1
05.065.1)65.1(
95.065.165.1
Z
Z
FZP
FZP
6464 6464 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
Given this result and that the level of confidence for this interval (1-) is .90, we conclude that
The area under the standard normal to the left of –1.65 is 0.05 The area under the standard normal to the right of 1.65 is 0.05
Thus, the two reliability factors represent the cutoffs -z/2 and z/2 for the standard normal
6565 6565 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
In general, a 100(1-)% confidence interval for the population mean when we draw samples from a normal distribution with known variance 2 is given by
where z/2 is the number for which
nzX
2/
22/
zZP
6666 6666 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
Note: We typically use the following reliability factors when constructing confidence intervals based on the standard normal distribution
90% interval: z0.05 = 1.65
95% interval: z0.025 = 1.96
99% interval: z0.005 = 2.58
Implication: As the degree of confidence increases the interval becomes wider
6767 6767 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Known Variance
Example: Suppose we draw a sample of 100 observations of returns on the Nikkei index, assumed to be normally distributed, with sample mean 4% and standard deviation 6%
What is the 95% confidence interval for the population mean?
The standard error is .06/ = .006
The confidence interval is .04 1.96(.006)
100
6868 6868 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Unknown Variance
In a more typical scenario, the population variance is unknown
Note that, if the sample size is large, the previous results can be modified as follows The population distribution need not be normal The population variance need not be known The sample standard deviation will be a sufficiently good estimator of
the population standard deviation
Thus, the confidence interval for the population mean derived above can be used by substituting s for
6969 6969 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Unknown Variance
However, if the sample size is small and the population variance is unknown, we cannot use the standard normal distribution
If we replace the unknown with the sample st. deviation s the following quantity
follows Student’s t distribution with (n – 1) degrees of freedom
ns
Xt
/
7070 7070 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Unknown Variance
The t-distribution has mean 0 and (n – 1) degrees of freedom
As degrees of freedom increase, the t-distribution approaches the standard normal distribution
Also, t-distributions have fatter tails, but as degrees of freedom increase (df = 8 or more) the tails become less fat and resemble that of a normal distribution
7171 7171 SlideSlide
SlideSlide
Confidence Interval for Mean of a Normal Distribution with Unknown Variance
In general, a 100(1-)% confidence interval for the population mean when we draw small samples from a normal distribution with an unknown variance 2 is given by
where tn-1,/2 is the number for which
n
stX n 2/,1
22/,11
nn ttP
7272 7272 SlideSlide
SlideSlide
Confidence Interval for the Population Variance of a Normal Population
Suppose we have obtained a random sample of n observations from a normal population with variance 2 and that the sample variance is s2. A 100(1 - )% confidence interval for the population variance is
2
2/1,1
22
22/,1
2 11
nn
snsn
7373 7373 SlideSlide
SlideSlide
EndEnd
top related