probability and statistics for management and engineering basics
TRANSCRIPT
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
1/17
Probability and Probability Distribution
Probability is the measure of the likeliness that anevent will occur.Probabilityis used to quantify an attitude of mind towards some proposition of whose
truth we are not certain. The proposition of interest is usually of the form "Will
a specificevent occur?" The attitude of mind is of the form "How certain are we
that the event will occur?" The certainty we adopt can be described in terms of
a numerical measure and this number, between 0 and 1 (where 0 indicates
impossibility and 1 indicates certainty), we call probability . Thus the higher
the probability of an event, the more certain we are that the event will occur. A
simple example would be the toss of a fair coin. Since the 2 outcomes are
deemed equiprobable, the probability of "heads" equals the probability of
"tails" and each probability is 1/2 or equivalently a 50% chance of either
"heads" or "tails".
History of Probability
"A gambler's dispute in 1654 led to the creation of a mathematical theory of
probability by two famous French mathematicians, Blaise Pascal and Pierre de
Fermat. Antoine Gombaud, Chevalier de Mr, a French nobleman with an
interest in gaming and gambling questions, called Pascal's attention to an
apparent contradiction concerning a popular dice game. The game consisted in
throwing a pair of dice 24 times; the problem was to decide whether or not to
bet even money on the occurrence of at least one "double six" during the 24
throws. A seemingly well-established gambling rule led de Mr to believe that
betting on a double six in 24 throws would be profitable, but his own
calculations indicated just the opposite.
This problem and others posed by de Mr led to an exchange of letters
between Pascal and Fermat in which the fundamental principles of probability
theory were formulated for the first time. Although a few special problems on
games of chance had been solved by some Italian mathematicians in the 15th
and 16th centuries, no general theory was developed before this famous
correspondence.
The Dutch scientist Christian Huygens, a teacher of Leibniz, learned of this
correspondence and shortly thereafter (in 1657) published the first book on
http://en.wikipedia.org/wiki/Measure_%28mathematics%29http://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Propositionhttp://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Propositionhttp://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Measure_%28mathematics%29 -
8/10/2019 Probability and Statistics for Management And Engineering Basics
2/17
probability; entitled De Ratiociniis in Ludo Aleae, it was a treatise on problems
associated with gambling. Because of the inherent appeal of games of chance,
probability theory soon became popular, and the subject developed rapidly
during the 18th century. The major contributors during this period were Jakob
Bernoulli (1654-1705) and Abraham de Moivre (1667-1754).
In 1812 Pierre de Laplace (1749-1827) introduced a host of new ideas and
mathematical techniques in his book, Thorie Analytique des Probabilits.
Before Laplace, probability theory was solely concerned with developing a
mathematical analysis of games of chance. Laplace applied probabilistic ideas
to many scientific and practical problems. The theory of errors, actuarial
mathematics, and statistical mechanics are examples of some of the important
applications of probability theory developed in the l9th century.
Like so many other branches of mathematics, the development of probability
theory has been stimulated by the variety of its applications. Conversely, each
advance in the theory has enlarged the scope of its influence. Mathematical
statistics is one important branch of applied probability; other applications
occur in such widely different fields as genetics, psychology, economics, and
engineering. Many workers have contributed to the theory since Laplace's time;
among the most important are Chebyshev, Markov, von Mises, and Kolmogorov.
One of the difficulties in developing a mathematical theory of probability has
been to arrive at a definition of probability that is precise enough for use in
mathematics, yet comprehensive enough to be applicable to a wide range of
phenomena. The search for a widely acceptable definition took nearly three
centuries and was marked by much controversy. The matter was finally
resolved in the 20th century by treating probability theory on an axiomatic
basis. In 1933 a monograph by a Russian mathematician A. Kolmogorov
outlined an axiomatic approach that forms the basis for the modern theory.
(Kolmogorov's monograph is available in English translation as Foundations of
Probability Theory, Chelsea, New York, 1950.) Since then the ideas have been
refined somewhat and probability theory is now part of a more general
discipline known as measure theory."
Basic Terminology
Trial
Can refer to each individual repetition when talking about an experiment
composed of any fixed number of them. As an example, one can think of an
http://en.wikipedia.org/wiki/Experiment_%28probability_theory%29http://en.wikipedia.org/wiki/Experiment_%28probability_theory%29 -
8/10/2019 Probability and Statistics for Management And Engineering Basics
3/17
experiment being any number from one to n coin tosses, say 17. In this case,
one toss can be called a trial to avoid confusion, since the whole experiment is
composed of 17 ones.
Sample space
The set of possible outcomes of an experiment. For example, the sample space
for rolling a six-sided die will be {1, 2, 3, 4, 5, 6}
Event
A subset of the sample space (a possible experiment's outcome), to which a
probability can be assigned. For example, on rolling a die, "getting a five or a
six" is an event (with a probability of one third if the die is fair)
Compound Event
A compound event consists of two or more simple events. Tossing two dice is a
compound event.
Exhaustive Events
The total possible outcomes of a trail.For example in In a throw of a
die,Number of exhaustive events = 6
Favorable Events:The outcomes of a trail which cause the happening of a particular event.For
example in a event A = Getting an even number = {2, 4, 6},Number of favorable
events = 3
Equally likely Events:
The events are said to be equally likely events, if none of them is expected to
occur in preference to other.For example In a toss of an unbiased coin,P (H) = P
(T) =
Mutually Exclusive/ Disjoint Event
The events which can not occur simultaneously.For example In a draw of a card
from a deck of playing car ds if A = The card drawn is a club B = The card drawn
http://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Event_%28probability_theory%29http://en.wikipedia.org/wiki/Sample_space -
8/10/2019 Probability and Statistics for Management And Engineering Basics
4/17
is a heart then Events A and B are mutually exclusive events.
Complementary Event()
The non-happening of event E is called complementary event EC of event E.
P() = 1 P(E).For example if probablity of getting a head is 0.4 thenprobabilty of tail is 0.6.
Independent Event
The happening/non-happening of one event does not depend on the occurrence
of other event
Dependent Event
The events which are not independent events.
Fomulae for Probablitity
If a random experiment results in n exhaustive, mutually exclusive and equally
likely events, out of which m are favorable to the happening of event E, then
the probability of occurrence of event E is
Probability can be expressed in terms of fraction, percentage, decimal or ratio.
Facts about probability
Probability of each event is a number between 0 and 1 incl usive i. e., 0
P(E) 1
Probability of impossible event is zero.
[Note: the converse is not necessarily true]
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
5/17
Probability of certain event is one.
The sum of probabilities of all possible events is equals to one i.e., P(E)
=1
Approaches of Probability
Classical theory of probability
The classical approach to probability is to count the number of favorable
outcomes, the number of total outcomes (outcomes are assumed to be mutually
exclusive and equiprobable), and express the probability as a ratio of these two
numbers. Here, "favorable" refers not to any subjective value given to the
outcomes, but is rather the classical terminology used to indicate that anoutcome belongs to a given event of interest. What is meant by this will be
made clear by an example, and formalized with the introduction of axiomatic
probability theory.
Classical definition of probability
If the number of outcomes belonging to an event is ,
and the total number of outcomes is , then the
probability of event is defined as .
Empirical or Statistical Probability or Frequency of
occurrence
This approach to probability is well-suited to a wide range of scientific
disciplines. It is based on the idea that the underlying probability of an event
can be measured by repeated trials.
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
6/17
Empirical or Statistical Probability as a measure of
frequency
Let be the number of times event occurs after
trials. We define the probability of event as
It is of course impossible to conduct an infinite number of trials. However, it
usually suffices to conduct a large number of trials, where the standard of large
depends on the probability being measured and how accurate a measurement
we need.
Axiomatic probability theory
Axiomatic probability theory, although it is often frightening to beginners, is
the most general approach to probability, and has been employed in tackling
some of the more difficult problems in probability. We start with a set of
axioms, which serve to define a probability space. Although these axioms may
not be immediately intuitive, be assured that the development is guided by the
more familiar classical probability theory. Let S be the sample space of a
random experiment. The probability P is a real valued function whose domain is
the power set of S and range is the interval [0,1] satisfying the following
axioms:
(i) For any event E, P (E ) 0
(ii) P (S) = 1
(iii) If E and F are mutually exclusive events, then P(E F) = P(E) + P(F).
It follows from (iii) that P() = 0. To prove this, we take F = and note that E
and are disjoint events. Therefore, from axiom (iii), we get P (E ) = P (E) +P () or P(E) = P(E) + P () i.e. P () = 0. Let S be a sample space containing
outcomes 1 , 2 ,...,n , i.e., S = {1, 2, ... , n}
It follows from the axiomatic definition of probability that:
(i) 0 P (i) 1 for each i S
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
7/17
(ii) P (1) + P (2) + ... + P (n) = 1
(iii) For any event A, P(A) = P (i ), i A .
Conditional Probability P(A|B) : The probability of event A provided
event B has already happened.
P (A|B) =
Concept
If an event B has occurred, instead of S, we consider B only.
The conditional probability of A given B will be the ratio of that part of A which
is included in B i.e. P(AB) to the probability of B.
Addition Theorem:For two events A and B, probability of happeningatleast one of them is
P(AB) = P(A) + P(B) P (AB)
Multiplication Theorem : For two events A and B, probability of theirsimultaneous happening is P(A B) = P(A) * P(B|A), P(A) > 0
Or P(A B) = P(B) *P(A|B), P(B) > 0
If the events A and B are independent i.e. P(A|B) = P(A) & P(B|A) = P(B), then
P(A B) = P(A) *P(B)
Bayes Theorem (aka, Bayes Rule)
Bayes' theoremprovides a mathematicalrule for revising anestimate or
forecast in light of experience and observation. It differs from other methods of
hypothesis testing in that itassigns 'after thefact'(posterior)probabilities to
the hypotheses instead of just accepting or rejecting them. Named after its
proponent, the UK mathematician Thomas Bayes (1702-1761) who researched
probability andstatistical inference
)(
)(
BP
BAP
http://www.businessdictionary.com/definition/provide.htmlhttp://www.businessdictionary.com/definition/rule.htmlhttp://www.businessdictionary.com/definition/estimate.htmlhttp://www.businessdictionary.com/definition/forecast.htmlhttp://www.businessdictionary.com/definition/experience.htmlhttp://www.businessdictionary.com/definition/method.htmlhttp://www.businessdictionary.com/definition/hypothesis-testing.htmlhttp://www.businessdictionary.com/definition/assign.htmlhttp://www.businessdictionary.com/definition/fact.htmlhttp://www.businessdictionary.com/definition/probability.htmlhttp://www.businessdictionary.com/definition/statistical-inference.htmlhttp://www.businessdictionary.com/definition/statistical-inference.htmlhttp://www.businessdictionary.com/definition/probability.htmlhttp://www.businessdictionary.com/definition/fact.htmlhttp://www.businessdictionary.com/definition/assign.htmlhttp://www.businessdictionary.com/definition/hypothesis-testing.htmlhttp://www.businessdictionary.com/definition/method.htmlhttp://www.businessdictionary.com/definition/experience.htmlhttp://www.businessdictionary.com/definition/forecast.htmlhttp://www.businessdictionary.com/definition/estimate.htmlhttp://www.businessdictionary.com/definition/rule.htmlhttp://www.businessdictionary.com/definition/provide.html -
8/10/2019 Probability and Statistics for Management And Engineering Basics
8/17
Applications of Bayes theorem
Bayes's theorem is significantly important in inverse problem theory, where the
a posterioriprobability density function is obtained from the product of priorprobability density function and the likelihood probability density function. An
important application is constructing computational models of oil reservoirs
given the observed data.
Although Bayes's theorem is commonly used to determine the probability of an
event occurring, it can also be applied to verify someone's credibility as a
prognosticator. Many pundits claim to be able to predict the outcome of an
event; political elections, trials, the weather and even sporting events. Larry
Sabato,founder ofSabato's Crystal Ball, is a perfect example. His website
provides free political analysis and election predictions. His success at
predictions has even led him to be called a "pundit with an opinion for every
reporter's phone call." We even havePunxsutawney Phil,the famous
groundhog, who tells us whether or not we can expect a longer winter or an
early spring. Bayes's theorem tells us the difference between who's on a hot
streak and who is what they claim to be.
Let's say we live in an area where everyone gambles on the outcome of coin
flips. Because of that, there is a big business for predicting coin flips. Suppose
that 5% of predictors can actually win in the long run, and 80% of those are
winners over a 2-year period. 95% of predictors are pretenders who are just
guessing, and 20% of them are winners over a 2-year period (everyone gets
lucky once in a while). This means that 82.6% of bettors who are winners over
a 2-year period are actually long-term losers who are winning above their real
average.
Probability Distribution
Random Variables
Introduction
A random variable associates each outcome of a random experiment with a
probability and it represents a measurable aspect or property of the outcomes.
http://en.wikipedia.org/wiki/Probability_density_functionhttp://en.wikipedia.org/wiki/Larry_Sabatohttp://en.wikipedia.org/wiki/Larry_Sabatohttp://en.wikipedia.org/wiki/Sabato%27s_Crystal_Ballhttp://en.wikipedia.org/wiki/Predictionshttp://en.wikipedia.org/wiki/Punxsutawney_Philhttp://en.wikipedia.org/wiki/Punxsutawney_Philhttp://en.wikipedia.org/wiki/Predictionshttp://en.wikipedia.org/wiki/Sabato%27s_Crystal_Ballhttp://en.wikipedia.org/wiki/Larry_Sabatohttp://en.wikipedia.org/wiki/Larry_Sabatohttp://en.wikipedia.org/wiki/Probability_density_function -
8/10/2019 Probability and Statistics for Management And Engineering Basics
9/17
For these reasons, a random variable is essential to be defined to assume a real
number to each and every outcome of the random experiment.
Random Variable
Let S be a sample space of a random experiment. A random variable is afunction that associates a real number with each element in the sample space, S.
Types of Random Variables
Random variables are commonly classified into two types namely, discrete
random variables and continuous random variables.
Discrete Random Variable
A random variable on a random sample S is said to be discrete if the range of X
is countable. In other words, a discrete random variable is a random variable
that can assume at most a finite or a countably infinite number of possible
values.
Probability Function (or) Probability Mass Function of a
Discrete Random Variable
Let X be a discrete random variable on a sample space S, which can take the
values
,1x ,2x
nx
such thatiii pxPxXP )()(
(called the probability ofix
).The function i
pis called the probability mass function (p.m.f) of the random
variable X i f ip
must satisfy the following conditions: (i)0ip for all i and
(ii)
n
iip 1
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
10/17
Distribution Function of the Discrete Random Variable
The distribution function or briefly the cumulative distribution function (c.d.f)
for a discrete random variable X is defined by
x
iX xXPxXPxFxF )()()()(
where x is any real number, that is, x .
Continuous random variable
A random variable X is said to be continuous if it takes all possible values
between certain limits o r in an interval which may be finite or infinite.
Probability Density function of a Continuous Random Variable
If X is a continuous random variable such that dxxf
dx
xX
dx
xP )(}22{ ,
then )(xf is called the probability density function (p.d.f) of X provided )(xf
satisfies the following conditions:
(i) 0)( xf , for all Rx ; (ii)
1)( dxxf
and (iii)
b
a
dxxfbxaP )()(
.
Distribution Function of the Continuous Random Variable
The cumulative distribution function (c.d.f.) or briefly the distribution function
for a continuous random variable X is defined by
x
X dxxfxXPxFxF )()()()(
, . x
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
11/17
Relationship between probability distribution function )(xF and probability
density function )(xf :
From the definition of probability distribution function, we can conclude
that)())(( xfxF
dx
d
, that is, the probability density function of a continuous
random variable X , )(xf is the derivative of the probability distribution
function of the random variable, )(xF
Binomial distribution
Frequency distribution where only two (mutually exclusive)outcomes are
possible, such as better or worse, gain or loss, head ortail,rise or fall,success
or failure,yes or no. Therefore, if the probability of success in any given trial is
known, binomial distributions can beemployed to compute a given number of
successes in a given number of trials.And it can be determined if anempirical
distribution deviates significantly from a typical outcome. Alsocalled Bernoullidistribution after its discoverer, the Swiss mathematician Jacques Bernoulli
(1654-1705)
A binomial experiment is one that possesses the following properties:
The experiment consists of n repeated trials;
Each trial results in an outcome that may be classified as a success or a failure
(hence the name, binomial);
The probability of a success, denoted by p, remains constant from trial to trialand repeated trials are independent.
The number of successes X in n trials of a binomial experiment is called a
binomial random variable.
The probability distribution of the random variable X is called a binomial
distribution, and is given by the formula:
http://www.businessdictionary.com/definition/frequency-distribution.htmlhttp://www.businessdictionary.com/definition/mutually-exclusive.htmlhttp://www.businessdictionary.com/definition/gain.htmlhttp://www.businessdictionary.com/definition/tail.htmlhttp://www.businessdictionary.com/definition/success.htmlhttp://www.businessdictionary.com/definition/failure.htmlhttp://www.businessdictionary.com/definition/probability.htmlhttp://www.businessdictionary.com/definition/distribution.htmlhttp://www.businessdictionary.com/definition/employed.htmlhttp://www.businessdictionary.com/definition/trial.htmlhttp://www.businessdictionary.com/definition/empirical.htmlhttp://www.businessdictionary.com/definition/typical.htmlhttp://www.businessdictionary.com/definition/call.htmlhttp://www.businessdictionary.com/definition/call.htmlhttp://www.businessdictionary.com/definition/typical.htmlhttp://www.businessdictionary.com/definition/empirical.htmlhttp://www.businessdictionary.com/definition/trial.htmlhttp://www.businessdictionary.com/definition/employed.htmlhttp://www.businessdictionary.com/definition/distribution.htmlhttp://www.businessdictionary.com/definition/probability.htmlhttp://www.businessdictionary.com/definition/failure.htmlhttp://www.businessdictionary.com/definition/success.htmlhttp://www.businessdictionary.com/definition/tail.htmlhttp://www.businessdictionary.com/definition/gain.htmlhttp://www.businessdictionary.com/definition/mutually-exclusive.htmlhttp://www.businessdictionary.com/definition/frequency-distribution.html -
8/10/2019 Probability and Statistics for Management And Engineering Basics
12/17
where
n = the number of trials
x = 0, 1, 2, .. . n
p = the probability of success in a single trial
q = the probability of failure in a single trial
(i.e. q = 1 p)
P(X) gives the probability of successes in n binomial trials.
Mean and Variance of Binomial Distribution
If p is the probability of success and q is the probability of failure in a binomial
trial, then the expected number of successes in n trials (i.e. the mean value of
the binomial distribution) is
E(X) = = np
The variance of the binomial distribution is
V(X) = 2 = npq
Note: In a binomial distribution, only 2 parameters, namely n and p, are needed
to determine the probability.
Poisson Distribution
The Poisson Distribution was developed by the French mathematician Simeon
Denis Poisson in 1837.
The Poisson random variable satisfies the following conditions:
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
13/17
The number of successes in two disjoint time intervals is independent.
The probability of a success during a small time interval is proportional
to the entire length of the time interval.
Apart from disjoint time intervals, the Poisson random variable alsoapplies to disjoint regions of space.
Applications
the number of deaths by horse kicking in the P russian army (first
application)
birth defects and genetic mutations
rare diseases (like Leukemia, but not AIDS because it is infectious and so
not independent) - especially in legal cases
car accidents
traffic flow and ideal gap distance
number of typing errors on a page
spread of an endangered animal in Africa
failure of a machine in one month
The probability distribution of a Poisson random variable X representing the
number of successes occurring in a given time interval or a specified region of
space is given by the formula:
Mean and Variance of Poisson Distribution
If is the average number of successes occurring in a given time interval or
region in the Poisson distribution, then the mean and the variance of the
Poisson distribution are both equal to .
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
14/17
E(X) = and V(X) = 2 =
Note: In a Poisson distribution, only one parameter, is needed to determine
the probability of an event
Normal Distibrution
The Normal Probability Distribution is very common in the field of statistics.
Whenever you measure things like people's height, weight, salary, opinions or
votes, the graph of the results is very often a normal curve.
The Normal Distribution
A random variable X whose distribution has the shape of a normal curve iscalled a normal random variable.
Normal Curve
This random variable X is said to be normally distributed with mean and
standard deviation if its probability dist ribution is given by
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
15/17
Properties of a Normal Distribution
The normal curve is symmetrical about the mean ;
The mean is at the middle and divides the area into halves;
The total area under the curve is equal to 1;
It is completely determined by its mean and standar d deviation (or
variance 2)
Note: In a normal distribution, on ly 2 parameters are needed, namely and 2.
Area Under the Normal Curve using Integration
The probability of a continuous normal variable X found in a particular interval
[a, b] is the area under the curve bounded by x=a and x=b and is given by
P(a
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
16/17
We can transform all the observations of any normal random variable X with
mean and variance to a new set o f observations of another normal random
variable Z with mean 0 and variance 1 using the following transformation:
Z=
A standard normal distribution is a normal distribution with mean 0 and
standard deviation 1. Areas under this curve can be found using a standard
normal table.
Standard Normal table
-
8/10/2019 Probability and Statistics for Management And Engineering Basics
17/17