Sie sind auf Seite 1von 91

Chapter 2

Probability Concepts and


Applications

To accompany
Quantitative Analysis for Management, Tenth Edition,
by Render, Stair, and Hanna © 2008 Prentice-Hall, Inc.
Power Point slides created by Jeff Heyl © 2009 Prentice-Hall, Inc.
Learning Objectives
After completing this chapter, students will be able to:
1. Understand the basic foundations of
probability analysis
2. Describe statistically dependent and
independent events
3. Use Bayes’ theorem to establish posterior
probabilities
4. Describe and provide examples of both
discrete and continuous random variables
5. Explain the difference between discrete and
continuous probability distributions
6. Calculate expected values and variances
and use the normal table
© 2008 Prentice-Hall, Inc. 2–2
Chapter Outline
2.1 Introduction
2.2 Fundamental Concepts
2.3 Mutually Exclusive and Collectively
Exhaustive Events
2.4 Statistically Independent Events
2.5 Statistically Dependent Events
2.6 Revising Probabilities with Bayes’
Theorem
2.7 Further Probability Revisions

© 2008 Prentice-Hall, Inc. 2–3


Chapter Outline
2.8 Random Variables
2.9 Probability Distributions
2.10 The Binomial Distribution
2.11 The Normal Distribution
2.12 The F Distribution
2.13 The Exponential Distribution
2.14 The Poisson Distribution

© 2008 Prentice-Hall, Inc. 2–4


Introduction

 Life is uncertain, we are not sure


what the future will bring
 Risk and probability is a part of
our daily lives
 Probability is a numerical
statement about the likelihood
that an event will occur

© 2008 Prentice-Hall, Inc. 2–5


Fundamental Concepts

1. The probability, P, of any event or


state of nature occurring is greater
than or equal to 0 and less than or
equal to 1. That is:

0  P (event)  1

2. The sum of the simple probabilities


for all possible outcomes of an
activity must equal 1
© 2008 Prentice-Hall, Inc. 2–6
Chapters in This Book
That Use Probability
CHAPTER TITLE
3 Decision Analysis
4 Regression Models
5 Forecasting
6 Inventory Control Models
13 Project Management
14 Waiting Lines and Queuing Theory Models
15 Simulation Modeling
16 Markov Analysis
17 Statistical Quality Control
Module 3 Decision Theory and the Normal Distribution
Module 4 Game Theory

Table 2.1
© 2008 Prentice-Hall, Inc. 2–7
Diversey Paint Example
 Demand for white latex paint at Diversey Paint
and Supply has always been either 0, 1, 2, 3, or 4
gallons per day
 Over the past 200 days, the owner has observed
the following frequencies of demand
QUANTITY
NUMBER OF DAYS PROBABILITY
DEMANDED
0 40 0.20 (= 40/200)
1 80 0.40 (= 80/200)
2 50 0.25 (= 50/200)
3 20 0.10 (= 20/200)
4 10 0.05 (= 10/200)
Total 200 Total 1.00 (= 200/200)

© 2008 Prentice-Hall, Inc. 2–8


Diversey Paint Example
 Demand for white latex paint at Diversey Paint
and Supply
Notice has always
the individual been either 0, 1, 2, 3, or 4
probabilities
aregallons per day
all between 0 and 1
 Over the0 ≤past 200 days,
P (event) ≤ 1 the owner has observed
the following frequencies of demand
And the total of all event
QUANTITY equals 1
probabilities NUMBER OF DAYS PROBABILITY
DEMANDED
0 ∑ P (event) = 1.00
40 0.20 (= 40/200)
1 80 0.40 (= 80/200)
2 50 0.25 (= 50/200)
3 20 0.10 (= 20/200)
4 10 0.05 (= 10/200)
Total 200 Total 1.00 (= 200/200)

© 2008 Prentice-Hall, Inc. 2–9


Types of Probability
Determining objective probability
 Relative frequency
 Typically based on historical data

Number of occurrences of the event


P (event) =
Total number of trials or outcomes

 Classical or logical method


 Logically determine probabilities without
trials
1 Number of ways of getting a head
P (head) =
2 Number of possible outcomes (head or tail)

© 2008 Prentice-Hall, Inc. 2 – 10


Types of Probability

Subjective probability is based on


the experience and judgment of the
person making the estimate
 Opinion polls
 Judgment of experts
 Delphi method
 Other methods

© 2008 Prentice-Hall, Inc. 2 – 11


Mutually Exclusive Events

Events are said to be mutually


exclusive if only one of the events can
occur on any one trial

 Tossing a coin will result


in either a head or a tail
 Rolling a die will result in
only one of six possible
outcomes

© 2008 Prentice-Hall, Inc. 2 – 12


Collectively Exhaustive Events

Events are said to be collectively


exhaustive if the list of outcomes
includes every possible outcome
 Both heads and
OUTCOME
tails as possible OF ROLL
PROBABILITY
outcomes of 1 1/
6

coin flips 2 1/
6
3 1/
 All six possible 1/
6
4 6
outcomes 5 1/
6
of the roll 6 1/
6
of a die Total 1

© 2008 Prentice-Hall, Inc. 2 – 13


Drawing a Card

Draw one card from a deck of 52 playing cards

P (drawing a 7) = 4/52 = 1/13


P (drawing a heart) = 13/52 = 1/4

 These two events are not mutually exclusive


since a 7 of hearts can be drawn
 These two events are not collectively
exhaustive since there are other cards in the
deck besides 7s and hearts

© 2008 Prentice-Hall, Inc. 2 – 14


Table of Differences

MUTUALLY COLLECTIVELY
DRAWS
EXCLUSIVE EXHAUSTIVE
1. Draws a spade and a club Yes No
2. Draw a face card and a Yes Yes
number card
3. Draw an ace and a 3 Yes No
4. Draw a club and a nonclub Yes Yes
5. Draw a 5 and a diamond No No
6. Draw a red card and a No No
diamond

© 2008 Prentice-Hall, Inc. 2 – 15


Adding Mutually Exclusive Events

We often want to know whether one or a


second event will occur
 When two events are mutually
exclusive, the law of addition is –

P (event A or event B) = P (event A) + P (event B)

P (spade or club) = P (spade) + P (club)


= 13/52 + 13/52
= 26/52 = 1/2 = 0.50 = 50%

© 2008 Prentice-Hall, Inc. 2 – 16


Adding Not Mutually Exclusive Events

The equation must be modified to account


for double counting
 The probability is reduced by
subtracting the chance of both events
occurring together
P (event A or event B) = P (event A) + P (event B)
– P (event A and event B both occurring)
P (A or B) = P (A) + P (B) – P (A and B)
P(five or diamond) = P(five) + P(diamond) – P(five and diamond)
= 4/52 + 13/52 – 1/52
= 16/52 = 4/13

© 2008 Prentice-Hall, Inc. 2 – 17


Venn Diagrams

P (A and B)

P (A) P (B) P (A) P (B)

Events that are mutually Events that are not


exclusive mutually exclusive

P (A or B) = P (A) + P (B) P (A or B) = P (A) + P (B)


– P (A and B)
Figure 2.1 Figure 2.2
© 2008 Prentice-Hall, Inc. 2 – 18
Statistically Independent Events

Events may be either independent or


dependent
 For independent events, the occurrence
of one event has no effect on the
probability of occurrence of the second
event

© 2008 Prentice-Hall, Inc. 2 – 19


Which Sets of Events Are Independent?

1. (a) Your education


Dependent events
(b) Your income level

2. (a) Draw a jack of hearts from a full 52-card deck Independent


(b) Draw a jack of clubs from a full 52-card deck events

3. (a) Chicago Cubs win the National League pennant Dependent


(b) Chicago Cubs win the World Series events

4. (a) Snow in Santiago, Chile


Independent events
(b) Rain in Tel Aviv, Israel

© 2008 Prentice-Hall, Inc. 2 – 20


Three Types of Probabilities
 Marginal (or simple) probability is just the
probability of an event occurring
P (A)
 Joint probability is the probability of two or more
events occurring and is the product of their
marginal probabilities for independent events
P (AB) = P (A) x P (B)
 Conditional probability is the probability of event
B given that event A has occurred
P (B | A) = P (B)
 Or the probability of event A given that event B
has occurred
P (A | B) = P (A)
© 2008 Prentice-Hall, Inc. 2 – 21
Joint Probability Example

The probability of tossing a 6 on the first


roll of the die and a 2 on the second roll
P (6 on first and 2 on second)
= P (tossing a 6) x P (tossing a 2)
= 1/6 x 1/6 = 1/36 = 0.028

© 2008 Prentice-Hall, Inc. 2 – 22


Independent Events
A bucket contains 3 black balls and 7 green balls
 We draw a ball from the bucket, replace it, and
draw a second ball

1. A black ball drawn on first draw


P (B) = 0.30 (a marginal probability)
2. Two green balls drawn
P (GG) = P (G) x P (G) = 0.7 x 0.7 = 0.49
(a joint probability for two independent events)

© 2008 Prentice-Hall, Inc. 2 – 23


Independent Events
A bucket contains 3 black balls and 7 green balls
 We draw a ball from the bucket, replace it, and
draw a second ball
3. A black ball drawn on second draw if the first
draw is green
P (B | G) = P (B) = 0.30
(a conditional probability but equal to the marginal
because the two draws are independent events)
4. A green ball is drawn on the second if the first
draw was green
P (G | G) = P (G) = 0.70
(a conditional probability as in event 3)

© 2008 Prentice-Hall, Inc. 2 – 24


Statistically Dependent Events
The marginal probability of an event occurring is
computed the same
P (A)

Calculating conditional probabilities is slightly more


complicated. The probability of event A given that
event B has occurred is
P (AB)
P (A | B) =
P (B)

The formula for the joint probability of two events is


P (AB) = P (B | A) P (A)

© 2008 Prentice-Hall, Inc. 2 – 25


When Events Are Dependent

Assume that we have an urn containing 10 balls of


the following descriptions
 4 are white (W) and lettered (L)
 2 are white (W) and numbered (N)
 3 are yellow (Y) and lettered (L)
 1 is yellow (Y) and numbered (N)

P (WL) = 4/10 = 0.4 P (YL) = 3/10 = 0.3


P (WN) = 2/10 = 0.2 P (YN) = 1/10 = 0.1
P (W) = 6/10 = 0.6 P (L) = 7/10 = 0.7
P (Y) = 4/10 = 0.4 P (N) = 3/10 = 0.3

© 2008 Prentice-Hall, Inc. 2 – 26


When Events Are Dependent

4 balls
White (W) 4
and Probability (WL) =
10
Lettered (L)

2 balls
Urn contains White (W) 2
10 balls Probability (WN) =
and 10
Numbered (N)
3 balls
Yellow (Y) 3
Probability (YL) =
and 10
Lettered (L)
1 ball Yellow (Y) 1
Figure 2.3 and Numbered (N) Probability (YN) =
10
© 2008 Prentice-Hall, Inc. 2 – 27
When Events Are Dependent

The conditional probability that the ball drawn


is lettered, given that it is yellow, is

P (YL) 0.3
P (L | Y) = = = 0.75
P (Y) 0.4

Verify P (YL) using the joint probability formula

P (YL) = P (L | Y) x P (Y) = (0.75)(0.4) = 0.3

© 2008 Prentice-Hall, Inc. 2 – 28


Joint Probabilities
for Dependent Events
If the stock market reaches 12,500 point by January,
there is a 70% probability that Tubeless Electronics
will go up
 There is a 40% chance the stock market will
reach 12,500
 Let M represent the event of the stock market
reaching 12,500 and let T be the event that
Tubeless goes up in value

P (MT) = P (T | M) x P (M) = (0.70)(0.40) = 0.28

© 2008 Prentice-Hall, Inc. 2 – 29


Revising Probabilities with
Bayes’ Theorem
Bayes’ theorem is used to incorporate additional
information and help create posterior probabilities

Prior
Probabilities

Bayes’ Posterior
Process Probabilities

New
Information

Figure 2.4
© 2008 Prentice-Hall, Inc. 2 – 30
Posterior Probabilities
A cup contains two dice identical in appearance but
one is fair (unbiased), the other is loaded (biased)
 The probability of rolling a 3 on the fair die is 1/6 or 0.166
 The probability of tossing the same number on the loaded
die is 0.60
 We select one by chance,
toss it, and get a result of a 3
 What is the probability that
the die rolled was fair?
 What is the probability that
the loaded die was rolled?

© 2008 Prentice-Hall, Inc. 2 – 31


Posterior Probabilities
We know the probability of the die being fair or
loaded is
P (fair) = 0.50 P (loaded) = 0.50
And that
P (3 | fair) = 0.166 P (3 | loaded) = 0.60

We compute the probabilities of P (3 and fair)


and P (3 and loaded)

P (3 and fair) = P (3 | fair) x P (fair)


= (0.166)(0.50) = 0.083
P (3 and loaded) = P (3 | loaded) x P (loaded)
= (0.60)(0.50) = 0.300
© 2008 Prentice-Hall, Inc. 2 – 32
Posterior Probabilities
We know the probability
The sumof the die being
of these fair or
probabilities
loaded is gives us the unconditional
probability
P (fair) = 0.50 of tossing
P (loaded) = 0.50a 3
And that
P (3 | fair) = 0.166P (3)P=(30.083 + 0.300
| loaded) = 0.383
= 0.60

We compute the probabilities of P (3 and fair)


and P (3 and loaded)

P (3 and fair) = P (3 | fair) x P (fair)


= (0.166)(0.50) = 0.083
P (3 and loaded) = P (3 | loaded) x P (loaded)
= (0.60)(0.50) = 0.300
© 2008 Prentice-Hall, Inc. 2 – 33
Posterior Probabilities
If a 3 does occur, the probability that the die rolled
was the fair one is
P (fair and 3) 0.083
P (fair | 3) = = = 0.22
P (3) 0.383

The probability that the die was loaded is

P (loaded and 3) 0.300


P (loaded | 3) = = = 0.78
P (3) 0.383

 These are the revised or posterior probabilities for the


next roll of the die
 We use these to revise our prior probability estimates

© 2008 Prentice-Hall, Inc. 2 – 34


Bayes Calculations
Given event B has occurred
STATE OF P (B | STATE PRIOR JOINT POSTERIOR
NATURE OF NATURE) PROBABILITY PROBABILITY PROBABILITY

A P(B | A) x P(A) = P(B and A) P(B and A)/P(B) = P(A|B)


A’ P(B | A’) x P(A’) = P(B and A’) P(B and A’)/P(B) = P(A’|B)
P(B)

Table 2.2
Given a 3 was rolled
STATE OF P (B | STATE PRIOR JOINT POSTERIOR
NATURE OF NATURE) PROBABILITY PROBABILITY PROBABILITY

Fair die 0.166 x 0.5 = 0.083 0.083 / 0.383 = 0.22


Loaded die 0.600 x 0.5 = 0.300 0.300 / 0.383 = 0.78
P(3) = 0.383

Table 2.3

© 2008 Prentice-Hall, Inc. 2 – 35


General Form of Bayes’ Theorem

We can compute revised probabilities more


directly by using

P ( B | A) P ( A)
P( A | B) 
P ( B | A) P ( A)  P ( B | A ) P ( A )

where
A  the complement of the event A;
for example, if A is the event “fair die”,
then A is “loaded die”

© 2008 Prentice-Hall, Inc. 2 – 36


General Form of Bayes’ Theorem

This is basically what we did in the previous example


If we replace A with “fair die”
Replace A with “loaded die
Replace B with “3 rolled”
We get

P ( fair die | 3 rolled)


P (3 | fair ) P ( fair )

P (3 | fair ) P ( fair )  P (3 | loaded) P (loaded)
(0.166 )(0.50 ) 0.083
  0.22
(0.166 )(0.50 )  (0.60 )(0.50 ) 0.383
© 2008 Prentice-Hall, Inc. 2 – 37
Further Probability Revisions
We can obtain additional information by performing
the experiment a second time
 If you can afford it, perform experiments
several times
We roll the die again and again get a 3
P ( fair )  0.50 and P (loaded)  0.50
P (3,3 | fair )  (0.166 )(0.166 )  0.027
P (3,3 | loaded)  (0.6)(0.6)  0.36

© 2008 Prentice-Hall, Inc. 2 – 38


Further Probability Revisions
We can obtain
P (3,3 and fair )  Pinformation
additional (3,3 | fair )  Pby performing
( fair )
the experiment a second  (0time
.027 )(0.5)  0.013
 If you can afford it, perform experiments
P (3,3 and
several loaded)  P (3,3 | loaded)  P (loaded)
times
 (0.36)(0.5)  0.18
We roll the die again and again get a 3
P ( fair )  0.50 and P (loaded)  0.50
P (3,3 | fair )  (0.166 )(0.166 )  0.027
P (3,3 | loaded)  (0.6)(0.6)  0.36

© 2008 Prentice-Hall, Inc. 2 – 39


Further Probability Revisions
We can obtain
P (3,3 and fair )  Pinformation
additional (3,3 | fair )  Pby performing
( fair )
the experiment a second  (0time
.027 )(0.5)  0.013
 If you can afford it, perform experiments
P (3,3 and
several loaded)  P (3,3 | loaded)  P (loaded)
times
 (0.36)(0.5)  0.18
We roll the die again and again get a 3
P(fair )  0.50 and P(loaded
P (3,3 )  0) .500.013
and fair
P((3fair
P ) (0.166)(0.166)  0.
| 3,)3
,3 | fair 027  0.067
P (3,3) 0.193
P(3,3 | loaded)  (0.6)(0.6)  0.36
P (3,3 and loaded) 0.18
P (loaded | 3,3)    0.933
P (3,3) 0.193

© 2008 Prentice-Hall, Inc. 2 – 40


Further Probability Revisions

After the first roll of the die –

probability the die is fair = 0.22


probability the die is loaded = 0.78

After the second roll of the die –

probability the die is fair = 0.067


probability the die is loaded = 0.933

© 2008 Prentice-Hall, Inc. 2 – 41


Random Variables

A random variable assigns a real number


to every possible outcome or event in an
experiment
X = number of refrigerators sold during the day

Discrete random variables can assume only


a finite or limited set of values
Continuous random variables can assume
any one of an infinite set of values

© 2008 Prentice-Hall, Inc. 2 – 42


Random Variables – Numbers
RANGE OF
RANDOM
EXPERIMENT OUTCOME RANDOM
VARIABLES
VARIABLES
Stock 50 Number of Christmas X 0, 1, 2,…, 50
Christmas trees trees sold
Inspect 600 Number of acceptable Y 0, 1, 2,…, 600
items items
Send out 5,000 Number of people Z 0, 1, 2,…, 5,000
sales letters responding to the
letters
Build an Percent of building R 0 ≤ R ≤ 100
apartment completed after 4
building months
Test the lifetime Length of time the S 0 ≤ S ≤ 80,000
of a lightbulb bulb lasts up to 80,000
(minutes) minutes
Table 2.4
© 2008 Prentice-Hall, Inc. 2 – 43
Random Variables – Not Numbers

RANGE OF
RANDOM
EXPERIMENT OUTCOME RANDOM
VARIABLES
VARIABLES
Students Strongly agree (SA) 5 if SA 1, 2, 3, 4, 5
respond to a Agree (A) 4 if A..
questionnaire Neutral (N) X= 3 if N..
Disagree (D) 2 if D..
Strongly disagree (SD) 1 if SD

One machine Defective Y= 0 if defective 0, 1


is inspected Not defective 1 if not defective

Consumers Good 3 if good…. 1, 2, 3


respond to Average Z= 2 if average
how they like Poor 1 if poor…..
a product

Table 2.5
© 2008 Prentice-Hall, Inc. 2 – 44
Probability Distribution of a
Discrete Random Variable
Selecting the right probability distribution
is important
 For discrete random variables a
probability is assigned to each event

Dr. Shannon asked students


to respond to the statement, 5. Strongly agree
4. Agree
“The textbook was well 3. Neutral
written and helped me 2. Disagree
acquire the necessary 1. Strongly disagree
information.”

© 2008 Prentice-Hall, Inc. 2 – 45


Probability Distribution of a
Discrete Random Variable
RANDOM NUMBER PROBABILITY
OUTCOME VARIABLE (X) RESPONDING P (X)
Strongly agree 5 10 0.1 = 10/100
Agree 4 20 0.2 = 20/100
Neutral 3 30 0.3 = 30/100
Disagree 2 30 0.3 = 30/100
Strongly disagree 1 10 0.1 = 10/100
Total 100 1.0 = 100/100
Table 2.6

Distribution follows all three rules


1. Events are mutually exclusive and collectively exhaustive
2. Individual probability values are between 0 and 1
3. Total of all probability values equals 1
© 2008 Prentice-Hall, Inc. 2 – 46
Probability Distribution for
Dr. Shannon’s Class

0.4 –

0.3 –
P (X)

0.2 –

0.1 –

0–
| | | | | |
1 2 3 4 5
Figure 2.5 X
© 2008 Prentice-Hall, Inc. 2 – 47
Probability Distribution for
Dr. Shannon’s Class

0.4 – Central tendency of the


distribution is the mean
or expected value
0.3 –
Amount of variability or
spread is the variance
P (X)

0.2 –

0.1 –

0–
| | | | | |
1 2 3 4 5
Figure 2.5 X
© 2008 Prentice-Hall, Inc. 2 – 48
Expected Value of a Discrete
Probability Distribution
The expected value is a measure of the central
tendency of the distribution and is a weighted
average of the values of the random variable
n
E X    X i P X i 
i 1

 X 1 P  X 1   X 2 P ( X 2 )  ...  X n P ( X n )
where
X i = random variable’s possible values
P ( X i ) = probability of each of the random variable’s
n possible values

i 1
= summation sign indicating we are adding all n
possible values
E ( X ) = expected value or mean of the random sample
© 2008 Prentice-Hall, Inc. 2 – 49
Variance of a
Discrete Probability Distribution
For a discrete probability distribution the
variance can be computed by
n
σ2  Variance   [ X i  E ( X )]2 P ( X i )
i 1

where
X i = random variable’s possible values
E ( X ) = expected value of the random variable
[ X i  E ( X )]= difference between each value of the random
variable and the expected mean
P ( X i ) = probability of each possible value of the
random sample

© 2008 Prentice-Hall, Inc. 2 – 50


Variance of a
Discrete Probability Distribution

For Dr. Shannon’s class


5
variance   [ X i  E ( X )]2 P ( X i )
i 1

variance  (5  2.9)2 (0.1)  ( 4  2.9)2 (0.2) 


(3  2.9)2 (0.3)  (2  2.9)2 (0.3) 
(1 2.9)2 (0.1)
 0.441  0.242  0.003  0.243  0.361
 1.29

© 2008 Prentice-Hall, Inc. 2 – 51


Variance of a
Discrete Probability Distribution

A related measure of dispersion is the


standard deviation

σ  Variance  σ 2

where
= square root
σ = standard deviation

© 2008 Prentice-Hall, Inc. 2 – 52


Variance of a
Discrete Probability Distribution

A related measure of dispersion is the


standard deviation

σ  Variance  σ 2

where
= square
Forroot
the textbook question
σ = standard deviation
σ  Variance
 1.29  1.14

© 2008 Prentice-Hall, Inc. 2 – 53


Probability Distribution of a
Continuous Random Variable
Since random variables can take on an infinite
number of values, the fundamental rules for
continuous random variables must be modified
 The sum of the probability values must still
equal 1
 But the probability of each value of the
random variable must equal 0 or the sum
would be infinitely large
The probability distribution is defined by a
continuous mathematical function called the
probability density function or just the probability
function
 Represented by f (X)

© 2008 Prentice-Hall, Inc. 2 – 54


Probability Distribution of a
Continuous Random Variable
Probability

| | | | | | |
5.06 5.10 5.14 5.18 5.22 5.26 5.30
Weight (grams)

Figure 2.6

© 2008 Prentice-Hall, Inc. 2 – 55


The Binomial Distribution

 Many business experiments can be


characterized by the Bernoulli process
 The Bernoulli process is described by the
binomial probability distribution
1. Each trial has only two possible outcomes
2. The probability stays the same from one trial
to the next
3. The trials are statistically independent
4. The number of trials is a positive integer

© 2008 Prentice-Hall, Inc. 2 – 56


The Binomial Distribution

The binomial distribution is used to find the


probability of a specific number of successes
out of n trials
We need to know
n = number of trials
p = the probability of success on any
single trial

We let
r = number of successes
q = 1 – p = the probability of a failure

© 2008 Prentice-Hall, Inc. 2 – 57


The Binomial Distribution
The binomial formula is

n!
Probability of r successes in n trials  p r q n r
r! ( n  r )!

The symbol ! means factorial, and


n! = n(n – 1)(n – 2)…(1)
For example
4! = (4)(3)(2)(1) = 24
By definition
1! = 1 and 0! = 1

© 2008 Prentice-Hall, Inc. 2 – 58


The Binomial Distribution

NUMBER OF 5!
Probability = (0.5)r(0.5)5 – r
HEADS (r) r!(5 – r)!

0 0.03125 = 5! (0.5)0(0.5)5 – 0
0!(5 – 0)!
1 0.15625 = 5! (0.5)1(0.5)5 – 1
1!(5 – 1)!
2 0.31250 = 5! (0.5)2(0.5)5 – 2
2!(5 – 2)!
3 0.31250 = 5! (0.5)3(0.5)5 – 3
3!(5 – 3)!
4 0.15625 = 5! (0.5)4(0.5)5 – 4
4!(5 – 4)!
5 0.03125 = 5! (0.5)5(0.5)5 – 5
5!(5 – 5)!

Table 2.7
© 2008 Prentice-Hall, Inc. 2 – 59
Solving Problems with the
Binomial Formula

We want to find the probability of 4 heads in 5 tosses

n = 5, r = 4, p = 0.5, and q = 1 – 0.5 = 0.5

Thus
5!
P  ( 4 successes in 5 trials)  0.5 40.5 5  4
4! (5  4 )!
5( 4)(3)(2)(1)
 (0.0625 )(0.5)  0.15625
4(3)(2)(1)(1! )

Or about 16%

© 2008 Prentice-Hall, Inc. 2 – 60


Solving Problems with the
Binomial Formula
0.4 –

0.3 –
Probability P (r)

0.2 –

0.1 –

0 –| | | | | | |
1 2 3 4 5 6
Values of r (number of successes)
Figure 2.7

© 2008 Prentice-Hall, Inc. 2 – 61


Solving Problems with
Binomial Tables
MSA Electronics is experimenting with the
manufacture of a new transistor
 Every hour a sample of 5 transistors is taken
 The probability of one transistor being
defective is 0.15
What is the probability of finding 3, 4, or 5 defective?

So n = 5, p = 0.15, and r = 3, 4, or 5

We could use the formula to solve this problem,


but using the table is easier

© 2008 Prentice-Hall, Inc. 2 – 62


Solving Problems with
Binomial Tables
P
n r 0.05 0.10 0.15
5 0 0.7738 0.5905 0.4437
1 0.2036 0.3281 0.3915
2 0.0214 0.0729 0.1382
3 0.0011 0.0081 0.0244
4 0.0000 0.0005 0.0022
5 0.0000 0.0000 0.0001
Table 2.8 (partial)

We find the three probabilities in the table


for n = 5, p = 0.15, and r = 3, 4, and 5 and
add them together

© 2008 Prentice-Hall, Inc. 2 – 63


Solving Problems with
Binomial Tables
P
P (3 or
n more
r 0.05 )  P (3)  P
defects ( 4)  P (5)
0.10 0.15
5 0 0.7738  0.0244  0.0022  0.0001
0.5905 0.4437 0.0267
1 0.2036 0.3281 0.3915
2 0.0214 0.0729 0.1382
3 0.0011 0.0081 0.0244
4 0.0000 0.0005 0.0022
5 0.0000 0.0000 0.0001
Table 2.8 (partial)

We find the three probabilities in the table


for n = 5, p = 0.15, and r = 3, 4, and 5 and
add them together

© 2008 Prentice-Hall, Inc. 2 – 64


Solving Problems with
Binomial Tables
It is easy to find the expected value (or mean) and
variance of a binomial distribution

Expected value (mean) = np


Variance = np(1 – p)

For the MSA example


Expected value  np  5(0.15 )  0.75
Variance  np(1 p)  5(0.15 )(0.85 )  0.6375

© 2008 Prentice-Hall, Inc. 2 – 65


The Normal Distribution
The normal distribution is the most popular
and useful continuous probability
distribution
 The formula for the probability density
function is rather complex
 ( x   )2
1
f (X )  e 2 2
 2

 The normal distribution is specified


completely when we know the mean, µ,
and the standard deviation, 

© 2008 Prentice-Hall, Inc. 2 – 66


The Normal Distribution

 The normal distribution is symmetrical,


with the midpoint representing the mean
 Shifting the mean does not change the
shape of the distribution
 Values on the X axis are measured in the
number of standard deviations away from
the mean
 As the standard deviation becomes larger,
the curve flattens
 As the standard deviation becomes
smaller, the curve becomes steeper
© 2008 Prentice-Hall, Inc. 2 – 67
The Normal Distribution

| | |

40 µ = 50 60

Smaller µ, same 

| | |

µ = 40 50 60

Larger µ, same 

| | |

40 50 µ = 60
Figure 2.8
© 2008 Prentice-Hall, Inc. 2 – 68
The Normal Distribution

Same µ, smaller 

Same µ, larger 

Figure 2.9
µ
© 2008 Prentice-Hall, Inc. 2 – 69
The Normal Distribution

16% 68% 16%

–1 +1
a µ b

2.3% 95.4% 2.3%

–2 +2
a µ b

0.15% 99.7% 0.15%

–3 +3
a µ b
Figure 2.10
© 2008 Prentice-Hall, Inc. 2 – 70
The Normal Distribution

If IQs in the United States were normally distributed


with µ = 100 and  = 15, then

1. 68% of the population would have IQs


between 85 and 115 points (±1)
2. 95.4% of the people have IQs between 70
and 130 (±2)
3. 99.7% of the population have IQs in the
range from 55 to 145 points (±3)
4. Only 16% of the people have IQs greater
than 115 points (from the first graph, the
area to the right of +1)

© 2008 Prentice-Hall, Inc. 2 – 71


Using the Standard Normal Table
Step 1
Convert the normal distribution into a standard
normal distribution
 A standard normal distribution has a mean
of 0 and a standard deviation of 1
 The new standard random variable is Z

X 
Z

where
X = value of the random variable we want to measure
µ = mean of the distribution
 = standard deviation of the distribution
Z = number of standard deviations from X to the mean, µ
© 2008 Prentice-Hall, Inc. 2 – 72
Using the Standard Normal Table
For example, µ = 100,  = 15, and we want to find
the probability that X is less than 130
X  130  100
Z 
 15
30
  2 std dev µ = 100
15 P(X < 130)  = 15

| | | | | | |
X = IQ
55 70 85 100 115 130 145

X 
| | | | | | | Z
–3 –2 –1 0 1 2 3 
Figure 2.11
© 2008 Prentice-Hall, Inc. 2 – 73
Using the Standard Normal Table
Step 2
Look up the probability from a table of normal
curve areas
 Use Appendix A or Table 2.9 (portion below)
 The column on the left has Z values
 The row at the top has second decimal
places for the Z values
AREA UNDER THE NORMAL CURVE
Z 0.00 0.01 0.02 0.03
1.8 0.96407 0.96485 0.96562 0.96638
P(X < 130)
1.9 0.97128 0.97193 0.97257 0.97320 = (Z < 2.00)
2.0 0.97725 0.97784 0.97831 0.97882
2.1 0.98214 0.98257 0.98300 0.98341 = 97.7%
2.2 0.98610 0.98645 0.98679 0.98713

Table 2.9
© 2008 Prentice-Hall, Inc. 2 – 74
Haynes Construction Company

Haynes builds apartment buildings


 Total construction time follows a normal
distribution
 For triplexes, µ = 100 days and  = 20 days
 Contract calls for completion in 125 days
 Late completion will incur a severe penalty
fee
 What is the probability of completing in 125
days?

© 2008 Prentice-Hall, Inc. 2 – 75


Haynes Construction Company

X  125  100
Z 
 20
25
  1.25
20

µ = 100 days X = 125 days


 = 20 days
Figure 2.12

From Appendix A, for Z = 1.25 the area is 0.89435


 There is about an 89% probability Haynes
will not violate the contract
© 2008 Prentice-Hall, Inc. 2 – 76
Haynes Construction Company

Haynes builds apartment buildings


 Total construction time follows a normal
distribution
 For triplexes, µ = 100 days and  = 20 days
 Completion in 75 days or less will earn a
bonus of $5,000
 What is the probability of getting the
bonus?

© 2008 Prentice-Hall, Inc. 2 – 77


Haynes Construction Company

X  75  100
Z 
 20
 25
  1.25 P(X < 75 days)
20
Area of
Interest

X = 75 days µ = 100 days


Figure 2.12

 But Appendix A has only positive Z values, the


probability we are looking for is in the negative tail

© 2008 Prentice-Hall, Inc. 2 – 78


Haynes Construction Company

X  75  100
Z 
 20
 25
  1.25 P(X > 125 days)
20 Area of
Interest

µ = 100 days X = 125 days

 Because the curve is symmetrical, we can look


at the probability in the positive tail for the
same distance away from the mean
© 2008 Prentice-Hall, Inc. 2 – 79
Haynes Construction Company
 We know the probability
completing in
125 days is 0.89435
 So the probability
completing in more
than 125 days is
1 – 0.89435 = 0.10565

µ = 100 days X = 125 days

© 2008 Prentice-Hall, Inc. 2 – 80


Haynes Construction Company
 The probability
completing in more
than 125 days is
1 – 0.89435 = 0.10565
 Going back to
the left tail of the
distribution

X = 75 days µ = 100 days

 The probability of completing in less than


75 days is 0.10565 or about 11%
© 2008 Prentice-Hall, Inc. 2 – 81
Haynes Construction Company

Haynes builds apartment buildings


 Total construction time follows a normal
distribution
 For triplexes, µ = 100 days and  = 20 days
 What is the probability of completing
between 110 and 125 days?

 We know the probability of completing in 125


days, P(X < 125) = 0.89435
 We have to complete the probability of
completing in 110 days and find the area
between those two events
© 2008 Prentice-Hall, Inc. 2 – 82
Haynes Construction Company

X  110  100
Z   = 20 days
 20
10
  0.5
20

µ = 100 110 125


days days days
Figure 2.14

 From Appendix A, for Z = 0.5 the area is 0.69146


 P(110 < X < 125) = 0.89435 – 0.69146 = 0.20289
or about 20%
© 2008 Prentice-Hall, Inc. 2 – 83
The F Distribution

 A continuous probability distribution


 The F statistic is the ratio of two sample variances
 F distributions have two sets of degrees of
freedom
 Degrees of freedom are based on sample size and
used to calculate the numerator and denominator

df1 = degrees of freedom for the numerator


df2 = degrees of freedom for the denominator

© 2008 Prentice-Hall, Inc. 2 – 84


The F Distribution

Consider the example: df1 = 5


df2 = 6
 = 0.05
From Appendix D, we get
F, df1, df2 = F0.05, 5, 6 = 4.39
This means
P(F > 4.39) = 0.05

There is only a 5% probability that F will exceed 4.39


© 2008 Prentice-Hall, Inc. 2 – 85
The F Distribution

F

Figure 2.15
© 2008 Prentice-Hall, Inc. 2 – 86
The F Distribution

F value for 0.05 probability


with 5 and 6 degrees of
freedom

0.05

F = 4.39

Figure 2.16
© 2008 Prentice-Hall, Inc. 2 – 87
The Exponential Distribution
 The exponential distribution (also called
the negative exponential distribution) is a
continuous distribution often used in
queuing models to describe the time
required to service a customer
f ( X )  e  x
where
X = random variable (service times)
µ = average number of units the service facility can
handle in a specific period of time
e = 2.718 (the base of natural logarithms)

© 2008 Prentice-Hall, Inc. 2 – 88


The Exponential Distribution
1
Expected value   Average service time

1
Variance  f(X)
2

X
Figure 2.17
© 2008 Prentice-Hall, Inc. 2 – 89
The Poisson Distribution
 The Poisson distribution is a discrete
distribution that is often used in queuing
models to describe arrival rates over time
 x e 
P( X ) 
X!
where
P(X) = probability of exactly X arrivals or occurrences
 = average number of arrivals per unit of time
(the mean arrival rate)
e = 2.718, the base of natural logarithms
X = specific value (0, 1, 2, 3, and so on) of the random
variable

© 2008 Prentice-Hall, Inc. 2 – 90


The Poisson Distribution
 The mean and variance of the distribution are
both 
Expected value = 
Variance = 
0.3 –

0.2 –
P(X)

0.1 –

0 –| | | | | | | | | |
0 1 2 3 4 5 6 7 8
Figure 2.18 X
© 2008 Prentice-Hall, Inc. 2 – 91

Das könnte Ihnen auch gefallen