Engineering Data Analysis Chapter 3 - Discrete Probability Distribution
Engineering Data Analysis Chapter 3 - Discrete Probability Distribution
Introduction
Many physical systems can be modelled by a similar or the same random variables
and random experiments. The distribution of the random variables involved in each of
these common systems can be analyzed, and the result of that analysis can be used in
different applications and examples. In this chapter, the analysis of several random
experiments and discrete random variables that often appear in applications is discussed.
A discussion of the basic sample space of the random experiment is frequently omitted
discrete random variable. A discrete random variable is a random variable that has
With a discrete probability distribution, each possible value of the discrete random
Video: https://www.youtube.com/watch?v=mrCxwEZ_22o
3.1 Random Variables and Their Probability Distributions
Random Variables
other mathematical variables, a random variable conceptually does not have a single,
fixed value (even if unknown); rather, it can take on a set of possible different values,
imprecise measurements). They may also conceptually represent either the results of an
“objectively” random process (such as rolling a die), or the “subjective” randomness that
either discrete (that is, taking any of a specified list of exact values) or as continuous
function describing the possible values of a random variable and their associated
Discrete random variables can take on either a finite or at most a countably infinite
set of discrete values (for example, the integers). Their probability distribution is given by
a probability mass function which directly maps each value of the random variable to a
probability. For example, the value of x1 takes on the probability p1, the value of x2 takes
on the probability p2, and so on. The probabilities pi must satisfy two requirements: every
probability pi is a number between 0 and 1, and the sum of all the probabilities is 1.
(p1+p2+⋯+pk=1)
This shows the probability mass function of a discrete probability distribution. The
probabilities of the singletons {1}, {3}, and {7} are respectively 0.2, 0.5, 0.3. A set not
Examples of discrete random variables include the values obtained from rolling a
of possible values. The probability distribution of a discrete random variable x lists the
values and their probabilities, where value x1 has probability p1, value x2 has
probability x2, and so on. Every probability pi is a number between 0 and 1, and the sum
• The number of eggs that a hen lays in a given day (it can’t be 2.3)
graph. For example, suppose that xx is a random variable that represents the number of
people waiting at the line at a fast-food restaurant and it happens to only take the values
2, 3, or 5 with probabilities 2/10, 3/10, and 5/10 respectively. This can be expressed
through the function f(x) = x/10, x=2, 3, 5 or through the table below. Of the conditional
probabilities of the event BB given that A1 is the case or that A2 is the case, respectively.
Notice that these two representations are equivalent, and that this can be represented
Video: https://www.youtube.com/watch?v=cqK3uRoPtk0
Probability Histogram: This histogram displays the probabilities of each of the three
discrete random variables. The formula, table, and probability histogram satisfy the
1. 0≤f(x) ≤1, i.e., the values of f(x) are probabilities, hence between 0 and 1.
2. ∑f(x) =1, i.e., adding the probabilities of all disjoint cases, we obtain the probability
histogram, and displays specific probabilities for each discrete random variable.
This shows the graph of a probability mass function. All the values of this function
x f(x)
2 0.2
3 0.3
5 0.5
Discrete Probability Distribution: This table shows the values of the discrete random
are defective. If a school makes a random purchase of 2 of these computers, find the
Solution:
Let X be a random variable whose values x are the possible numbers of defective
computers purchased by the school. Then x can only take the numbers 0, 1, and 2.
Now,
x 0 1 2
You might recall that the cumulative distribution function is defined for discrete
Again, F(x) accumulates all of the probability less than or equal to x. The cumulative
that of the discrete case. All we need to do is replace the summation with an integral.
defined as:
𝑥
𝐹(𝑥) = ∫ 𝑓(𝑡)𝑑𝑡
−∞
For -∞<x<∞
Example 1. Suppose that a day’s production of 850 manufactured parts contains 50 parts
that do not con- form to customer requirements. Two parts are selected at random,
without replacement, from the batch. Let the random variable X equal the number of
Solution:
The question can be answered by first finding the probability mass function of X.
Therefore,
The cumulative distribution function for this example is graphed in the figure below. Note
that F(x) is defined for all x from - < x < and not only for 0, 1, and 2.
The expected value of a random variable is the weighted average of all possible
probability distribution of a discrete random variable X lists the values and their
probabilities, such that xi has a probability of pi. The probabilities pi must satisfy two
requirements:
expectation, EV, mean, or first moment) of a random variable is the weighted average of
all possible values that this random variable can take on. The weights used in computing
The expected value may be intuitively understood by the law of large numbers: the
expected value, when it exists, is almost surely the limit of the sample mean as sample
size grows to infinity. More informally, it can be interpreted as the long-run average of the
results of many independent repetitions of an experiment (e.g. a dice roll). The value may
not be expected in the ordinary sense—the “expected value” itself may be unlikely or even
impossible (such as having 2.5 children), as is also the case with the sample mean.
How To Calculate Expected Value
Suppose random variable X can take value x1 with probability p1, value x2 with
probability p2, and so on, up to value xi with probability pi. Then the expectation value of
a random variable XX is defined as E[X] = x1 p1+ x2 p2+⋯+xi pi, which can also be written
If all outcomes xi are equally likely (that is, p 1= p2 =⋯=pi), then the weighted average
turns into the simple average. This is intuitive: the expected value of a random variable is
the average of all values it can take; thus, the expected value is what one expects to
happen on average. If the outcomes xi are not equally probable, then the simple average
must be replaced with the weighted average, which takes into account the fact that some
outcomes are more likely than the others. The intuition, however, remains the same: the
For example, let X represent the outcome of a roll of a six-sided die. The possible values
for X are 1, 2, 3, 4, 5, and 6, all equally likely (each having the probability of 1/6). The
expectation of X is:
In this case, since all outcomes are equally likely, we could have simply averaged the
numbers together:
(1 + 2 + 3 + 4 + 5 + 6) /6 = 3.5.
Average Dice Value Against Number of Rolls
Binomial Experiment
• Each trial can result in just two possible outcomes. We call one of these outcomes
• The trials are independent; that is, the outcome on one trial does not affect the
Consider the following statistical experiment. You flip a coin 2 times and count the
2. Each trial can result in just two possible outcomes - heads or tails.
4. The trials are independent; that is, getting heads on one trial does not affect
• b (x; n, P): Binomial probability - the probability that an n-trial binomial experiment
trial is P.
Binomial Distribution
a binomial distribution.
Suppose we flip a coin two times and count the number of heads (successes). The
binomial random variable is the number of heads, which can take on values of 0, 1, or 2.
0 0.25
1 0.50
2 0.25
results in exactly x successes. For example, in the above table, we see that the binomial
Given x, n, and P, we can compute the binomial probability based on the binomial
formula.
probability is:
b (x; n, P) = nCx * Px * (1 - P) n - x
or
Example 1.Suppose a die is tossed 5 times. What is the probability of getting exactly 2
fours?
Solution:
This is a binomial experiment in which the number of trials is equal to 5, the number
of successes is equal to 2, and the probability of success on a single trial is 1/6 or about
random variable falls within a specified range (e.g., is greater than or equal to a stated
45 or fewer heads in 100 tosses of a coin. This would be the sum of all these individual
binomial probabilities.
b (x = 0; 100, 0.5) + b (x = 1; 100, 0.5) + ... + b (x = 44; 100, 0.5) + b (x = 45; 100, 0.5)
Example 1. What is the probability of obtaining 45 or fewer heads in 100 tosses of a
coin?
Solution:
formula. The sum of all these probabilities is the answer we seek. Thus,
b (x < 45; 100, 0.5) = b (x = 0; 100, 0.5) + b (x = 1; 100, 0.5) + . . . + b (x = 45; 100, 0.5)
students from the same school apply, what is the probability that at most 2 are
accepted?
Solution:
formula. The sum of all these probabilities is the answer we seek. Thus,
experiment.
failures.
▪ The average number of successes (μ) that occurs in a specified region is known.
▪ The probability that a success will occur is proportional to the size of the region.
▪ The probability that a success will occur in an extremely small region is virtually
zero.
Note that the specified region could take many forms. For instance, it could be a length,
Notation
The following notation is helpful, when we talk about the Poisson distribution.
logarithm system.)
• P (x; μ): The Poisson probability that exactly x successes occur in a Poisson
a Poisson distribution.
Given the mean number of successes (μ) that occur in a specified region, we can
number of successes within a given region is μ. Then, the Poisson probability is:
where x is the actual number of successes that result from the experiment, and e is
Example 1. The average number of homes sold by the Acme Realty company is 2
homes per day. What is the probability that exactly 3 homes will be sold tomorrow?
Solution:
▪ x = 3; since we want to find the likelihood that 3 homes will be sold tomorrow.
▪ e = 2.71828; since e is a constant equal to approximately 2.71828.
P (3; 2) = 0.180
random variable is greater than some specified lower limit and less than some specified
upper limit.
Example. Suppose the average number of lions seen on a 1-day safari is 5. What is the
probability that tourists will see fewer than four lions on the next 1-day safari?
▪ x = 0, 1, 2, or 3; since we want to find the likelihood that tourists will see fewer
than 4 lions; that is, we want the probability that they will see 0, 1, 2, or 3 lions.
To solve this problem, we need to find the probability that tourists will see 0, 1, 2, or 3
lions. Thus, we need to calculate the sum of four probabilities: P (0; 5) + P (1; 5) + P (2;
5) + P (3; 5).
To compute this sum, we use the Poisson formula:
P (x < 3, 5) = [ (e-5) (50) / 0!] + [ (e-5) (51) / 1!] + [(e-5) (52) / 2!] + [(e-5) (53) / 3!]
[(0.006738) (125) / 6]
P (x < 3, 5) = 0.2650
REFERENCES:
Montgomery, D. C. et al. (2003). Applied Statistics and Probability for Engineers 3rd Edition. USA.
John Wiley & Sons, Inc.
Walpole, R. E. et al. (2016). Probability & Statistics for Engineers & Scientists 9th Edition. England.
Pearson Education Limited
https://courses.lumenlearning.com/boundless-statistics/chapter/discrete-random-variables/
https://newonlinecourses.science.psu.edu/stat414/node/98/
https://stattrek.com/probability-distributions/binomial.aspx
https://stattrek.com/probability-distributions/poisson.aspx