Frequency With Which That Outcome Would Be Obtained If The Process Were
Frequency With Which That Outcome Would Be Obtained If The Process Were
Despite the fact that the concept of probability is such a common and natural
part of our experience, no single scientific interpretation of the term probability is
accepted by all statisticians.
The frequency interpretation applies only to a problem in which there can be, at
least in principle, a large number of similar repetitions of a certain process. Many
important problems are not of this type. For example, the probability that a
particular medical research project will lead to the development of a new
treatment for a certain disease within a specified period of time.
The interesting feature of an experiment is that each of its possible outcomes can
be specified before the experiment is performed, and probabilities can be
assigned to various combinations of outcomes that are of interest.
3 Set Theory
Let S denote the sample space of some experiment. Then each possible outcome
s of the experiment is said to be a member of the space S, or to belong to the
space S. The statement that s is a member of S is denoted symbolically by the
relation s 2 S.
An event C is a certain subset of possible outcomes in the space S. We say the
event C occurs if the outcome of the experiment is in the set C.
Example 3.1 A six-sided die is rolled. The sample space is S = {1, 2, 3, 4, 5, 6}. Let A =
{2, 4, 6} denote the event A that an even number is obtained, and B = {4, 5, 6} denote
the event that a number greater than 3 is obtained. We say A occurs if we get
the number 2 in a roll.
A = B , A B and B A (equality)
A subset of S that contains no outcomes is called the empty set, or null set, and
it is denoted by the symbol ;. For each event, it is true that ; A S .
3.3 Operations of set theory
Given any two events (or sets) A a nd B, we have the following elementary set
operations:
Union: The union of A a nd B, written A [ B , is the set of elements that belong to
either A or B or both:
A [ B = {x : x 2 A or x 2 B}
Intersection: The intersection of A and B, written A \ B, is the set of elements that
belong to both A and B:
A \ B = {x : x 2 A and x 2 B}
Complements: The complement of A, written A c , is the set of all elements that are not
in A:
Ac = {x : x A}
Disjoint events
It is said that two events A and B are disjoint, or mutually exclusive, if A and B
have no outcomes in common. It follows that A and B are disjoint if and only if
A\B=;.
The events A1, . . . ,An or the events A1,A2, . . . are disjoint if for every i 6= j, we
have that Ai and Aj are disjoint, that is, Ai \ Aj = ; for all i 6= j.
Example 3.2 (Event operations) Consider the experiment of selecting a card at
random from a standard deck and noting its suit: clubs (C), diamond (D), hearts
(H), or spades (S). The sample space is
S = {C,D,H, S},
and some possible events are
Axiom 1 For every event A, P(A) 0. In words, the probability of every event must
be nonnegative.
Axiom 2 P(S) = 1. If an event is certain to occur, then the probability of that
event is 1.
pi 0 , for i = 1, . . . , n
and
Example 5.1 Consider an experiment in which five fibers having different lengths
are subject to a testing process to learn which fiber will break first. Suppose that
the lengths of the five fibers are 1,2,3,4, and 5 inches, respectively. Suppose also
that the probability that any given fiber will be the first to break is proportional to
the length of that fiber.
Example 5.2 (Tossing coins) Suppose that three fair coins are tossed
simultaneously. We shall determine the probability of obtaining exactly two
heads.
6 Counting Methods
For example, suppose that there are three different routes from city A to city B
and five different routes from city B t o city C. Then the number of different
routes from A to C that pass through B is 3 5 = 15.
6.2 Permutations
Suppose that k c ards are to be selected one at a time and removed from a deck
of n c ards (k = 1, 2, . . . , n). Then each possible outcome of this experiment will
be a permutation of k cards from the deck, and the total number of these
permutations will be Pn,k = n(n 1) (n k + 1). This number Pn,k is called the
number of permutations of n e lements taken k a t a time.
Example 6.1 (Arranging books) Suppose that six different books are to be
arranged on a shelf. The number of possible permutations of the books is 6! =
720.
P25,2 = 25 24 = 600.
7 Combinatorial Methods
7.1 Combinations
The number Cn,k is also denoted by the symbol . When this notation is used,
this number is called a binomial coefficient because it appears in the binomial
theorem.
Binomial theorem. For all numbers x and y and each positive integer n,
Examples
Conditional Probability
Example 1.1 Rolling Dice Suppose that two dices were rolled and it was observed
that the sum T of the two numbers was odd. We shall determine the probability
that T was less than 8.
1.1 The multiplication rule for conditional probability
P(AB) = P(B)P(A | B)
Or
Example 1.3 Suppose that two balls are to be selected at random, without
replacement, from a box containing r red balls and b blue balls. The probability
that the first ball will be red and the second ball will be blue is
2 Independent Events
Suppose that P(A) > 0 and P(B) > 0, then it follows from the definitions of
independence and conditional probability that A and B are independent if and
only if:
P(A | B) = P(A) and
P(B | A) = P(B),
that is, learning that B (A) occurred does not change the probability of A (B).
Example 2.1 Machine operations. Suppose that two machines 1 and 2 in a factory
are operated independently of each other. Let A be the event that machine 1 will
become inoperative during a given 8-hour period; let B be the event that
machine 2 will become inoperative during the same period; and suppose that
P(A) = 1/3 and P(B) = 1/4. What is the probability that both machines will become
inoperative during the period.
Theorem 2.1 If two events A and B are independent, then the events A and B c
Definition 2.2 The k events A1, . . . , Ak are independent if, for every subset Ai1 , . . . , Aij
of j of these events (j = 2, 3, . . . , k), P(Ai1 \ \ Aij ) = P(Ai1 ) P(Aij ).
Suppose that a machine produces a defective item with probability p (0 < p < 1)
and produces a nondefective item with probability q = 1 - p. Suppose further that
six items produced by the machine are selected at random and inspected, and
that the outcomes for these six items are independent. The probability that at
least one of the six items will be defective is
p = 1 - P(D1D2D3D4D5D6) = 1 - q 6 ,
The probability that exactly two of six items are defective is:
Example 2.5 Suppose that the probability p that the machine produces a
defective item takes one of two values, either 0.01 or 0.4, the first corresponding
to normal operation and the sec-ond corresponding to a need for maintenance.
Let B1 be the event that p = 0.01, and let B2 be the event that p = 0.4. If we knew that
B1 had occurred, then we would proceed under the assumption that the events
D1, D2, . . . were independent with P(Di | B1) = 0.01 for all i. Let A be the event that we
observe exactly two defectives in a random sample of six items. Then
Definition 1.1 Consider an experiment for which the sample space is defined by
S. A random variable is a function from a sample space into the real numbers. In
other words, in a particular experiment a random variable X would be some
function that assigns a real number X(s) to each possible outcome s 2 S.
Example 1.1 Consider an experiment in which a coin is tossed 10 times. In this
experiment the sample space can be regarded as the set of outcomes consisting
of the 210 different sequences of 10 heads and tails that are possible, and the
random variable X could be the number of heads obtained on the 10 tosses. For
each possible sequence s consisting of 10 heads and tails, this random variable
would assign a number X(s) equal to the number of heads in the sequence. Thus,
if s is the sequence HHTTTHTTTH, then X(s) = 4.
for x = 0, 1, . . . , 10.
Random variables that can take every value in an interval are called continuous
random variables.
Suppose that the value of a random variable X is equally likely to be each of the
k integers 1, 2, . . . , k. Then the p.d.f. of X is as follows:
The function f is called the probability density function of X. Every p.d.f. f must
satisfy the following two requirements:
Let a and b be two given real numbers such that a < b, and consider an
experiment in which a point X is selected from the interval S = {x : a x b} in such
a way that the probability that X will belong to each subinterval of S is
proportional to the length of that subinterval. This distribution of the random
variable X is called the uniform distribution on the interval [a, b].
For example, if a random variable X has a uniform distribution on the interval [1,
4], then the pdf of X is
Example 2.2
Suppose that the pdf of a certain random variable X has the following form
3 The Distribution Function
Property 3.1 The function F(x) is nondecreasing as x increases; that is, if x1 < x2,
then F(x1) F(x2).
Example 1.1 Suppose that a random variable X can have only the four different
values -2, 0, 1, and 4, and that P(X = 2) = 0.1, P(X = 0) = 0.4, P(X = 1) = 0.3,
and P(X = 4) = 0.2. Then
E(X) = 2(0.1) + 0(0.4) + 1(0.3) + 4(0.2) = 0.9.
Expectation for a continuous distribution
If a random variable X has a continuous distribution for which the pdf is f, then
the expectation E(X) is defined as follows:
The number E(X) is also called the expected value of X or the mean of X; and the
terms expectation, expected value, and mean can be used interchangeably.
3 Variance
5 The Median
Example:. The set of voltages generated by thermal electron motion in a large number of identical resistors.
Ergodicity
A Random Process is Ergodic if all time averages of sample functions equal the corresponding ensemble averages.