Probability: 201A: Econometrics I Peter E. Rossi Anderson School of Management
Probability: 201A: Econometrics I Peter E. Rossi Anderson School of Management
201A: Econometrics I
Peter E. Rossi
Anderson School of Management
2018
1 / 25
Outline
2 / 25
Role of Probability in Econometrics
3 / 25
Sample Spaces and Events
4 / 25
Set Operations
5 / 25
Partitions, Indicator Functions, and Monotone Sequences
A partition of the sample space. A sequence of disjoint sets such that
S
Ai = Ω .
Example:
consider a sequence of N coin tosses, then the number of heads in each
sequence defines a partition of the sample space. Looking ahead: if our
goal is to make inferences regarding the probability of a head H given the
realization of one sequence, does knowing which element of this partition
“summarize” the information in this seqence completely?
An indicator function of an event is defined by
(
1 if ω ∈ A
IA (ω) =
0 if ω ∈/A
6 / 25
Probability
Assign a real number, P (A) , to each event. With some restrictions, this
mapping from events to real numbers can be called a probability measure
or probability distribution.
Axioms of Probability.
Axiom 1: P (A) ≥ 0
Axiom 2: P (Ω) = 1 S∞ P∞
Axiom 3: If A1, A2, . . . are disjoint, then P ( i=1 Ai ) = i=1 P (Ai ).
How can we interpet what “probability” means or where this measure
comes from?
There are three major schools of thought.
I Frequency interpretation: probabilities are long-run frequencies (e.g.
coin toss).
I Subjective: probabilities represent degrees of belief
I Probabilities are useful building blocks for models of observed data
7 / 25
Interpretation of Probability
8 / 25
Properties of Probability Measures
9 / 25
Finite Sample Spaces and Counting Measure
Suppose the sample space is finite (is this much of a restriction?). Classic
examples include the coin toss experiment but this covers anytime in
which the set of possible outcomes are finite. If all outcomes are equally
likely, then you get what is called a counting measure. That is, to
determine the probability of an event, we simply count the number of
outcomes in the event and express that as a ratio to the total number of
possible outcomes.
kAk
P (A) =
kΩk
This probability measure is called the uniform probability measure.
Computing probabilities with this sort of measure involves simply
counting the total number of outcomes in the sample space and counting
the number of outcomes that constitute an event. This can get
somewhat tricky and is called combinatorics. Given we have computers,
these considerations are less relevant today than in the heyday of
probability theory among starving Russian probabilists.
10 / 25
How to Count if You Must
We only need to know two “counting” methods. Most other identities can
be derived using these two ideas.
I The number of ways of ordering n objects is
n! = n (n − 1) . . . 3 × 2 × 1.
I The number of ways of choosing k objects from n is given by the
so-called binomial coefficients. (think we have to choose a
committee of size k from among n students).
n n (n − 1) . . . (n − k + 1) n!
= =
k k! k! (n − k)!
11 / 25
Independent Events
12 / 25
Equally Probable and Disjoint Events
Equally probable events do not have to be independent!
The classic example is a disjoint event. If we toss a coin, then the event
of a Head is equally probable as the event of a tail even though they are
not independent. We know that for any two events (with positive prob),
P(A) , P (B) > 0 and we also know (by definition of disjoint) that
P (AB) = 0.
Example of computations that exploit disjoint events and independence.
Consider the probability that at least one head is obtained in a sequence
of n coin tosses. One way is to simply count all events with one head and
divide by the number of total events. Another is to express the problem
using the complementary event (no heads obtained).
13 / 25
Birthday Problem
14 / 25
Birthday Problem (continued)
Now let’s compute the probability that no one has the same birthday. Let
A be the event that there are two or more duplicate birthdays
15 / 25
R Implementation
16 / 25
R Implementation
1.0
0.8
0.6
prob
0.4
0.2
0.0
0 20 40 60
class_size
17 / 25
Conditional Probability
P (A and B)
P (A|B) =
P (B)
18 / 25
Conditional Probability as a Partition
0 1 2 3
TTT HTT HHT HHH
THT HTH
TTH THH
19 / 25
Conditional Probability and Independence
What this means that if A and B are independent, we don’t change our
views regarding the probability of A occurring if we observe B.
20 / 25
Inverse Probability and Fallacies
21 / 25
Bayes Theorem
P (B|Ai ) P (Ai )
P (Ai |B) = P
j P (B|Aj ) P (Aj )
The denominator comes from the fact that the {Ai } constitute a
partition and is sometimes called the “law of total probability.” Some call
P (Ai ) the prior (before) probability of the event and is P (Ai |B) is called
the posterior (after) probability. This is an updating formula that tells
you how the event B updates probabilities from the prior to posterior.
22 / 25
The Classic Game Show Host Problem
Problem. You are on a game show where you “compete” with the host
for prizes. This problem is styled after the “Price is Right” hosted by
Monty Hall.
I There are three doors (doors A, B, C). There is a prize (a car)
behind one of the doors and a “goat” or nothing behind the others.
I The host asks you to choose one door. Let’s say you pick door A.
I The host reveals what is behind door B and there is no prize behind
it.
I The host ask to if you want to switch from door A to door C.
(sometimes the host might ask you to “bet” or pay for this privilege).
I Should you switch?
23 / 25
The Classic Game Show Host Problem
24 / 25
The Classic Game Show Host Problem
Solution. Let C P be the event that the prize is behind door C. Let B M
be the event that the host (“Monty”) shows you door B. Let ACh be the
event that you choose door A to begin with.
P B M |ACh = P B M |AP P AP + P B M |B P P B P
+ P B M |C P P C P
1
= .5 × 1/3 + 0 × 1/3 + 1 × 1/3 =
M P Ch
P Ch
2
P B |C , A P C |A 1 × 1/3 2
P C P |B M , ACh =
M Ch
= 1 =
P (B |A ) /2 3
25 / 25