Probability
Probability
Outline of Lecture
Random Variable
– Discrete Random Variable.
– Continuous Random Variables.
Probability Distribution Function.
– Discrete and Continuous.
– PDF and PMF.
Expectation of Random Variables.
Propagation through Linear and Nonlinear model.
Multivariate Probability Density Functions.
Some Important Probability Distribution
Functions.
2
Random Variables
3
Discrete Random Variables
A random variable X and the corresponding
distribution are said to be discrete, if the number of
values for which X has non-zero probability is finite.
Probability Mass Function of X:
pj when x x j
f ( x)
0 otherwise
0 x0
(1 p) 2 0 x 1
F ( x)
p(1 p)1 x 2
p 2
x2
X denote the random variable that is equal to sum of two fair
dices.
– Random variable can take any integral value between 1 and 12.
5
Continuous Random Variables and Distributions
F '( x) f ( x)
The integrand f(y) is called a probability density
function.
f ( x)dx 1
Properties: b
P(a X b) F (b) F (a) f ( x)dx
a
6
Continuous Random Variables and Distributions
7
Difference Between PDF and PMF
Probability density function does not defines a probability but
probability density.
– To obtain the probability we must integrate it in an interval.
Probability mass function gives the true probability.
– It does not need to be integrate to obtain the probability.
a b
Probability distribution function is either continuous or has a jump
discontinuity.
1) P(a X b) 3) P(a X b)
2) P(a X b) 4) P(a X b)
x1 x2 xn y1 y2 yn
x y
n n
Recall, probability approx. equal to relative frequency.
– Approx. Np1 number of xi’s have value u1
x1 x2 xn np1u1 npmum
x ui pi
n n
9
Statistical Characterization of Random Variables
Expected Value:
–The expected value of a discrete random variable, x
is found by multiplying each value of random
variable by its probability and then summing over all
values of x.
Expected value of x: E[ x] xP( x) xf ( x)
x x
– Expected value is equivalent to center of mass concept.
r mi ri mi
– That’s why name first moment also.
– Body is perfectly balanced abt. Center of mass
The expectation value of x is the “balancing point” for the
probability mass function of x
– Expected value is equal to the point of symmetry in case of
symmetric pmf/pdf.
10
Statistical Characterization of Random Variables
Law of Unconscious Statistician (LOTUS): We can
take an expectation of any function of a random
variable.
Expected value of (y=g(x)) = E[y]= yf y g(x)f x
y x
11
Example
13
Expectation Rules
14
Variance of Random Variable
V ( x) E[ x 2 2 x 2 ]
E[ x 2 ] 2( E[ x]) 2 ( E[ x]) 2
E[ x 2 ] ( E[ x]) 2
15
Propagation of moments and density function through
linear models
y=ax+b
(x )
Let us define z
Here, a = 1/ and b = - /
Therefore, E[z] = 0 and V[z] = 1
z is generally known as “Standardized variable”
16
Propagation of moments and density function through
non-linear models
If x is a random variable with probability density function p(x) and y =
f(x) is a one to one transformation that is differentiable for all x then the
probability function of y is given by
E ( y ) a x2 and V ( y ) 2a 4 x4
17
Random Variables
n F ( X )
f ( x)
X 1X 2 ...X n
Marginal Probability Functions: A marginal probability
functions are obtained by integrating out the
variables that are of no interest.
y
F ( x) P( x, y ) or f ( x, y )dy
y y
19
Multivariate Expectations
E( X ) xf
X ( x)dx
What abt. g(X,Y)=X+Y
f X ( x)
f X ,Y ( x, y )dy
E( X ) xf
X ( x)dx xf
X ,Y ( x, y )dxdy
E (Y )
yf Y ( y )dy yf
X ,Y ( x, y )dydx
E ( g ( X )) g ( x) f
X ( x)dx g ( x) f
X ,Y ( x, y )dydx
E (h(Y )) h( y ) f
X ,Y ( x, y )dxdy
E ( g ( X , Y )) g ( x, y ) f
X ,Y ( x, y )dxdy
20
Multivariate Expectations
Mean Vector:
E[x] [ E[ x1 ] E[ x2 ] ...... E[ xn ]]
Expected value of g(x1,x2,…….,xn) is given by
E[ g (x)] ..... g (x) f (x) or ..... g (x) f (x)dx
xn xn1 x1 xn xn-1 x1
Covariance Matrix:
cov[x] P E[(x )(x )T ] E[xxT ] T
where, S E[xxT ] is known as autocorrelation matrix.
1 0 0 1 12 1n 1 0 0
0 0 1 2 n 0 2 0
NOTE: P R 2 21
0 0 n n1 n 2 1 0 0 n
22
Independent Variables
23
Gaussian or Normal Distribution
The normal distribution is the most widely known and used distribution
in the field of statistics.
– Many natural phenomena can be approximated by Normal
distribution.
Central Limit Theorem:
– The central limit theorem states that given a distribution with a mean
and variance 2, the sampling distribution of the mean approaches
a normal distribution with a mean and a variance 2/N as N, the
sample size increases.
0.399
Normal Density Function:
( x )2
1
f ( x) e 2 2
x
2
n
2 R 2
25
Multivariate Normal Distribution
26
Summary of Probability Distribution Functions
Standardized Symmetric 1 x2
2
Exponential Skewed e T
0 1/ 1/2
Positively
A distribution is skewed if it has most of its values either to the right or to the left of its mean
27
Properties of Estimators
Unbiasedness
– On average the value of parameter being estimated is equal
to true value.
E[ xˆ ] x
Efficiency
– Have a relatively small variance.
– The values of parameters being estimated should not vary
with samples.
Sufficiency
– Use as much as possible information available from the
samples.
Consistency
– As the sample size increases, the estimated value
approaches the true value.
28