Positive Definite Matrix Econometrics
Positive Definite Matrix Econometrics
Convex functions
Rudi Pendavingh
Optimization in Rn , lecture 7
Corollary
Let p ∈ R[X1 , . . . , Xn ] be a homogeneous quadratic polynomial. Then
p(x1 , . . . , xn ) ≥ 0 for all x1 , . . . , xn ∈ R if and only if
p = s12 + · · · + sn2
Lemma
A∼
= B if and only if B = Y t AY for some invertible Y .
Lemma
Let A, B be symmetric. If A ∼
= B, then A is PSD ⇔ B is PSD.
Rudi Pendavingh (TUE) Semidefinite matrices & Convex functions ORN7 3 / 17
Recognizing PSD matrices
Example
1 2 −1 1 0 0 1 0 0
A= 2 5 1 =∼ 0 1 3 ∼
= 0 1 0 = D.
−1 1 12 0 3 11 0 0 2
Lemma
A is PD ⇐⇒ A is PSD and det(A) 6= 0.
Theorem
Let A be a symmetric matrix. The following are equivalent:
1 A is PD, i.e. x t Ax > 0 for all nonzero x ∈ Rn .
2 all eigenvalues of A are positive.
3 A = Z t Z for some real matrix Z so that det(Z ) 6= 0.
Lemma
Let A, B be symmetric. If A ∼
= B, then A is PD ⇔ B is PD.
Let A be a square matric and let I ⊆ {1, . . . , n}. Then AI denotes the
restriction of A to the rows and columns indexed by I .
Lemma
Let A be an n × n matrix. Then A is PSD ⇐⇒ det(AI ) ≥ 0 for all
I ⊆ {1, . . . , n}.
Lemma
Let A be an n × n matrix. Then A is PD ⇐⇒ det(AI ) > 0 for
I = {1}, . . . , {1, . . . , n}.
Proof.
Let x ∈ Cn be a nonzero vector such that λx = Ax.
Let p be such that |xp | = maxi |xi |.
We have λxp = nj=1 apj xj , hence (λ − app )xp = j6=p apj xj .
P P
P P
Taking norms, |λ − app ||xp | = | j6=p apj xj | ≤ j6=p |apj ||xj | ≤ ρp |xp |.
Dividing by |xp |, the Theorem follows.
Definition
A function f : Rn → R is convex if Dom(f ) is convex, and
Definition
f is concave if −f is convex.
Lemma
f is convex and concave if and only if f is affine.
Theorem
Let f : Rn → R be a norm. Then f is convex.
Lemma
Let f : Rn → R be a convex function. Then {x ∈ Rn | f (x) ≤ α} is a
convex set for any α ∈ R.
Lemma
f is a convex function if and only if epi(f ) is a convex set.
Definition
A subgradient of f at x is a row vector w such that
Theorem
f is convex ⇐⇒ Dom(f ) is convex, and there exists a subgradient of f at
each x in the interior of Dom(f ).
Lemma
If f : Rn → R is convex, then the following functions are convex as well:
x 7→ αf (x), for any α > 0
x 7→ f (x + t), for any fixed t ∈ Rn
x 7→ f (Ax), for any n × m matrix A
Lemma
If f1 , . . . , fm : Rn → R are convex, then the following functions are convex
as well:
x 7→ max{f1 (x), . . . , fm (x)}
x 7→ f1 (x) + · · · + fm (x)
Lemma
If w is a subgradient of f at x, and f is differentiable at x, then
w = ∇f (x).
f (y ) ≥ f (x) + ∇f (x)(y − x)
Definition
The Hessian of a function f : Rn → R at x is the matrix
∂2f 2f
∂x12
(x) · · · ∂x∂1 ∂x n
(x)
∇2 f (x) :=
.. ..
.
. .
2
∂ f 2
∂ f
∂xn ∂x1 (x) · · · ∂x 2
(x)
n
Example
x t Qx
Let f : x 7→ 2 + px, where Q is a symmetric matrix, p a row vector.
∇f (x) = xtQ + p for all x
∇2 f (x) = Q for all x
f is convex if and only if Q is PSD
if f is convex, then f (y ) = min{f (x) | x ∈ Rn } if and only if
0 = ∇f (y ) = y t Q + p
X 7→ log(det(X )) is concave on
{X | X is PSD, det(X ) 6= 0}
g : t 7→ − log(det(X + tY ))
= − log(det(X )) − log(det(I + tY 0 ))
where Y 0 = Z −1 YZ −1Q. Let λ1 , . . . , λn be the eigenvalues of Y 0 .
Then det(I + tY 0 ) = i (1 + tλi ), hence
X
g (t) = − log(det(X )) − log(1 + tλi ).
i