Algebraic Methods in Data Science: Lesson 1: Dan Garber
Algebraic Methods in Data Science: Lesson 1: Dan Garber
Algebraic Methods in Data Science: Lesson 1: Dan Garber
Dan Garber
https://dangar.net.technion.ac.il/
Introduction
Course staff:
1 Lecturer: Dan Garber (dangar@technion.ac.il)
2 TA in charge: Ido Botzer (idobotzer@campus.technion.ac.il)
3 TA: Or Markovetzki (ormar@campus.technion.ac.il)
Grade:
1 Homework: - 15% of grade - TAKEF
We will first develop similar notions for the more general and more
abstract linear (vector) spaces and in particular for Rn and Rm×n .
This will in turn lead to the theory of eigenvalues and eigenvectors for real
matrices and (basically) to everything we will do in this course.
A norm is a function that assigns a strictly positive length (or size) to each
vector in a linear space, except for the zero vector, which is assigned a
length of zero. In the following, let X be a linear space.
Definition
A function k·k : X → R is a norm, if
1 ∀x ∈ X kxk ≥ 0, and kxk = 0 if and only if x = 0 (positivity)
2 ∀x, y ∈ X kx + yk ≤ kxk + kyk (triangle inequality)
3 ∀α ∈ R, x ∈ X : kαxk = |α| · kxk (homogeneity)
where (1) follows from the usual triangle inequality for scalars. The case
p = 2: simply the Euclidean-norm (basic geometry).
The case p = ∞:
where again, (1) follows from the triangle inequality for scalars.
Lets get to proving the general case, i.e., for all p ∈ [1, ∞].
@ Dan Garber (Technion) Lesson 1 Winter 2020-21 8 / 25
Proof of triangle inequality for p-norms
First, note that if x = 0 or y = 0 then the proof is trivial: if w.l.o.g.
x = 0 we have kx + ykp = kykp = kykp + 0 = kykp + kxkp .
Consider now the case that kxkp + kykp = 1. It suffices to show that
kx + ykpp ≤ 1 = (kxkp + kykp )p .
Definition: a function g(x) : R → R is convex on interval (a, b) if for any
x, y ∈ (a, b), λ ∈ [0, 1] we have g(λx + (1 − λ)y) ≤ λg(x) + (1 − λ)g(y).
Consider now the case that kxkp + kykp = 1. It suffices to show that
Fact: the scalar function f (x) = |x|p is convex on (−∞, ∞). That is, for
any x, y ∈ R and λ ∈ [0, 1] it holds that
And so we have proved the claim for the case x, y 6= 0, kxkp + kykp = 1.
@ Dan Garber (Technion) Lesson 1 Winter 2020-21 11 / 25
It is not difficult to show (try it for yourself) that it indeed satisfies the
inner product properties:
1 hx, xi ≥ 0, and hx, xi = 0 if and only if x = 0 (positivity)
2 hx + y, zi = hx, zi + hy, zi (additivity)
3 hαx, yi = αhx, yi (homogeneity)
4 hx, yi = hy, xi (symmetry)
Theorem
Let X be an
pinner product space. Then, the function k·k : X → R given
by kxk := hx, xi is a norm.
The fact that ∀x: kxk ≥ 0 and kxk = 0 ⇔ x = 0, follows directly from
the first property of inner products.
To prove the homogeneity, fix some x ∈ X and scalar α ∈ R. We have
p p p
2
kαxk = hαx, αxi = α hx, yi = |α| hx, xi = |α|kxk,
(1)
where (1) follows from homogeneity and symmetry of the inner product.
It remains to prove the triangle inequality.
@ Dan Garber (Technion) Lesson 1 Winter 2020-21 17 / 25
Theorem
Let X be an
pinner product space. Then, the function k·k : X → R given
by kxk := hx, xi is a norm.
We have seen that for the standard inner product in Rn it holds for any
two vectors x, y ∈ Rn with angle θ between them that
> x> y
x y = kxk2 kyk2 cos θ, cos θ =
kxk2 kyk2
Orthogonality
Orthogonality generalizes to notion of two perpendicular lines from plane
geometry to abstract inner product spaces. It will be central to everything
we will do in this course.
Definition
Given an inner product space X and vectors x, y ∈ X , we say that x, y
are orthogonal if hx, yi = 0, and we write x ⊥ y.
@ Dan Garber (Technion) Lesson 1 Winter 2020-21 22 / 25
Orthogonality
Definition
Given an inner product space X and vectors x, y ∈ X , we say that x, y
are orthogonal if hx, yi = 0, and we write x ⊥ y.
Orthogonality
Definition
Given an inner product space X and vectors x(1) , . . . , x(n) in X , all are
non-zero, we say that x(1) , . . . , x(n) are mutually orthogonal if
hx(i) , x(j) i = 0 for all i 6= j.
Theorem
Given an inner product space X , any mutually orthogonal vectors
x(1) , . . . , x(n) are linearly independent.
Theorem
Given an inner product space X , any mutually orthogonal vectors
x(1) , . . . , x(n) are linearly independent.
= αj hx(j) , x(j) i =
6 0 {x(i) , x(j) are orthogonal, αj 6= 0}