Statistical Machine Learning-The Basic Approach and Current Research Challenges
Statistical Machine Learning-The Basic Approach and Current Research Challenges
Herbert Simon
research.
Detection of fraudulent activity
(credit card transactions, intrusion
detection, stock market manipulation)
Analysis of genome functionality
Email spam detection.
Spatial prediction of landslide hazards.
Classification prediction
we wish to classify data points into categories, and we
are given already classified samples as our training
input.
For example:
Training a spam filter
Medical Diagnosis (Patient info
High/Low risk).
Stock market prediction ( Predict tomorrows market
trend from companies performance data)
Clustering
the grouping data into representative collections
- a fundamental tool for data analysis.
Examples :
Learning Theory:
The fundamental dilemma
Tradeoff between
accuracy and simplicity
y=f(x)
Y
Good models
should enable
Prediction
of new data
X
Accuracy
Limited data
Possible
Models/representati
ons
Complexity
Problem Outline
We are interested in
(automated) Hypothesis Generation,
rather than traditional Hypothesis Testing
1
VC dim( H ) ln( )
| {(x, y) S : h( x) y} |
Pr( x , y )D (h( x) y )
c
|S|
|S|
Training error
Complexity Term
Hardness-of-Approximation Results
For each of the following classes, approximating the
best agreement rate for h in H (on a given input
sample S ) up to some constant ratio, is NP-hard :
Monomials
Constant width
Monotone Monomials
Half-spaces
Balls
BD-Eiron-Long
Axis aligned Rectangles
Threshold NNs
Bartlett- BD
The Class H
Approximation Error
Estimation Error
Computational Error
Total Error
The predictor h:
Sign ( wi xi+b)
Potential problem
data may not be linearly separable
x (x, x2)
Kernel-Based Algorithms
Rather than define the embedding explicitly, define
just the matrix of the inner products in the range
space.
K(x1x1)
K(x1xm)
K(x1x2)
K(xixj)
K(xmx1)
............
.......
.......
........
K(xmxm
)
max
separating h
min
xi
wn xi