Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
0% found this document useful (0 votes)
8 views

Machine Learning

Uploaded by

khitas mehdi
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

Machine Learning

Uploaded by

khitas mehdi
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 20

Classification, Machine learning

et Deep learning
Traditional Programming

Data
Computer Output
Program

Machine Learning

Data
Computer Program
Output
Clustering Strategies
• K-means
– Iteratively re-assign points to the nearest cluster
center.
Sample Applications
• Face recognition
• Character recognition
• Speech recognition
• Medical diagnosis
• Industrial applications
• Web search
• Space exploration
• Robotics
• Information extraction
• Social networks
Face Recognition

Training examples of a person

Test images
The machine learning
framework
• Apply a prediction function to a feature representation of
the image to get the desired output:

f( ) = “apple”
f( ) = “tomato”
f( ) = “cow”
The machine learning
framework
y = f(x)
output prediction Image
function feature

• Training: given a training set of labeled examples {(x1,y1),


…, (xN,yN)}, estimate the prediction function f by minimizing
the prediction error on the training set
• Testing: apply f to a never before seen test example x and
output the predicted value y = f(x)
Steps
Training Training
Labels
Training
Images
Image Learned
Training
Features model

Testing

Image Learned
Prediction
Features model
Test Image
Features
• Raw pixels

• Histograms

• Other descriptors
Many classifiers to choose from

• K-nearest neighbor
• Neural networks
• SVM
• Deep Neural networks
• Etc.
Classifiers: Nearest neighbor

Training
Training Test
examples
examples example
from class 2
from class 1

f(x) = label of the training example nearest to x

• All we need is a distance function for our inputs


• No training required!
(Artificial) Neural Networks
• Motivation: human brain
– massively parallel (1011
neurons, ~20 types)
– small computational units with
simple low-bandwidth
communication (1014 synapses,
1-10ms cycle time)

• Realization: neural network


– units (≈ neurons) connected by
directed weighted links
– activation function from inputs
to output
Neural Networks (continued)

• neural network = parameterized family of nonlinear functions


• types
– feed-forward (acyclic): single-layer perceptrons, multi-layer networks
– recurrent (cyclic): Hopfield networks, Boltzmann machines
Neural Network Learning
Key Idea: Adjusting the weights changes the function
represented by the neural network (learning = optimization in
weight space).

Iteratively adjust weights to reduce error (difference between


network output and target output).

• Weight Update
– backpropagation
Deep Learning
• Deep learning (DL) is a subtype of machine learning (ML). DL can process a wider
range of data resources, requires less data preprocessing by humans (e.g. feature
labelling), and can sometimes produce more accurate results than traditional ML
approaches (although it requires a larger amount of data to do so).
• However, it is computationally more expensive in time to execute, hardware costs
and data quantities.
Deep Learning
Deep Learning

You might also like