This document provides an overview of deep learning, including a brief history of machine learning and neural networks. It discusses various deep learning models such as deep belief networks, convolutional neural networks, and recurrent neural networks. Applications of deep learning in areas like computer vision, natural language processing, and robotics are also covered. Finally, popular platforms, frameworks and libraries for developing deep learning systems are mentioned.
1 of 53
More Related Content
Big Data Malaysia - A Primer on Deep Learning
1. A primer on Deep
Learning
Poo Kuan Hoong
19th September 2016
2. Data Science Institute
• The Data Science Institute is a research
center based in the Faculty of Computing
& Informatics, Multimedia University.
• The members comprise of expertise
across faculties such as Faculty of
Computing and Informatics, Faculty of
Engineering, Faculty of Management &
Faculty of Information Science and
Technology.
• Conduct research in leading data science
areas including stream mining, video
analytics, machine learning, deep
learning, next generation data
visualization and advanced data
modelling.
6. Acknowledgement
Andrew Ng: Deep Learning, Self-Taught
Learning and Unsupervised Feature
Learning [Youtube]
Yann LeCun: Deep Learning
Tutorial, ICML, Atlanta, 2013 [PDF]
Geoff Hinton, Yoshua Bengio & Yann LeCun: Deep
Learning: NIPS2015 Tutorial [PDF]
Yoshua Bengio: Theano: A Python
framework for fast computation of
mathematical expressions. [URL]
7. Outline
• A brief history of machine learning
• Understanding the human brain
• Neural Network: Concept, implementation and challenges
• Deep Belief Network (DBN): Concept and Application
• Convolutional Neural Network (CNN): Concept and Application
• Recurrent Neural Network (RNN): Concept and Application
• Deep Learning: Strengths, weaknesses and applications
• Deep Learning: Platforms, frameworks and libraries
• Demo
8. Introduction
• In the past 10 years, machine learning and
Artificial Intelligence (AI) have shown
tremendous progress
• The recent success can be attributed to:
• Explosion of data
• Cheap computing cost – CPUs and GPUs
• Improvement of machine learning models
• Much of the current excitement concerns a
subfield of it called “deep learning”.
9. A brief history of Machine learning
• Most of the machine learning methods are based on supervised
learning
Input
Feature
Representation
Learning Algorithm
16. Neural Network
• Deep Learning is primarily about neural networks, where a network is
an interconnected web of nodes and edges.
• Neural nets were designed to perform complex tasks, such as the task
of placing objects into categories based on a few attributes.
• Neural nets are highly structured networks, and have three kinds of
layers - an input, an output, and so called hidden layers, which refer
to any layers between the input and the output layers.
• Each node (also called a neuron) in the hidden and output layers has
a classifier.
18. Neural Network: Forward Propagation
• The input neurons first receive the
data features of the object. After
processing the data, they send their
output to the first hidden layer.
• The hidden layer processes this output
and sends the results to the next
hidden layer.
• This continues until the data reaches
the final output layer, where the
output value determines the object's
classification.
• This entire process is known as
Forward Propagation, or Forward prop.
19. Neural Network: Backward Propagation
• To train a neural network over a large set of labelled data, you must
continuously compute the difference between the network’s
predicted output and the actual output.
• This difference is called the cost, and the process for training a net is
known as backpropagation, or backprop
• During backprop, weights and biases are tweaked slightly until the
lowest possible cost is achieved.
• An important aspect of this process is the gradient, which is a
measure of how much the cost changes with respect to a change in a
weight or bias value.
20. The 1990s view of what was wrong with back-
propagation
• It required a lot of labelled training data
• almost all data is unlabeled
• The learning time did not scale well
• It was very slow in networks with multiple hidden layers.
• It got stuck at local optima
• These were often surprisingly good but there was no good theory
21. Deep Belief Network (DBN)
• The Deep Belief Network, or DBN,
was also conceived by Geoff Hinton.
• Used by Google for their work on the
image recognition problem.
• DBN is trained two layers at a time,
and these two layers are treated like
an RBM.
• Throughout the net, the hidden layer
of an RBM acts as the input layer of
the adjacent one. So the first RBM is
trained, and its outputs are then
used as inputs to the next RBM. This
procedure is repeated until the
output layer is reached.
22. Deep Belief Network (DBN)
• DBN is capable of recognizing the inherent patterns in the data. In
other words, it’s a sophisticated, multilayer feature extractor.
• The unique aspect of this type of net is that each layer ends up
learning the full input structure.
• Layers generally learn progressively complex patterns – for facial
recognition, early layers could detect edges and later layers would
combine them to form facial features.
• DBN learns the hidden patterns globally, like a camera slowly bringing
an image into focus.
• DBN still requires a set of labels to apply to the resulting patterns. As
a final step, the DBN is fine-tuned with supervised learning and a
small set of labeled examples.
24. Convolutional Neural Network (CNN)
• CNN inspired by the Visual Cortex.
• CNNs are deep nets that are used for image, object, and even speech
recognition.
• Pioneered by Yann Lecun (NYU)
• Deep supervised neural networks are generally too difficult to train.
• CNNs have multiple types of layers, the first of which is the
convolutional layer.
25. Convolutional Neural Network (CNN)
• A series of filters forms layer one, called the convolutional layer. The weights and
biases in this layer determine the effectiveness of the filtering process.
• Each flashlight represents a single neuron. Typically, neurons in a layer activate or
fire. On the other hand, in the convolutional layer, neurons search for patterns
through convolution. Neurons from different filters search for different patterns,
and thus they will process the input differently.
Filter 2/ Neural 2
W1=10
W3=4
W2=5
27. CNN: Application
• Classify a scene in an image
• Image Classifier Demo (NYU): http://horatio.cs.nyu.edu/
• Describe or understanding an image
• Toronto Deep Learning Demo: http://deeplearning.cs.toronto.edu/i2t
• MIT Scene Recognition Demo: http://places.csail.mit.edu/demo.html
• Handwriting recognition
• Handwritten digits recognition:
http://cs.stanford.edu/people/karpathy/convnetjs/demo/mnist.html
• Video classification
• Large-scale Video Classification with Convolutional Neural Networks
http://cs.stanford.edu/people/karpathy/deepvideo/
28. Recurrent Neural Network (RNN)
• The Recurrent Neural Net (RNN) is
the brainchild of Juergen
Schmidhuber and Sepp Hochreiter.
• RNNs have a feedback loop where
the net’s output is fed back into
the net along with the next input.
• RNNs receive an input and produce
an output. Unlike other nets, the
inputs and outputs can come in a
sequence.
• Variant of RNN is Long Term Short
Memory (LSTM)
29. RNN: Application
• RNN is suitable for time series data, where an output can be the next
value in a sequence, or the next several values
Classify Image frame by
frame
Image captioning Document Classification
30. Deep Learning: Benefits
• Robust
• No need to design the features ahead of time – features are automatically
learned to be optimal for the task at hand
• Robustness to natural variations in the data is automatically learned
• Generalizable
• The same neural net approach can be used for many different applications
and data types
• Scalable
• Performance improves with more data, method is massively parallelizable
31. Deep Learning: Weaknesses
• Deep Learning requires a large dataset, hence long training period.
• In term of cost, Machine Learning methods like SVMs and other tree
ensembles are very easily deployed even by relative machine learning
novices and can usually get you reasonably good results.
• Deep learning methods tend to learn everything. It’s better to encode prior
knowledge about structure of images (or audio or text).
• The learned features are often difficult to understand. Many vision features
are also not really human-understandable (e.g,
concatenations/combinations of different features).
• Requires a good understanding of how to model multiple modalities with
traditional tools.
32. Deep Learning: Application in our daily lives
Better voice recognition
Search your own photos without labeling
Facebook’s DeepFace Recognition has accuracy of
97.25%
35. Deep Learning: Platform & Frameworks & Libraries
Platform
• H20 – deep learning framework that comes with R and Python interfaces
[http://www.h2o.ai/verticals/algos/deep-learning/]
Framework
• Caffe - deep learning framework made with expression, speed, and modularity in mind.
Developed by the Berkeley Vision and Learning Center (BVLC) [http://caffe.berkeleyvision.org/]
• Torch - scientific computing framework with wide support for machine learning algorithms that
puts GPUs first. Based on Lua programming language [http://torch.ch/]
Library
• Tensorflow - open source software library for numerical computation using data flow graphs
from Google [https://www.tensorflow.org/]
• Theano - a python library developed by Yoshua Bengio’s team
[http://deeplearning.net/software/theano/]
• Microsoft/CNTK: Computational Network Toolkit
https://github.com/Microsoft/CNTK
• Baidu’s PaddlePaddle: Deep Learning AI Platform
https://github.com/baidu/Paddle
36. Learned Models
• Trained Models can be shared with others
• Save the training time
• For example: AlexNet (2012: 5 conv layers & 3 conn layers),
GoogLeNet (2014: 22 layers deep network), ParseNet (2015) , etc
• URLs:
• https://github.com/BVLC/caffe/wiki/Model-Zoo
• http://deeplearning4j.org/model-zoo
38. Nvidia: Digits
• The NVIDIA Deep Learning GPU Training System (DIGITS) puts the
power of deep learning in the hands of data scientists and
researchers.
• Quickly design the best deep neural network (DNN) for your data
using real-time network behavior visualization.
• https://developer.nvidia.com/digits