Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Deep Learning

Download as pdf or txt
Download as pdf or txt
You are on page 1of 15

BENGAL COLLEGE OF ENGINEERING AND

TECHNOLOGY
DEPARTMENT OF INFORMATION TECHNOLOGY

SESSION (2018-2019)

A SEMINAR REPORT

SUBMITTED IN BENGAL COLLEGE OF ENGG. AND TECHNOLOGY

DURGAPUR (WEST BENGAL)

“DEEP LEARNING”

SUBMITTED TO SUBMITTED BY
Mr. Biplab Palye Abhishek Dwivedi(04)
HOD(IT Dept.) B.Tech 6th Semester
Index
1. Introduction
2. Overview
3. Definition
4. Neural Networks
5. Applications
6. Conclusion
7. Bibliography
Introduction
Deep learning(also known as deep structured
learning or hierarchical learning) is part of a broader family
of machine learningmethods based on the layers used in artificial neural
networks. Learning can be supervised, semi-supervised or unsupervised.

Deep learning architectures such as deep neural networks, deep belief


networks, recurrent neural networks and convolutional neural
networks have been applied to fields including computer vision, speech
recognition, natural language processing, audio recognition, social
network filtering, machine translation, bioinformatics, drug design,
medical image analysis, material inspection and board gameprograms,
where they have produced results comparable to and in some cases
superior to human experts.

Neural networks were originally inspired by information processing and


distributed communication nodes in biological systems synaptic
structures yet have various differences from the structural and functional
properties of biological brains, which make them incompatible with
the neurological evidence. Specifically, Neural networks tend to be static
and symbolic, while the biological brain of most living organisms is
dynamic (plasticity) and analog.
Overview
Most modern deep learning models are based on an artificial neural
networks, specifically, Convolutional Neural Networks (CNN)s,
although they can also include propositional formulas or latent variables
organized layer-wise in deep generative models such as the nodes
in deep belief networks and deep Boltzmann machines.

In deep learning, each level learns to transform its input data into a
slightly more abstract and composite representation. In an image
recognition application, the raw input may be a matrix of pixels; the first
representational layer may abstract the pixels and encode edges; the
second layer may compose and encode arrangements of edges; the third
layer may encode a nose and eyes; and the fourth layer may recognize
that the image contains a face. Importantly, a deep learning process can
learn which features to optimally place in which level on its own. (Of
course, this does not completely obviate the need for hand-tuning; for
example, varying numbers of layers and layer sizes can provide different
degrees of abstraction.)

The "deep" in "deep learning" refers to the number of layers through


which the data is transformed. More precisely, deep learning systems
have a substantial credit assignment path(CAP) depth. The CAP is the
chain of transformations from input to output. CAPs describe potentially
causal connections between input and output. For a feedforward neural
network, the depth of the CAPs is that of the network and is the number
of hidden layers plus one (as the output layer is also parameterized).
For recurrent neural networks, in which a signal may propagate through
a layer more than once, the CAP depth is potentially unlimited.[2] No
universally agreed upon threshold of depth divides shallow learning
from deep learning, but most researchers agree that deep learning
involves CAP depth > 2. CAP of depth 2 has been shown to be a
universal approximator in the sense that it can emulate any
function. Beyond that more layers do not add to the function
approximator ability of the network. Deep models (CAP > 2) are able to
extract better features than shallow models and hence, extra layers help
in learning features.

Deep learning architectures are often constructed with a greedy layer-by-


layer method. Deep learning helps to disentangle these abstractions and
pick out which features improve performance.

For supervised learning tasks, deep learning methods obviate feature


engineering, by translating the data into compact intermediate
representations akin to principal components, and derive layered
structures that remove redundancy in representation.

Deep learning algorithms can be applied to unsupervised learning tasks.


This is an important benefit because unlabeled data are more abundant
than labeled data. Examples of deep structures that can be trained in an
unsupervised manner are neural history compressors and deep belief
networks.
Definition
Deep learning is a class of machine learning algorithms that:

 use a cascade of multiple layers of nonlinear processing units


for feature extraction and transformation. Each successive layer uses
the output from the previous layer as input.
 learn in supervised (e.g., classification) and/or unsupervised (e.g.,
pattern analysis) manners.
 learn multiple levels of representations that correspond to different
levels of abstraction; the levels form a hierarchy of concepts.

Deep neural networks are generally interpreted in terms of the universal


approximation theorem or probabilistic inference.
The classic universal approximation theorem concerns the capacity
of feedforward neural networks with a single hidden layer of finite size
to approximate continuous functions. In 1989, the first proof was
published by George Cybenko for sigmoid activation functions and was
generalised to feed-forward multi-layer architectures in 1991 by Kurt
Hornik.
The universal approximation theorem for deep neural networks concerns
the capacity of networks with bounded width but the depth is allowed to
grow. Lu et al. proved that if the width of a deep neural
network with ReLU activation is strictly larger than the input dimension,
then the network can approximate any Lebesgue integrable function; If
the width is smaller or equal to the input dimension, then deep neural
network is not a universal approximator.
The probabilistic interpretation derives from the field of machine
learning. It features inference, as well as the optimization concepts
of training and testing, related to fitting and generalization, respectively.
More specifically, the probabilistic interpretation considers the
activation nonlinearity as a cumulative distribution function. The
probabilistic interpretation led to the introduction
of dropout as regularizer in neural networks. The probabilistic
interpretation was introduced by researchers
including Hopfield, Widrow and Narendra and popularized in surveys
such as the one by Bishop.
Neural Networks
Artificial Neural Networks
Artificial neural networks (ANNs) or connectionist systems are
computing systems inspired by the biological neural networks that
constitute animal brains. Such systems learn (progressively improve
their ability) to do tasks by considering examples, generally without
task-specific programming. For example, in image recognition, they
might learn to identify images that contain cats by analyzing example
images that have been manually labeled as "cat" or "no cat" and using
the analytic results to identify cats in other images. They have found
most use in applications difficult to express with a traditional computer
algorithm using rule-based programming.
An ANN is based on a collection of connected units called artificial
neurons, (analogous to biological neurons in a biological brain). Each
connection (synapse) between neurons can transmit a signal to another
neuron. The receiving (postsynaptic) neuron can process the signal(s)
and then signal downstream neurons connected to it. Neurons may have
state, generally represented by real numbers, typically between 0 and 1.
Neurons and synapses may also have a weight that varies as learning
proceeds, which can increase or decrease the strength of the signal that it
sends downstream.
Typically, neurons are organized in layers. Different layers may perform
different kinds of transformations on their inputs. Signals travel from the
first (input), to the last (output) layer, possibly after traversing the layers
multiple times.
The original goal of the neural network approach was to solve problems
in the same way that a human brain would. Over time, attention focused
on matching specific mental abilities, leading to deviations from biology
such as backpropagation, or passing information in the reverse direction
and adjusting the network to reflect that information.
Neural networks have been used on a variety of tasks, including
computer vision, speech recognition, machine translation, social
network filtering, playing board and video gamesand medical diagnosis.
As of 2017, neural networks typically have a few thousand to a few
million units and millions of connections. Despite this number being
several order of magnitude less than the number of neurons on a human
brain, these networks can perform many tasks at a level beyond that of
humans (e.g., recognizing faces, playing "Go" ).

Deep Neural Networks


A deep neural network (DNN) is an artificial neural network (ANN)
with multiple layers between the input and output layers.[11][2] The DNN
finds the correct mathematical manipulation to turn the input into the
output, whether it be a linear relationship or a non-linear relationship.
The network moves through the layers calculating the probability of
each output. For example, a DNN that is trained to recognize dog breeds
will go over the given image and calculate the probability that the dog in
the image is a certain breed. The user can review the results and select
which probabilities the network should display (above a certain
threshold, etc.) and return the proposed label. Each mathematical
manipulation as such is considered a layer, and complex DNN have
many layers, hence the name "deep" networks. The goal is that
eventually, the network will be trained to decompose an image
into features, identify trends that exist across all samples and classify
new images by their similarities without requiring human input.[101]
DNNs can model complex non-linear relationships. DNN architectures
generate compositional models where the object is expressed as a
layered composition of primitives. The extra layers enable composition
of features from lower layers, potentially modeling complex data with
fewer units than a similarly performing shallow network.
Deep architectures include many variants of a few basic approaches.
Each architecture has found success in specific domains. It is not always
possible to compare the performance of multiple architectures, unless
they have been evaluated on the same data sets.
DNNs are typically feedforward networks in which data flows from the
input layer to the output layer without looping back. At first, the DNN
creates a map of virtual neurons and assigns random numerical values,
or "weights", to connections between them. The weights and inputs are
multiplied and return an output between 0 and 1. If the network didn’t
accurately recognize a particular pattern, an algorithm would adjust the
weights. That way the algorithm can make certain parameters more
influential, until it determines the correct mathematical manipulation to
fully process the data.
Recurrent neural networks (RNNs), in which data can flow in any
direction, are used for applications such as language modeling. Long
short-term memory is particularly effective for this use.
Convolutional deep neural networks (CNNs) are used in computer
vision. CNNs also have been applied to acoustic modeling for automatic
speech recognition (ASR).
Applications
Automatic Speech Recognition
Large-scale automatic speech recognition is the first and most
convincing successful case of deep learning. LSTM RNNs can learn
"Very Deep Learning" tasks[2] that involve multi-second intervals
containing speech events separated by thousands of discrete time steps,
where one time step corresponds to about 10 ms. LSTM with forget
gates is competitive with traditional speech recognizers on certain tasks.
The initial success in speech recognition was based on small-scale
recognition tasks based on TIMIT. The data set contains 630 speakers
from eight major dialects of American English, where each speaker
reads 10 sentences. Its small size lets many configurations be tried. More
importantly, the TIMIT task concerns phone-sequence recognition,
which, unlike word-sequence recognition, allows weak
phone bigram language models. This lets the strength of the acoustic
modeling aspects of speech recognition be more easily analyzed. The
error rates listed below, including these early results and measured as
percent phone error rates (PER), have been summarized since 1991.
Image Recognition
A common evaluation set for image classification is the MNIST
database data set. MNIST is composed of handwritten digits and
includes 60,000 training examples and 10,000 test examples. As with
TIMIT, its small size lets users test multiple configurations. A
comprehensive list of results on this set is available.
Deep learning-based image recognition has become "superhuman",
producing more accurate results than human contestants. This first
occurred in 2011.
Deep learning-trained vehicles now interpret 360° camera
views. Another example is Facial Dysmorphology Novel Analysis
(FDNA) used to analyze cases of human malformation connected to a
large database of genetic syndromes.
Visual Art Processing
Closely related to the progress that has been made in image recognition
is the increasing application of deep learning techniques to various
visual art tasks. DNNs have proven themselves capable, for example, of
a) identifying the style period of a given painting, b) Neural Style
Transfer - capturing the style of a given artwork and applying it in a
visually pleasing manner to an arbitrary photograph or video, and c)
generating striking imagery based on random visual input fields.
Natural Language Processing
Neural networks have been used for implementing language models
since the early 2000s. LSTM helped to improve machine translation and
language modeling.
Other key techniques in this field are negative sampling and word
embedding. Word embedding, such as word2vec, can be thought of as a
representational layer in a deep learning architecture that transforms an
atomic word into a positional representation of the word relative to other
words in the dataset; the position is represented as a point in a vector
space. Using word embedding as an RNN input layer allows the network
to parse sentences and phrases using an effective compositional vector
grammar. A compositional vector grammar can be thought of
as probabilistic context free grammar (PCFG) implemented by an
RNN. Recursive auto-encoders built atop word embeddings can assess
sentence similarity and detect paraphrasing. Deep neural architectures
provide the best results for constituency parsing, sentiment
analysis,[136] information retrieval, spoken language
understanding, machine translation, contextual entity linking, writing
style recognition, Text classification and others.
Recent developments generalize word embedding to sentence
embedding.
Google Translate (GT) uses a large end-to-end long short-term memory
network. Google Neural Machine Translation (GNMT) uses an example-
based machine translation method in which the system "learns from
millions of examples." It translates "whole sentences at a time, rather
than pieces. Google Translate supports over one hundred languages. The
network encodes the "semantics of the sentence rather than simply
memorizing phrase-to-phrase translations". GT uses English as an
intermediate between most language pairs.
Recommendation Systems
Recommendation systems have used deep learning to extract meaningful
features for a latent factor model for content-based music
recommendations. Multiview deep learning has been applied for
learning user preferences from multiple domains. The model uses a
hybrid collaborative and content-based approach and enhances
recommendations in multiple tasks.
Conclusion

Deep learning is a quite resource-demanding technology. It requires


more powerful GPUs, high-performance graphics processing units, large
amounts of storage to train the models, etc. Furthermore, this technology
needs more time to train in comparison with traditional machine
learning.

Despite all its challenges, deep learning discovers new improved


methods of unstructured big data analytics for those with the intention
to use it. Indeed, businesses can gain significant benefits from using
deep learning within their tasks of data processing. Though, the question
is not whether this technology is useful, rather how companies can
implement it in their projects to improve the way they process data.
Bibliography

1. https://www.upwork.com/hiring/for-clients/deep-learning-
definition-benefits-challenges/
2. https://en.wikipedia.org/wiki/Deep_learning

You might also like