This document discusses transfer learning using TensorFlow. It begins with an introduction to deep learning and its applications. TensorFlow is introduced as an open-source library for machine learning using data flow graphs. Transfer learning is described as a technique where a model trained on one domain is reused on another domain by retraining or fine-tuning the last layers while keeping earlier layers fixed. This allows building accurate models using small datasets by leveraging knowledge gained from large datasets. The document demonstrates performing transfer learning using TensorFlow to retrain an Inception V3 model for a new image classification task.
1 of 33
More Related Content
Explore and Have Fun with TensorFlow: Transfer Learning
3. TensorFlow & Deep Learning Malaysia Group
The TensorFlow & Deep Learning
Malaysia group's aims are:
• To enable people to create and deploy
their own Deep Learning models built
using primarily TensorFlow or other
Deep Learning libraries.
• To build the key skill sets for this group
from the combination of both beginner
and intermediate models as well as
advancing to the next level
• A knowledge sharing and presentations
platform in relation to the cutting edge
deep learning research papers and
techniques.
11. Deep Learning: Strengths
• Robust
• No need to design the features ahead of time - features are
automatically learned to be optimal for the task at hand
• Robustness to natural variations in the data is automatically learned
• Generalizable
• The same neural net approach can be used for many different
applications and data types
• Scalable
• Performance improves with more data, method is massively
parallelizable
15. What is TensorFlow?
• URL: https://www.tensorflow.org/
• Released under the open source license on
November 9, 2015
• Current version 1.2
• Open source software library for
numerical computation using data flow
graphs
• Originally developed by Google Brain Team
to conduct machine learning and deep
neural networks research
• General enough to be applicable in a wide
variety of other domains as well
• TensorFlow provides an extensive suite of
functions and classes that allow users to
build various models from scratch.
18. TensorFlow Models
https://github.com/tensorflow/models
Models
• adversarial_crypto: protecting communications with adversarial neural cryptography.
• adversarial_text: semi-supervised sequence learning with adversarial training.
• attention_ocr: a model for real-world image text extraction.
• autoencoder: various autoencoders.
• cognitive_mapping_and_planning: implementation of a spatial memory based mapping
and planning architecture for visual navigation.
• compression: compressing and decompressing images using a pre-trained Residual GRU
network.
• differential_privacy: privacy-preserving student models from multiple teachers.
• domain_adaptation: domain separation networks.
• im2txt: image-to-text neural network for image captioning.
• inception: deep convolutional networks for computer vision.
25. Inception V3
• Inception-v3 is trained for the ImageNet Large Visual Recognition
Challenge using the data from 2012.
• This is a standard task in computer vision, where models try to
classify entire images into 1000 classes, like "Zebra", "Dalmatian", and
"Dishwasher".
30. Transfer Learning with TensorFlow
• Transfer learning does not require GPUs to train
• Training across the training set (2,000 images) took less than a minute
on my Macbook Pro without GPU support. This is not entirely
surprising though, as the final model is just a softmax regression.
• With TensorBoard, it is able to provide summaries that make it easier
to understand, debug, and optimize the retraining.
• It is also able visualize the graph and statistics, such as how the
weights or accuracy varied during training.
• https://www.tensorflow.org/tutorials/image_retraining