SlideShare a Scribd company logo
Lesson 6
Deep Learning
Legal Notices and Disclaimers
This presentation is for informational purposes only. INTEL MAKES NO
WARRANTIES, EXPRESS OR IMPLIED, IN THIS SUMMARY.
Intel technologies’ features and benefits depend on system configuration and may
require enabled hardware, software or service activation. Performance varies
depending on system configuration. Check with your system manufacturer or retailer
or learn more at intel.com.
This sample source code is released under the Intel Sample Source Code License
Agreement.
Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other
countries.
*Other names and brands may be claimed as the property of others.
Copyright © 2018, Intel Corporation. All rights reserved.
3
Learning Objectives
• Identify the types of problems Deep Learning resolves
• Describe the steps in building a neural network model
• Describe a convolutional neural network
• Explain transfer learning and why it’s useful
• Identify common Deep Learning architectures
You will be able to:
Deep learning summary
5
Deep Learning
“Machine learning that involves using
very complicated models called “deep
neural networks”." (Intel)
Models determine best representation
of original data; in classic machine
learning, humans must do this.
6
Classic
Machine
Learning
Step 1: Determine
features.
Step 2: Feed them
through model.
“Arjun"
Neural Network
“Arjun"
Deep
Learning
Steps 1 and 2
are combined
into 1 step.
Deep Learning Differences
Machine
Learning
Classifier
Algorithm
Feature
Detection
7
Deep Learning Problem Types
Deep Learning can solve multiple supervised and unsupervised problems.
• The majority of its success has been when working
with images, natural language, and audio data.
• Image classification and detection.
• Semantic segmentation.
• Natural language object retrieval.
• Speech recognition and language translation.
8
Classification and Detection
Detect and label the image
• Person
• Motor Bike
https://people.eecs.berkeley.edu/~jhoffman/talks/lsda-baylearn2014.pdf
9
Semantic Segmentation
Label every pixel
https://people.eecs.berkeley.edu/~jhoffman/talks/lsda-baylearn2014.pdf
10
Natural Language Object Retrieval
http://arxiv.org/pdf/1511.04164v3.pdf
11
Speech Recognition and Language Translation
http://svail.github.io/mandarin/
The same architecture can be used for speech recognition in English, or
in Mandarin Chinese.
Deep learning summary
Formulating Supervised Learning Tools
13
For a supervised learning problem:
• Collect a labeled dataset (features and target labels).
• Choose the model.
• Choose an evaluation metric:
“What to use to measure performance.”
• Choose an optimization method:1
“How to find the model configuration that gives the best performance.”
1 There are standard methods to use for different models and metrics.
Which Model?
14
There are many models that represent the problem and make decisions in different
ways, each with their own advantages and disadvantages.
• DL models are biologically inspired.
• The main building block is a neuron.
Neuron
Neuron
15
A neuron multiplies each feature by a weight and then adds these
values together.
• Z = X1W1+ X2W2+ X3W3
Z
X1
X2
X3
W2
W1
W3
Neuron
16
This value is then put through a function called the activation function.
• There are several activation functions
that can be used.
• The output of the neuron is the output
of the activation function.
Activation
Function
X1
X2
X3
W2
W1
W3
Output
Perceptron
17
A neuron with a simple activation function can solve linearly
separable problems.
• These are problems where the different
classes can be separated by a line.
• The Perceptron: one of the earliest neural
network models that used neurons with
simple activation functions.
Perceptron
18
Problems where the labels cannot be separated by a single line are not
solvable by a single neuron.
• This is a major limitation, and one of the
reasons for the first AI winter, that was
discussed in lesson 1.
19
Fully Connected Network
http://svail.github.io/mandarin/
More complicated problems can be solved by connecting multiple
neurons together and using more complicated activation functions.
• Organized into layers of neurons.
• Each neuron is connected to every
neuron in the previous layer.
• Each layer transforms the output of
the previous layer and then passes it
on to the next.
• Every connection has a separate weight.
Deep Learning
Deep Learning refers to when many layers are used to build deep networks.
• State-of-the-art models use hundreds of layers.
• Deep layers tend to decrease in width.
• Successive layers transform inputs with two effects:
• Compression: each layer is asked to summarize the
input in a way that best serves the task.
• Extraction: the model succeeds when each layer
extracts task-relevant information.
Input
Output
21
Steps in Building a Fully Connected Network
To build a fully connected network a user needs to:
• Define the network architecture.
• How many layers and neurons?
• Define what activation function to
use for each neuron.
• Define an evaluation metric.
• The values for the weights are learned
during model training.
http://svail.github.io/mandarin/
22
Evaluation Metric
The metric used will depend on the problem being solved.
Some examples include:
• Regression
• Mean Squared Error
• Classification
• Categorical Cross-Entropy
• Multi-Label classification
• Binary Cross-Entropy
23
Fully Connected Network Problems
Not optimal for detecting features.
• Computationally intensive – heavy memory usage.
Deep learning summary
25
Convolutional Neural Network
Convolutional neural networks reduce the required
computation and are good for detecting features.
• Each neuron is connected to a small set of nearby
neurons in the previous layer.
• The same set of weights are used for each neuron.
• Ideal for spatial feature recognition.
• Example: image recognition
• Cheaper on resources due to fewer connections.
http://svail.github.io/mandarin/
26
Convolutions as Feature Detectors
-1 1 -1
-1 1 -1
-1 1 -1
Vertical Line Detector
-1 -1 -1
1 1 1
-1 -1 -1
Horizontal Line Detector
-1 -1 -1
-1 1 1
-1 1 1
Corner Detector
Convolutions can be thought of as “local feature detectors”.
27
Convolutions
28
Convolutional Neural Network
Convolutions
Fully Connected
Deep learning summary
30
Transfer Learning
There are difficulties with building convolutional neural networks.
• They require huge datasets.
• There is a large amount of required computation.
• A large amount of time is spent experimenting to
get the hyper-parameters correct.
• It would be very difficult to train a famous,
competition-winning model from scratch
31
Transfer Learning
We can take advantage of existing DL models.
• Early layers in a neural network are the hardest (slowest) to train.
• These ”primitive” features should be general across many image
classification tasks.
• Later layers in the network capture features that are more particular to
the specific image classification problem.
• Later layers are easier (quicker) to train since adjusting their weights has
a more immediate impact on the final result.
32
Transfer Learning
Keeping the early layers of a pre-trained network, and re-training the
later layers for a specific application, is called transfer learning.
• Results of the training are weights (numbers) that are easy to store.
• Using pre-trained layers reduces the amount of required data.
33
Convolutional Neural Network
Convolutions
Fully Connected
Fix Weights
Train only the
last layer
34
Transfer Learning Options
The additional training of a pre-trained network on a specific new
dataset is referred to as “fine-tuning”.
• Choose “how much” and “how far back” to fine-tune.
• Should I train just the very last layer, or go back a few layers?
• Re-train the entire network (from the starting-point of the existing
network)?
Deep learning summary
36
AlexNet
37
VGG 16
38
Inception
39
MobileNets
Efficient models for mobile and embedded vision applications.
MobileNet models can be applied to various recognition tasks for efficient device intelligence.
40
Learning Objectives
• Identify the types of problems Deep Learning resolves.
• Describe the steps in building a neural network model.
• Describe a convolutional neural network.
• Explain transfer learning and why it’s useful.
• Identify common Deep Learning architectures.
In this session we worked to:
Deep learning summary

More Related Content

PDF
What is Deep Learning | Deep Learning Simplified | Deep Learning Tutorial | E...
PPTX
Convolutional neural network
PPTX
Image classification using cnn
PPTX
Convolutional Neural Network
PDF
Introduction to Deep Learning
PPTX
Google net
PDF
Anomaly Detection using Deep Auto-Encoders
PDF
Artificial Neural Network Tutorial | Deep Learning With Neural Networks | Edu...
What is Deep Learning | Deep Learning Simplified | Deep Learning Tutorial | E...
Convolutional neural network
Image classification using cnn
Convolutional Neural Network
Introduction to Deep Learning
Google net
Anomaly Detection using Deep Auto-Encoders
Artificial Neural Network Tutorial | Deep Learning With Neural Networks | Edu...

What's hot (20)

PDF
An introduction to Deep Learning
PPTX
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
PPTX
Batch normalization presentation
PDF
GANs and Applications
PPTX
Convolutional Neural Network (CNN)
PPTX
Convolutional Neural Network (CNN) - image recognition
PPTX
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
PDF
Chap4 Récursivité en python
PDF
LeNet-5
PDF
Graph neural networks overview
PDF
Convolutional Neural Networks (CNN)
PPTX
Introduction to Deep learning
PDF
Recursive Neural Networks
PPTX
Deep learning
PPTX
Convolutional neural network from VGG to DenseNet
PPTX
CONVOLUTIONAL NEURAL NETWORK
PPTX
Convolutional Neural Network (CNN)
PDF
Python avancé : Lecture et écriture de fichiers
PPTX
What is Deep Learning?
PDF
PyTorch Python Tutorial | Deep Learning Using PyTorch | Image Classifier Usin...
An introduction to Deep Learning
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
Batch normalization presentation
GANs and Applications
Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN) - image recognition
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
Chap4 Récursivité en python
LeNet-5
Graph neural networks overview
Convolutional Neural Networks (CNN)
Introduction to Deep learning
Recursive Neural Networks
Deep learning
Convolutional neural network from VGG to DenseNet
CONVOLUTIONAL NEURAL NETWORK
Convolutional Neural Network (CNN)
Python avancé : Lecture et écriture de fichiers
What is Deep Learning?
PyTorch Python Tutorial | Deep Learning Using PyTorch | Image Classifier Usin...
Ad

Similar to Deep learning summary (20)

PPTX
Introduction to deep learning
PDF
imageclassification-160206090009.pdf
PPTX
B4UConference_machine learning_deeplearning
PPTX
Deep learning
PPTX
Image classification with Deep Neural Networks
PDF
NLP and Deep Learning for non_experts
PPTX
Artificial Intelligence, Machine Learning and Deep Learning
PDF
Practical deepllearningv1
PPTX
Deep learning introduction
PPTX
tensorflow.pptx
PPTX
Strata London - Deep Learning 05-2015
PPTX
A simple presentation for deep learning.
PPTX
Automatic Attendace using convolutional neural network Face Recognition
PDF
Deep Learning Study _ FInalwithCNN_RNN_LSTM_GRU.pdf
PDF
Deep learning: Cutting through the Myths and Hype
PDF
Deep Learning & Tensor flow: An Intro
PDF
Hardware Acceleration for Machine Learning
PDF
Deep Learning Demystified
PPTX
presentation of IntroductionDeepLearning.pptx
PPTX
Session_2_Introduction_to_Deep_Learning.pptx
Introduction to deep learning
imageclassification-160206090009.pdf
B4UConference_machine learning_deeplearning
Deep learning
Image classification with Deep Neural Networks
NLP and Deep Learning for non_experts
Artificial Intelligence, Machine Learning and Deep Learning
Practical deepllearningv1
Deep learning introduction
tensorflow.pptx
Strata London - Deep Learning 05-2015
A simple presentation for deep learning.
Automatic Attendace using convolutional neural network Face Recognition
Deep Learning Study _ FInalwithCNN_RNN_LSTM_GRU.pdf
Deep learning: Cutting through the Myths and Hype
Deep Learning & Tensor flow: An Intro
Hardware Acceleration for Machine Learning
Deep Learning Demystified
presentation of IntroductionDeepLearning.pptx
Session_2_Introduction_to_Deep_Learning.pptx
Ad

More from ankit_ppt (20)

PPTX
08 neural networks
PPTX
07 learning
PPTX
06 image features
PPTX
05 contours seg_matching
PPTX
04 image transformations_ii
PPTX
03 image transformations_i
PPTX
02 image processing
PPTX
01 foundations
PPTX
Word2 vec
PPTX
Text similarity measures
PPTX
Text generation and_advanced_topics
PPTX
Nlp toolkits and_preprocessing_techniques
PPTX
Matrix decomposition and_applications_to_nlp
PPTX
Machine learning and_nlp
PPTX
Latent dirichlet allocation_and_topic_modeling
PPTX
Intro to nlp
PPTX
Ot regularization and_gradient_descent
PPTX
Ml10 dimensionality reduction-and_advanced_topics
PPTX
Ml9 introduction to-unsupervised_learning_and_clustering_methods
PPTX
Ml8 boosting and-stacking
08 neural networks
07 learning
06 image features
05 contours seg_matching
04 image transformations_ii
03 image transformations_i
02 image processing
01 foundations
Word2 vec
Text similarity measures
Text generation and_advanced_topics
Nlp toolkits and_preprocessing_techniques
Matrix decomposition and_applications_to_nlp
Machine learning and_nlp
Latent dirichlet allocation_and_topic_modeling
Intro to nlp
Ot regularization and_gradient_descent
Ml10 dimensionality reduction-and_advanced_topics
Ml9 introduction to-unsupervised_learning_and_clustering_methods
Ml8 boosting and-stacking

Recently uploaded (20)

PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PPTX
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
PPTX
Construction Project Organization Group 2.pptx
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PPTX
Internet of Things (IOT) - A guide to understanding
PDF
Well-logging-methods_new................
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PPTX
additive manufacturing of ss316l using mig welding
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PPTX
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
PPTX
UNIT 4 Total Quality Management .pptx
PDF
composite construction of structures.pdf
PPTX
Welding lecture in detail for understanding
PDF
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PPTX
Foundation to blockchain - A guide to Blockchain Tech
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Lecture Notes Electrical Wiring System Components
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
Construction Project Organization Group 2.pptx
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
Internet of Things (IOT) - A guide to understanding
Well-logging-methods_new................
Embodied AI: Ushering in the Next Era of Intelligent Systems
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
additive manufacturing of ss316l using mig welding
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
UNIT 4 Total Quality Management .pptx
composite construction of structures.pdf
Welding lecture in detail for understanding
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
Foundation to blockchain - A guide to Blockchain Tech

Deep learning summary

  • 2. Legal Notices and Disclaimers This presentation is for informational purposes only. INTEL MAKES NO WARRANTIES, EXPRESS OR IMPLIED, IN THIS SUMMARY. Intel technologies’ features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. Check with your system manufacturer or retailer or learn more at intel.com. This sample source code is released under the Intel Sample Source Code License Agreement. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as the property of others. Copyright © 2018, Intel Corporation. All rights reserved.
  • 3. 3 Learning Objectives • Identify the types of problems Deep Learning resolves • Describe the steps in building a neural network model • Describe a convolutional neural network • Explain transfer learning and why it’s useful • Identify common Deep Learning architectures You will be able to:
  • 5. 5 Deep Learning “Machine learning that involves using very complicated models called “deep neural networks”." (Intel) Models determine best representation of original data; in classic machine learning, humans must do this.
  • 6. 6 Classic Machine Learning Step 1: Determine features. Step 2: Feed them through model. “Arjun" Neural Network “Arjun" Deep Learning Steps 1 and 2 are combined into 1 step. Deep Learning Differences Machine Learning Classifier Algorithm Feature Detection
  • 7. 7 Deep Learning Problem Types Deep Learning can solve multiple supervised and unsupervised problems. • The majority of its success has been when working with images, natural language, and audio data. • Image classification and detection. • Semantic segmentation. • Natural language object retrieval. • Speech recognition and language translation.
  • 8. 8 Classification and Detection Detect and label the image • Person • Motor Bike https://people.eecs.berkeley.edu/~jhoffman/talks/lsda-baylearn2014.pdf
  • 9. 9 Semantic Segmentation Label every pixel https://people.eecs.berkeley.edu/~jhoffman/talks/lsda-baylearn2014.pdf
  • 10. 10 Natural Language Object Retrieval http://arxiv.org/pdf/1511.04164v3.pdf
  • 11. 11 Speech Recognition and Language Translation http://svail.github.io/mandarin/ The same architecture can be used for speech recognition in English, or in Mandarin Chinese.
  • 13. Formulating Supervised Learning Tools 13 For a supervised learning problem: • Collect a labeled dataset (features and target labels). • Choose the model. • Choose an evaluation metric: “What to use to measure performance.” • Choose an optimization method:1 “How to find the model configuration that gives the best performance.” 1 There are standard methods to use for different models and metrics.
  • 14. Which Model? 14 There are many models that represent the problem and make decisions in different ways, each with their own advantages and disadvantages. • DL models are biologically inspired. • The main building block is a neuron. Neuron
  • 15. Neuron 15 A neuron multiplies each feature by a weight and then adds these values together. • Z = X1W1+ X2W2+ X3W3 Z X1 X2 X3 W2 W1 W3
  • 16. Neuron 16 This value is then put through a function called the activation function. • There are several activation functions that can be used. • The output of the neuron is the output of the activation function. Activation Function X1 X2 X3 W2 W1 W3 Output
  • 17. Perceptron 17 A neuron with a simple activation function can solve linearly separable problems. • These are problems where the different classes can be separated by a line. • The Perceptron: one of the earliest neural network models that used neurons with simple activation functions.
  • 18. Perceptron 18 Problems where the labels cannot be separated by a single line are not solvable by a single neuron. • This is a major limitation, and one of the reasons for the first AI winter, that was discussed in lesson 1.
  • 19. 19 Fully Connected Network http://svail.github.io/mandarin/ More complicated problems can be solved by connecting multiple neurons together and using more complicated activation functions. • Organized into layers of neurons. • Each neuron is connected to every neuron in the previous layer. • Each layer transforms the output of the previous layer and then passes it on to the next. • Every connection has a separate weight.
  • 20. Deep Learning Deep Learning refers to when many layers are used to build deep networks. • State-of-the-art models use hundreds of layers. • Deep layers tend to decrease in width. • Successive layers transform inputs with two effects: • Compression: each layer is asked to summarize the input in a way that best serves the task. • Extraction: the model succeeds when each layer extracts task-relevant information. Input Output
  • 21. 21 Steps in Building a Fully Connected Network To build a fully connected network a user needs to: • Define the network architecture. • How many layers and neurons? • Define what activation function to use for each neuron. • Define an evaluation metric. • The values for the weights are learned during model training. http://svail.github.io/mandarin/
  • 22. 22 Evaluation Metric The metric used will depend on the problem being solved. Some examples include: • Regression • Mean Squared Error • Classification • Categorical Cross-Entropy • Multi-Label classification • Binary Cross-Entropy
  • 23. 23 Fully Connected Network Problems Not optimal for detecting features. • Computationally intensive – heavy memory usage.
  • 25. 25 Convolutional Neural Network Convolutional neural networks reduce the required computation and are good for detecting features. • Each neuron is connected to a small set of nearby neurons in the previous layer. • The same set of weights are used for each neuron. • Ideal for spatial feature recognition. • Example: image recognition • Cheaper on resources due to fewer connections. http://svail.github.io/mandarin/
  • 26. 26 Convolutions as Feature Detectors -1 1 -1 -1 1 -1 -1 1 -1 Vertical Line Detector -1 -1 -1 1 1 1 -1 -1 -1 Horizontal Line Detector -1 -1 -1 -1 1 1 -1 1 1 Corner Detector Convolutions can be thought of as “local feature detectors”.
  • 30. 30 Transfer Learning There are difficulties with building convolutional neural networks. • They require huge datasets. • There is a large amount of required computation. • A large amount of time is spent experimenting to get the hyper-parameters correct. • It would be very difficult to train a famous, competition-winning model from scratch
  • 31. 31 Transfer Learning We can take advantage of existing DL models. • Early layers in a neural network are the hardest (slowest) to train. • These ”primitive” features should be general across many image classification tasks. • Later layers in the network capture features that are more particular to the specific image classification problem. • Later layers are easier (quicker) to train since adjusting their weights has a more immediate impact on the final result.
  • 32. 32 Transfer Learning Keeping the early layers of a pre-trained network, and re-training the later layers for a specific application, is called transfer learning. • Results of the training are weights (numbers) that are easy to store. • Using pre-trained layers reduces the amount of required data.
  • 33. 33 Convolutional Neural Network Convolutions Fully Connected Fix Weights Train only the last layer
  • 34. 34 Transfer Learning Options The additional training of a pre-trained network on a specific new dataset is referred to as “fine-tuning”. • Choose “how much” and “how far back” to fine-tune. • Should I train just the very last layer, or go back a few layers? • Re-train the entire network (from the starting-point of the existing network)?
  • 39. 39 MobileNets Efficient models for mobile and embedded vision applications. MobileNet models can be applied to various recognition tasks for efficient device intelligence.
  • 40. 40 Learning Objectives • Identify the types of problems Deep Learning resolves. • Describe the steps in building a neural network model. • Describe a convolutional neural network. • Explain transfer learning and why it’s useful. • Identify common Deep Learning architectures. In this session we worked to:

Editor's Notes

  • #7: For some tasks, like image recognition, Deep Learning can eliminate feature extraction.
  • #11: localize a target object within a given image based on a natural language query of the object. Natural language object retrieval differs from text-based image retrieval task as it involves spatial information about objects within the scene and global scene context
  • #12: CTC – Connectionist Temporal Classification English – Non tonal language, pitch is not important. So the traditional speech detection systems which convert raw audio to frequency lose out on pitch and work for english. Mandarin – Tonal, so pitch is important. Deep speech does not convert to frequency, so using raw input sound, train the NN for both english and mandarin Traditional speech detection systems extract abstract units of sound (Phonemes), map every word in the vocabulary to a set of phonemes. Adapting this to madarin would require constructing a new lexicon. Deep speech – does not need a new lexicon since the R-CNNs maps variable length speech to characters can adapt equally well to both english and mandarin. Language modelling – how probable is a string of words in a given language. English – words are separated by space, mandarin not. Deep speech uses a character based segmentation. All in all, most significant change is extending the last layer to include more characters in the mandarin language and deep speech works equally well.
  • #40: We’ll explore this more in this lesson’s assignment.