SlideShare a Scribd company logo
1 of 5
https://medium.com/aiguys/deep-convolutional-neural-networks-dcnns-explained-in-layman-terms-
b990b2818061
Deep Convolutional Neural Networks
(DCNNs) explained in layman's terms
In deep learning, a convolutional neural network (CNN, or ConvNet) is a class of artificial neural
networks, most commonly applied to analyze visual imagery. They are also known as shift
invariant or space invariant artificial neural networks (SIANN), based on the shared-weight
architecture of the convolution kernels or filters that slide along input features and provide
translation equivariant responses known as feature maps.
Basic DCNN architecture
DCNN Architecture
1. Convolutional Layer + Relu
2. Pooling Layer
3. Fully Connected Layer
4. Dropout
5. Activation Functions
1. Convolutional Layer + Relu
This layer is the first layer that is used to extract the various features from the input images. In
this layer, the mathematical operation of convolution is performed between the input image and a
filter of a particular size MxM. By sliding the filter over the input image, the dot product is taken
between the filter and the parts of the input image with respect to the size of the filter (MxM).
The output is termed as the Feature map which gives us information about the image such as
the corners and edges. Later, this feature map is fed to other layers to learn several other features
of the input image.
2. Pooling Layer
In most cases, a Convolutional Layer is followed by a Pooling Layer. The primary aim of this
layer is to decrease the size of the convolved feature map to reduce computational costs. This is
performed by decreasing the connections between layers and independently operating on each
feature map. Depending upon the method used, there are several types of Pooling operations.
In Max Pooling, the largest element is taken from the feature map. Average Pooling calculates
the average of the elements in a predefined sized Image section. The total sum of the elements in
the predefined section is computed in Sum Pooling. The Pooling Layer usually serves as a bridge
between the Convolutional Layer and the FC Layer
3. Fully Connected Layer
The Fully Connected (FC) layer consists of the weights and biases along with the neurons and is
used to connect the neurons between two different layers. These layers are usually placed before
the output layer and form the last few layers of a CNN Architecture.
In this, the input image from the previous layers is flattened and fed to the FC layer. The
flattened vector then undergoes a few more FC layers where mathematical operations usually
take place. In this stage, the classification process begins to take place.
4. Dropout
Usually, when all the features are connected to the FC layer, it can cause overfitting in the
training dataset. Overfitting occurs when a particular model works so well on the training data
causing a negative impact on the model’s performance when used on new data.
To overcome this problem, a dropout layer is utilized wherein a few neurons are dropped from
the neural network during the training process resulting in reduced size of the model. On passing
a dropout of 0.3, 30% of the nodes are dropped out randomly from the neural network.
5. Activation Functions
Finally, one of the most important parameters of the CNN model is the activation function. They
are used to learn and approximate any kind of continuous and complex relationship between
variables of the network. In simple words, it decides which information of the model should fire
in the forward direction and which ones should not at the end of the network.
It adds non-linearity to the network. There are several commonly used activation functions such
as the ReLU, Softmax, tanH, and the Sigmoid functions. Each of these functions has a specific
usage. For a binary classification CNN model, sigmoid and softmax functions are preferred and
for multi-class classification, softmax is used.
How does a DCNN work?
CNN compares images piece by piece. The pieces that it looks for are called features which are
nothing but a bunch of MxM matrices with numbers(images are nothing but MxM number
matrices of pixel values for a computer). By finding rough feature matches in roughly the same
positions in two images, CNNs get a lot better at seeing similarities than whole-image matching
schemes. However, When presented with a new image, the CNN doesn’t know exactly where
these features will match so it tries them everywhere, in every possible position(matches feature
matrices in steps by shifting the defined steps at a time). In calculating the match to a feature
across the whole image, we make it a filter. The math we use to do this is called convolution,
from which Convolutional Neural Networks take their name.
The next step is to repeat the convolution process in its entirety for each of the other features.
The result is a set of filtered images, one for each of our filters. It’s convenient to think of this
whole collection of convolution operations as a single processing step.
Now comes the step where we introduce so-called “non-linearity” in our model so that our model
can predict and learn non-linear boundaries. A very common way to do this is using a non-linear
function (like Relu, gelu). The most popular non-linear function is RELU which performs a
simple math operation: wherever a negative number occurs, swap it out for a 0. This helps the
CNN stay mathematically healthy by keeping learned values from getting stuck near 0 or
blowing up toward infinity. Note that this convolution + Relu operation may create massive
feature maps and it is crucial to reduce the feature map size while keeping the identified feature
intact.
Pooling is a way to take large images and shrink them down while preserving the most important
information in them. It consists of stepping a small window across an image and taking the
maximum value from the window at each step. In practice, a window of 2 or 3 pixels on a side
and steps of 2 pixels work well. A pooling layer is just the operation of performing pooling on an
image or a collection of images. The output will have the same number of images, but they will
each have fewer pixels. This is also helpful in managing the computational load.
Once the desired amount of convolution operations are performed (depending upon the designed
model) it is now time to make use of the power of deep learning neural networks to harness the
full potential of the operations performed in earlier stages. But before we pass the pooled feature
maps to the neural network for learning, we need to flatten the matrices. The reason is very
obvious: neural network only accepts a single dimension input. So we stack them like Lego
bricks. In the end, raw images get filtered, rectified, and pooled to create a set of shrunken,
feature-filtered images and now it is ready to go into the world of neurons (Neural network).
The Fully connected layers in the neural network take the high-level filtered images (1
dimension rectified pooled feature map) and translate them into votes (or signals). These votes
are expressed as weights, or connection strengths, between each value and each category. When
a new image is presented to the CNN, it percolates through the lower layers until it reaches the
fully connected layer at the end. Then an election is held. The answer with the most votes wins
and is declared the category of the input.
And that is how a Deep CNN works. The below figure would summarize what we have talked
about above
DCNN process
DCNN model considerations
(Hyperparameter tuning)
Unfortunately, not every aspect of CNNs can be learned in so straightforward a manner. There is
still a long list of decisions that a CNN designer must make.
 For each convolution layer, How many features? How many pixels in each feature?
 For each pooling layer, What window size? What stride?
 What function should I use? How many epochs? Any early stopping?
 For each extra fully connected layer, How many hidden neurons? and so on...
In addition to these, there are also higher-level architectural decisions to make like how many of
each layer to include? In what order? There are lots of tweaks that we can try, such as new layer
types and more complex ways to connect layers with each other or simply increasing the number
of epochs or changing the activation function.
And the best way to decide is to do and see it for yourself.
Here is a simple notebook where you can see what this might look like and how you can come to
a conclusion for selecting the best CNN hyperparameter combination.
https://colab.re
search.google.com/drive/1gXenThfIViK2v14WJ2D-U9U3hcW5QjC3?usp=sharing
Do note that it may be computationally heavy and hence optimization of your image and batch
size might be essential.

More Related Content

Similar to Deep Neural Network DNN.docx

Visualizing and Understanding Convolutional Networks
Visualizing and Understanding Convolutional NetworksVisualizing and Understanding Convolutional Networks
Visualizing and Understanding Convolutional NetworksWilly Marroquin (WillyDevNET)
 
11_Saloni Malhotra_SummerTraining_PPT.pptx
11_Saloni Malhotra_SummerTraining_PPT.pptx11_Saloni Malhotra_SummerTraining_PPT.pptx
11_Saloni Malhotra_SummerTraining_PPT.pptxSaloniMalhotra23
 
Convolutional Neural Network and Its Applications
Convolutional Neural Network and Its ApplicationsConvolutional Neural Network and Its Applications
Convolutional Neural Network and Its ApplicationsKasun Chinthaka Piyarathna
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Gaurav Mittal
 
improving Profile detection using Deep Learning
improving Profile detection using Deep Learningimproving Profile detection using Deep Learning
improving Profile detection using Deep LearningSahil Kaw
 
BASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptxBASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptxRiteshPandey184067
 
[Revised] Intro to CNN
[Revised] Intro to CNN[Revised] Intro to CNN
[Revised] Intro to CNNVincent Tatan
 
Deep Computer Vision - 1.pptx
Deep Computer Vision - 1.pptxDeep Computer Vision - 1.pptx
Deep Computer Vision - 1.pptxJawadHaider36
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural NetworksIRJET Journal
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural NetworksTayleeGray
 
let's dive to deep learning
let's dive to deep learninglet's dive to deep learning
let's dive to deep learningMohamed Essam
 
Review-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningReview-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningTrong-An Bui
 
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012Florent Renucci
 
Unsupervised Object Detection
Unsupervised Object DetectionUnsupervised Object Detection
Unsupervised Object DetectionMahan Fathi
 
intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxssuser3aa461
 

Similar to Deep Neural Network DNN.docx (20)

Visualizing and Understanding Convolutional Networks
Visualizing and Understanding Convolutional NetworksVisualizing and Understanding Convolutional Networks
Visualizing and Understanding Convolutional Networks
 
11_Saloni Malhotra_SummerTraining_PPT.pptx
11_Saloni Malhotra_SummerTraining_PPT.pptx11_Saloni Malhotra_SummerTraining_PPT.pptx
11_Saloni Malhotra_SummerTraining_PPT.pptx
 
Convolutional Neural Network and Its Applications
Convolutional Neural Network and Its ApplicationsConvolutional Neural Network and Its Applications
Convolutional Neural Network and Its Applications
 
DL.pdf
DL.pdfDL.pdf
DL.pdf
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
improving Profile detection using Deep Learning
improving Profile detection using Deep Learningimproving Profile detection using Deep Learning
improving Profile detection using Deep Learning
 
BASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptxBASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptx
 
Deep Learning
Deep LearningDeep Learning
Deep Learning
 
[Revised] Intro to CNN
[Revised] Intro to CNN[Revised] Intro to CNN
[Revised] Intro to CNN
 
Deep Computer Vision - 1.pptx
Deep Computer Vision - 1.pptxDeep Computer Vision - 1.pptx
Deep Computer Vision - 1.pptx
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural Networks
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
 
CNN.pptx
CNN.pptxCNN.pptx
CNN.pptx
 
doc1.docx
doc1.docxdoc1.docx
doc1.docx
 
let's dive to deep learning
let's dive to deep learninglet's dive to deep learning
let's dive to deep learning
 
Review-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningReview-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learning
 
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012
Manifold Blurring Mean Shift algorithms for manifold denoising, report, 2012
 
Unsupervised Object Detection
Unsupervised Object DetectionUnsupervised Object Detection
Unsupervised Object Detection
 
intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptx
 
Deep learning (2)
Deep learning (2)Deep learning (2)
Deep learning (2)
 

More from jaffarbikat

Deep Learning Vocabulary.docx
Deep Learning Vocabulary.docxDeep Learning Vocabulary.docx
Deep Learning Vocabulary.docxjaffarbikat
 
Chapter 24 till slide 40.pptx
Chapter 24 till slide 40.pptxChapter 24 till slide 40.pptx
Chapter 24 till slide 40.pptxjaffarbikat
 
Chapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxChapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxjaffarbikat
 
Chapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxChapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxjaffarbikat
 
10-Sequences and summation.pptx
10-Sequences and summation.pptx10-Sequences and summation.pptx
10-Sequences and summation.pptxjaffarbikat
 
11-Induction CIIT.pptx
11-Induction CIIT.pptx11-Induction CIIT.pptx
11-Induction CIIT.pptxjaffarbikat
 
9-Functions.pptx
9-Functions.pptx9-Functions.pptx
9-Functions.pptxjaffarbikat
 

More from jaffarbikat (10)

Deep Learning Vocabulary.docx
Deep Learning Vocabulary.docxDeep Learning Vocabulary.docx
Deep Learning Vocabulary.docx
 
Chapter 24 till slide 40.pptx
Chapter 24 till slide 40.pptxChapter 24 till slide 40.pptx
Chapter 24 till slide 40.pptx
 
Chapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxChapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptx
 
Chapter 21.pdf
Chapter 21.pdfChapter 21.pdf
Chapter 21.pdf
 
Chapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptxChapter 21 Couloumb Law 25.pptx
Chapter 21 Couloumb Law 25.pptx
 
10-Sequences and summation.pptx
10-Sequences and summation.pptx10-Sequences and summation.pptx
10-Sequences and summation.pptx
 
11-Induction CIIT.pptx
11-Induction CIIT.pptx11-Induction CIIT.pptx
11-Induction CIIT.pptx
 
9-Functions.pptx
9-Functions.pptx9-Functions.pptx
9-Functions.pptx
 
8-Sets-2.ppt
8-Sets-2.ppt8-Sets-2.ppt
8-Sets-2.ppt
 
7-Sets-1.ppt
7-Sets-1.ppt7-Sets-1.ppt
7-Sets-1.ppt
 

Recently uploaded

3D Printing And Designing Final Report.pdf
3D Printing And Designing Final Report.pdf3D Printing And Designing Final Report.pdf
3D Printing And Designing Final Report.pdfSwaraliBorhade
 
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCRdollysharma2066
 
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档208367051
 
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一diploma 1
 
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)jennyeacort
 
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一F dds
 
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Call Girls Satellite 7397865700 Ridhima Hire Me Full Night
Call Girls Satellite 7397865700 Ridhima Hire Me Full NightCall Girls Satellite 7397865700 Ridhima Hire Me Full Night
Call Girls Satellite 7397865700 Ridhima Hire Me Full Nightssuser7cb4ff
 
FiveHypotheses_UIDMasterclass_18April2024.pdf
FiveHypotheses_UIDMasterclass_18April2024.pdfFiveHypotheses_UIDMasterclass_18April2024.pdf
FiveHypotheses_UIDMasterclass_18April2024.pdfShivakumar Viswanathan
 
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证nhjeo1gg
 
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一F La
 
Passbook project document_april_21__.pdf
Passbook project document_april_21__.pdfPassbook project document_april_21__.pdf
Passbook project document_april_21__.pdfvaibhavkanaujia
 
西北大学毕业证学位证成绩单-怎么样办伪造
西北大学毕业证学位证成绩单-怎么样办伪造西北大学毕业证学位证成绩单-怎么样办伪造
西北大学毕业证学位证成绩单-怎么样办伪造kbdhl05e
 
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...katerynaivanenko1
 
Call Girls Meghani Nagar 7397865700 Independent Call Girls
Call Girls Meghani Nagar 7397865700  Independent Call GirlsCall Girls Meghani Nagar 7397865700  Independent Call Girls
Call Girls Meghani Nagar 7397865700 Independent Call Girlsssuser7cb4ff
 
ARt app | UX Case Study
ARt app | UX Case StudyARt app | UX Case Study
ARt app | UX Case StudySophia Viganò
 
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Service
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts ServiceCall Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Service
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Servicejennyeacort
 
shot list for my tv series two steps back
shot list for my tv series two steps backshot list for my tv series two steps back
shot list for my tv series two steps back17lcow074
 

Recently uploaded (20)

3D Printing And Designing Final Report.pdf
3D Printing And Designing Final Report.pdf3D Printing And Designing Final Report.pdf
3D Printing And Designing Final Report.pdf
 
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR
8377877756 Full Enjoy @24/7 Call Girls in Nirman Vihar Delhi NCR
 
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制堪培拉大学毕业证(UC毕业证)#文凭成绩单#真实留信学历认证永久存档
 
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一
办理(USYD毕业证书)澳洲悉尼大学毕业证成绩单原版一比一
 
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)
Call Us ✡️97111⇛47426⇛Call In girls Vasant Vihar༒(Delhi)
 
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一
办理学位证(SFU证书)西蒙菲莎大学毕业证成绩单原版一比一
 
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Harsh Vihar (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
Call Girls Satellite 7397865700 Ridhima Hire Me Full Night
Call Girls Satellite 7397865700 Ridhima Hire Me Full NightCall Girls Satellite 7397865700 Ridhima Hire Me Full Night
Call Girls Satellite 7397865700 Ridhima Hire Me Full Night
 
FiveHypotheses_UIDMasterclass_18April2024.pdf
FiveHypotheses_UIDMasterclass_18April2024.pdfFiveHypotheses_UIDMasterclass_18April2024.pdf
FiveHypotheses_UIDMasterclass_18April2024.pdf
 
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证
在线办理ohio毕业证俄亥俄大学毕业证成绩单留信学历认证
 
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Okhla Delhi 💯Call Us 🔝8264348440🔝
 
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一
办理(宾州州立毕业证书)美国宾夕法尼亚州立大学毕业证成绩单原版一比一
 
Passbook project document_april_21__.pdf
Passbook project document_april_21__.pdfPassbook project document_april_21__.pdf
Passbook project document_april_21__.pdf
 
Cheap Rate ➥8448380779 ▻Call Girls In Iffco Chowk Gurgaon
Cheap Rate ➥8448380779 ▻Call Girls In Iffco Chowk GurgaonCheap Rate ➥8448380779 ▻Call Girls In Iffco Chowk Gurgaon
Cheap Rate ➥8448380779 ▻Call Girls In Iffco Chowk Gurgaon
 
西北大学毕业证学位证成绩单-怎么样办伪造
西北大学毕业证学位证成绩单-怎么样办伪造西北大学毕业证学位证成绩单-怎么样办伪造
西北大学毕业证学位证成绩单-怎么样办伪造
 
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...
MT. Marseille an Archipelago. Strategies for Integrating Residential Communit...
 
Call Girls Meghani Nagar 7397865700 Independent Call Girls
Call Girls Meghani Nagar 7397865700  Independent Call GirlsCall Girls Meghani Nagar 7397865700  Independent Call Girls
Call Girls Meghani Nagar 7397865700 Independent Call Girls
 
ARt app | UX Case Study
ARt app | UX Case StudyARt app | UX Case Study
ARt app | UX Case Study
 
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Service
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts ServiceCall Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Service
Call Girls in Ashok Nagar Delhi ✡️9711147426✡️ Escorts Service
 
shot list for my tv series two steps back
shot list for my tv series two steps backshot list for my tv series two steps back
shot list for my tv series two steps back
 

Deep Neural Network DNN.docx

  • 1. https://medium.com/aiguys/deep-convolutional-neural-networks-dcnns-explained-in-layman-terms- b990b2818061 Deep Convolutional Neural Networks (DCNNs) explained in layman's terms In deep learning, a convolutional neural network (CNN, or ConvNet) is a class of artificial neural networks, most commonly applied to analyze visual imagery. They are also known as shift invariant or space invariant artificial neural networks (SIANN), based on the shared-weight architecture of the convolution kernels or filters that slide along input features and provide translation equivariant responses known as feature maps. Basic DCNN architecture DCNN Architecture 1. Convolutional Layer + Relu 2. Pooling Layer 3. Fully Connected Layer 4. Dropout 5. Activation Functions
  • 2. 1. Convolutional Layer + Relu This layer is the first layer that is used to extract the various features from the input images. In this layer, the mathematical operation of convolution is performed between the input image and a filter of a particular size MxM. By sliding the filter over the input image, the dot product is taken between the filter and the parts of the input image with respect to the size of the filter (MxM). The output is termed as the Feature map which gives us information about the image such as the corners and edges. Later, this feature map is fed to other layers to learn several other features of the input image. 2. Pooling Layer In most cases, a Convolutional Layer is followed by a Pooling Layer. The primary aim of this layer is to decrease the size of the convolved feature map to reduce computational costs. This is performed by decreasing the connections between layers and independently operating on each feature map. Depending upon the method used, there are several types of Pooling operations. In Max Pooling, the largest element is taken from the feature map. Average Pooling calculates the average of the elements in a predefined sized Image section. The total sum of the elements in the predefined section is computed in Sum Pooling. The Pooling Layer usually serves as a bridge between the Convolutional Layer and the FC Layer 3. Fully Connected Layer The Fully Connected (FC) layer consists of the weights and biases along with the neurons and is used to connect the neurons between two different layers. These layers are usually placed before the output layer and form the last few layers of a CNN Architecture. In this, the input image from the previous layers is flattened and fed to the FC layer. The flattened vector then undergoes a few more FC layers where mathematical operations usually take place. In this stage, the classification process begins to take place. 4. Dropout Usually, when all the features are connected to the FC layer, it can cause overfitting in the training dataset. Overfitting occurs when a particular model works so well on the training data causing a negative impact on the model’s performance when used on new data. To overcome this problem, a dropout layer is utilized wherein a few neurons are dropped from the neural network during the training process resulting in reduced size of the model. On passing a dropout of 0.3, 30% of the nodes are dropped out randomly from the neural network.
  • 3. 5. Activation Functions Finally, one of the most important parameters of the CNN model is the activation function. They are used to learn and approximate any kind of continuous and complex relationship between variables of the network. In simple words, it decides which information of the model should fire in the forward direction and which ones should not at the end of the network. It adds non-linearity to the network. There are several commonly used activation functions such as the ReLU, Softmax, tanH, and the Sigmoid functions. Each of these functions has a specific usage. For a binary classification CNN model, sigmoid and softmax functions are preferred and for multi-class classification, softmax is used. How does a DCNN work? CNN compares images piece by piece. The pieces that it looks for are called features which are nothing but a bunch of MxM matrices with numbers(images are nothing but MxM number matrices of pixel values for a computer). By finding rough feature matches in roughly the same positions in two images, CNNs get a lot better at seeing similarities than whole-image matching schemes. However, When presented with a new image, the CNN doesn’t know exactly where these features will match so it tries them everywhere, in every possible position(matches feature matrices in steps by shifting the defined steps at a time). In calculating the match to a feature across the whole image, we make it a filter. The math we use to do this is called convolution, from which Convolutional Neural Networks take their name. The next step is to repeat the convolution process in its entirety for each of the other features. The result is a set of filtered images, one for each of our filters. It’s convenient to think of this whole collection of convolution operations as a single processing step. Now comes the step where we introduce so-called “non-linearity” in our model so that our model can predict and learn non-linear boundaries. A very common way to do this is using a non-linear function (like Relu, gelu). The most popular non-linear function is RELU which performs a simple math operation: wherever a negative number occurs, swap it out for a 0. This helps the CNN stay mathematically healthy by keeping learned values from getting stuck near 0 or blowing up toward infinity. Note that this convolution + Relu operation may create massive feature maps and it is crucial to reduce the feature map size while keeping the identified feature intact. Pooling is a way to take large images and shrink them down while preserving the most important information in them. It consists of stepping a small window across an image and taking the maximum value from the window at each step. In practice, a window of 2 or 3 pixels on a side and steps of 2 pixels work well. A pooling layer is just the operation of performing pooling on an image or a collection of images. The output will have the same number of images, but they will each have fewer pixels. This is also helpful in managing the computational load.
  • 4. Once the desired amount of convolution operations are performed (depending upon the designed model) it is now time to make use of the power of deep learning neural networks to harness the full potential of the operations performed in earlier stages. But before we pass the pooled feature maps to the neural network for learning, we need to flatten the matrices. The reason is very obvious: neural network only accepts a single dimension input. So we stack them like Lego bricks. In the end, raw images get filtered, rectified, and pooled to create a set of shrunken, feature-filtered images and now it is ready to go into the world of neurons (Neural network). The Fully connected layers in the neural network take the high-level filtered images (1 dimension rectified pooled feature map) and translate them into votes (or signals). These votes are expressed as weights, or connection strengths, between each value and each category. When a new image is presented to the CNN, it percolates through the lower layers until it reaches the fully connected layer at the end. Then an election is held. The answer with the most votes wins and is declared the category of the input. And that is how a Deep CNN works. The below figure would summarize what we have talked about above DCNN process DCNN model considerations (Hyperparameter tuning) Unfortunately, not every aspect of CNNs can be learned in so straightforward a manner. There is still a long list of decisions that a CNN designer must make.  For each convolution layer, How many features? How many pixels in each feature?  For each pooling layer, What window size? What stride?  What function should I use? How many epochs? Any early stopping?  For each extra fully connected layer, How many hidden neurons? and so on...
  • 5. In addition to these, there are also higher-level architectural decisions to make like how many of each layer to include? In what order? There are lots of tweaks that we can try, such as new layer types and more complex ways to connect layers with each other or simply increasing the number of epochs or changing the activation function. And the best way to decide is to do and see it for yourself. Here is a simple notebook where you can see what this might look like and how you can come to a conclusion for selecting the best CNN hyperparameter combination. https://colab.re search.google.com/drive/1gXenThfIViK2v14WJ2D-U9U3hcW5QjC3?usp=sharing Do note that it may be computationally heavy and hence optimization of your image and batch size might be essential.