SlideShare a Scribd company logo
1 of 25
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Prof. Rizwan Ali Naqvi
Department of Intelligent Mechatronics Engineering,
Sejong University, Republic of Korea.
Applied Machine Learning
Spring 2023
Lecture 9:
Convolutional Neural Networks (CNNs)
&
Recurrent Neural Networks (RNNs)
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Agenda
Convolutional Neural Network (CNN)
Basic CNN Components
Architectural Evolution of Deep CNNs
Recurrent Neural Network (RNN)
Summary
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Convolutional Neural Network
• Machine Learning (ML) algorithms are known to learn the underlying relationship in
data and thus make decisions without requiring explicit instructions.
• In 1989, a new class of Neural Networks (NN), called Convolutional Neural Networks
(CNN) was reported, which has shown enormous potential in Machine Vision (MV)
related tasks.
• CNNs are one of the best learning algorithms for understanding image content and have
shown exemplary performance in image segmentation, classification, detection, and
retrieval-related tasks
• The topology of CNN is divided into multiple learning stages composed of a
combination of the convolutional layers, non-linear processing units, and subsampling
layers.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Convolutional Neural Network
• CNN, with the automatic feature extraction ability, reduces the need for a separate
feature extractor.
• Thus, CNN without exhaustive processing can learn good internal representation from
raw pixels.
• Deep Convolutional Neural Networks (CNN) is a special type of Neural Network,
which has shown exemplary performance in several competitions related to Computer
Vision and Image Processing.
• The powerful learning ability of deep CNN is primarily due to the use of multiple
feature extraction stages that can automatically learn representations from the data.
• Several inspiring ideas to bring advancements in CNNs have been explored, such as the
use of different activation and loss functions, parameter optimization, and architectural
innovations.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Convolutional Neural Network
• However, the significant improvement in the representational capacity of the deep CNN
is achieved through architectural innovations.
• CNN first came to the limelight through the work of LeCuN in 1989 for the processing
of grid-like topological data (images and time series data)
• The architectural design of CNN was inspired by Hubel and Wiesel’s work and thus
mostly follows the basic structure of a primate’s visual cortex.
• Nowadays, CNN is considered as one of the most widely used ML techniques,
especially in vision-related applications. CNN can learn representations from grid-like
data, and recently it has shown substantial performance improvement in various ML
applications.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Convolutional Neural Network
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Convolutional Neural Network
• Since CNN possesses both good feature generation and discrimination ability, therefore in a
typical ML system, CNN capabilities are exploited for feature generation and classification.
• A typical CNN architecture generally comprises alternate layers of convolution and pooling
followed by one or more fully connected layers at the end.
• In some cases, a fully connected layer is replaced with a global average pooling layer.
• In addition to different mapping functions, different regulatory units such as batch
normalization and dropout are also incorporated to optimize CNN performance
• The arrangement of CNN components plays a fundamental role in designing new architectures
and thus achieving enhanced performance.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Basic CNN Components
Convolutional Layer
• The convolutional layer is composed of a set of convolutional
kernels where each neuron acts as a kernel.
• Convolutional kernel works by dividing the image into small
slices, commonly known as receptive fields.
• The division of an image into small blocks helps in extracting
feature patterns.
• Kernel convolves with the images using a specific set of weights
by multiplying its elements with the corresponding elements of
the receptive field
• Due to weight sharing ability of convolutional operation, different
sets of features within an image can be extracted by sliding kernel
with the same set of weights on the image
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Basic CNN Components
Pooling Layer
• Pooling or down-sampling is an interesting local
operation.
• It sums up similar information in the neighborhood of
the receptive field and outputs the dominant response
within this local region.
• The use of pooling operation helps to extract a
combination of features, which are invariant to
translational shifts and small distortions.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Basic CNN Components
Activation Function
• Activation function serves as a decision function and
helps in the learning of intricate patterns..
• Activation functions are necessary to prevent linearity.
Without them, the data would pass through the nodes
and layers of the network only going through linear
functions
• The selection of an appropriate activation function can
accelerate the learning process
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Dropout
• Dropout introduces regularization within the
network, which ultimately improves
generalization by randomly skipping some units
or connections
• In NNs, multiple connections that learn a non-
linear relation are sometimes co-adapted, which
causes overfitting.
• This random dropping of some connections or
units produces several thinned network
architectures, and finally, one representative
network is selected with small weights.
• This selected architecture is then considered as
an approximation of all of the proposed network
Basic CNN Components
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Fully Connected Layer
• Fully Connected Layer is simply, feed forward neural
networks.
• Fully Connected Layers form the last few layers in
the network.
• The input to the fully connected layer is the output
from the final Pooling or Convolutional Layer, which
is flattened and then fed into the fully connected
layer.
• The flattened output is fed to a feed-forward neural
network and backpropagation is applied to every
iteration of training.
• Over a series of epochs, the model is able to
distinguish between dominating and certain low-level
features in images and classify them using
the softmax classification technique.
Basic CNN Components
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
• Nowadays, CNNs are considered as the most widely used algorithms among biologically inspired Artificial
Intelligence (AI) techniques.
• CNN history begins with the neurobiological experiments conducted by Hubel and Wiesel.
• Their work provided a platform for many cognitive models, and CNN replaced almost all of these.
• Over the decades, different efforts have been carried out to improve the performance of CNNs.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
Stagnation of CNN: early 2000
• In the late 1990s and early 2000, researchers had little insight into the internal working of
CNN, and it was considered as a black box.
• Complicated architecture design and heavy processing made it hard to train CNN.
• It was widely presumed in early 2000 that the backpropagation algorithm used for training
CNN was not effective in converging to the global minima of the error surface.
• Thus, CNN was considered a less effective feature extractor compared to handcrafted features.
• Moreover, no comprehensive dataset of diverse categories of images was available at that time.
• Therefore, because of the insignificant improvement in CNN performance at the cost of high
computational time, little attention was given to exploring its role in different applications such
as object detection, video surveillance, etc.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
• At that time, other statistical methods and in particular, SVM became more popular than CNN
due to their relatively high performance
• Meanwhile, a few research groups kept on working with CNNs and tried to optimize their
performance.
• In 2003, Simard improved CNN architecture and showed good results compared to SVM on a
hand-digit benchmark dataset.
• This improvement in performance expedited the research in CNNs by extending their
applications beyond optical character recognition to other script’s character recognition,
deployment in image sensors for face detection in video conferencing, and regulation of street
crimes, etc.
• Likewise, CNN-based systems were industrialized in markets for customer tracking.
• Moreover, CNN’s potential in other applications such as medical image segmentation, anomaly
detection, and robot vision was also explored.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
Revival of CNN: 2006–2011
• Deep CNNs generally have complex architecture and time-intensive training phases that
sometimes may span over weeks.
• In early 2000, there were a few parallel processing techniques and limited hardware resources
for the training of deep Networks.
• Since 2006, significant efforts have been made to tackle the CNN optimization problem.
• Hinton reported the concept of greedy layerwise pre-training in 2006, which revived the
research in deep learning.
• Bengio and other researchers proposed that the sigmoid activation function is not suitable for
the training of deep architectures.
• This observation started the use of activation functions other than sigmoid such as ReLU, tanh,
etc.,
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
• The revival of deep learning was one of the factors, which brought deep CNNs into the limelight.
• In late 2006, researchers started using graphics processing units (GPUs) to accelerate the training
of deep NN and CNN architectures.
• In 2007, NVIDIA launched the CUDA programming platform, which allows the exploitation of
parallel processing capabilities of GPU to a greater degree.
• In essence, the use of GPUs for NN and CNN training and other hardware improvements were the
main factors, which revived the research in CNN.
• In 2010, Fei–Fei Li’s group at Stanford, established a large database of images known as
ImageNet, containing millions of annotated images belonging to a large number of classes.
• This database was coupled with the annual ImageNet Large Scale Visual Recognition Challenge
(ILSVRC), where the performances of various models have been evaluated and scored.
• Similarly, in the same year, Stanford released PASCAL 2010 VOC dataset for object detection.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
Rise of CNN: 2012–2014
• The availability of extensive training data and hardware advancements are the factors that
contributed to the advancement in CNN research.
• But the main driving forces that have accelerated the research and give rise to the use of CNNs
in image classification and recognition tasks are parameter optimization strategies and new
architectural ideas.
• The main breakthrough in CNN performance was brought by AlexNet, which showed
exemplary performance in 2012-ILSVRC (reduced error rate from 25.8 to 16.4) as compared to
conventional CV techniques.
• In this era, several attempts were made to improve the performance of CNN; depth and
parameter optimization strategies were explored with a significant reduction in computational
cost.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Architectural Evolution of Deep CNNs
• Similarly, different architectural designs were proposed, whereby each new architecture tried to
overcome the shortcomings of previously proposed architectures in combination with new
structural reformulations.
• With the trend of designing very deep CNNs, it generally becomes difficult to independently
determine filter dimensions, stride, padding, and other hyper-parameters for each layer.
• This problem is resolved by designing convolutional layers with a fixed topology that can be
repeated multiple times.
• This shifted the trend from custom layer design towards modular and uniform layer design.
• It should be noted that in this era, two different types of architectures, deep and narrow, as well
as deep and wide, were in use.
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Rapid Increase in Architectural Innovations and Applications of CNN:
2015-Present
• The research in CNN is still going on and has a significant potential for improvement.
• It is generally observed that significant improvements in CNN performance occurred from
2015 to 2019.
• The representational capacity of a CNN usually depends on its depth.
• Different experimental studies showed that state-of-the-art deep architectures such as VGG,
ResNet, ResNext, etc. also showed good results for challenging recognition and localization
problems like semantic and instance-based object segmentation, etc.
• However, two main concerns observed with deep and wide architectures are the high
computational cost and memory requirement.
Architectural Evolution of Deep CNNs
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
• As a result, it is very challenging to deploy state-of-the-art wide and deep CNN models in
resource-constrained environments.
• Conventional convolution operation requires a huge number of multiplications, which increases
the inference time and restricts the applicability of CNN to low memory and time-constraint
applications.
• Many real-world applications, such as autonomous vehicles, robotics, healthcare, and mobile
applications, perform the tasks that need to be carried out on computationally limited platforms
in a timely manner.
• Therefore, different modifications in CNN are performed to make them appropriate for
resource-constrained environments.
• GoogleNet exploited the idea of small networks, which replaces conventional convolution with
point-wise group convolution operation to make it computationally efficient.
Architectural Evolution of Deep CNNs
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
• Similarly, ShuffleNet used point-wise group convolution but with a new idea of channel
shuffle that significantly reduces the number of operations without affecting the accuracy.
• In the same way, ANTNet proposed a novel architectural block known as ANT Block, which at
low computational cost, achieved good performance on benchmark datasets.
Architectural Evolution of Deep CNNs
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning
Sejong University
Week 9: Deep Learning (Part II) Applied Machine Learning

More Related Content

Similar to Lecture-7 Applied ML.pptx

UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...ijscai
 
Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...IJSCAI Journal
 
improving Profile detection using Deep Learning
improving Profile detection using Deep Learningimproving Profile detection using Deep Learning
improving Profile detection using Deep LearningSahil Kaw
 
MDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A PrimerMDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A PrimerPoo Kuan Hoong
 
Transfer Learning and Fine-tuning Deep Neural Networks
 Transfer Learning and Fine-tuning Deep Neural Networks Transfer Learning and Fine-tuning Deep Neural Networks
Transfer Learning and Fine-tuning Deep Neural NetworksPyData
 
intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxssuser3aa461
 
FINAL_Team_4.pptx
FINAL_Team_4.pptxFINAL_Team_4.pptx
FINAL_Team_4.pptxnitin571047
 
Predicting Steering Angle for Self Driving Vehicles
Predicting Steering Angle for Self Driving VehiclesPredicting Steering Angle for Self Driving Vehicles
Predicting Steering Angle for Self Driving VehiclesIRJET Journal
 
cnn explainer.pptx
cnn explainer.pptxcnn explainer.pptx
cnn explainer.pptxINAM352782
 
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITIONTRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITIONijaia
 
Transfer Learning with Convolutional Neural Networks for IRIS Recognition
Transfer Learning with Convolutional Neural Networks for IRIS RecognitionTransfer Learning with Convolutional Neural Networks for IRIS Recognition
Transfer Learning with Convolutional Neural Networks for IRIS Recognitiongerogepatton
 
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION gerogepatton
 
Convolutional Neural Network.pdf
Convolutional Neural Network.pdfConvolutional Neural Network.pdf
Convolutional Neural Network.pdfAiblogtech
 
Classification of Images Using CNN Model and its Variants
Classification of Images Using CNN Model and its VariantsClassification of Images Using CNN Model and its Variants
Classification of Images Using CNN Model and its VariantsIRJET Journal
 
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkRunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
 
Transformer models for FER
Transformer models for FERTransformer models for FER
Transformer models for FERIRJET Journal
 
Image Segmentation and Classification using Neural Network
Image Segmentation and Classification using Neural NetworkImage Segmentation and Classification using Neural Network
Image Segmentation and Classification using Neural NetworkAIRCC Publishing Corporation
 

Similar to Lecture-7 Applied ML.pptx (20)

UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
 
Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...
 
improving Profile detection using Deep Learning
improving Profile detection using Deep Learningimproving Profile detection using Deep Learning
improving Profile detection using Deep Learning
 
Deep Learning
Deep LearningDeep Learning
Deep Learning
 
MDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A PrimerMDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A Primer
 
Transfer Learning and Fine-tuning Deep Neural Networks
 Transfer Learning and Fine-tuning Deep Neural Networks Transfer Learning and Fine-tuning Deep Neural Networks
Transfer Learning and Fine-tuning Deep Neural Networks
 
intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptx
 
CNN.pptx.pdf
CNN.pptx.pdfCNN.pptx.pdf
CNN.pptx.pdf
 
FINAL_Team_4.pptx
FINAL_Team_4.pptxFINAL_Team_4.pptx
FINAL_Team_4.pptx
 
Predicting Steering Angle for Self Driving Vehicles
Predicting Steering Angle for Self Driving VehiclesPredicting Steering Angle for Self Driving Vehicles
Predicting Steering Angle for Self Driving Vehicles
 
cnn explainer.pptx
cnn explainer.pptxcnn explainer.pptx
cnn explainer.pptx
 
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITIONTRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
 
Transfer Learning with Convolutional Neural Networks for IRIS Recognition
Transfer Learning with Convolutional Neural Networks for IRIS RecognitionTransfer Learning with Convolutional Neural Networks for IRIS Recognition
Transfer Learning with Convolutional Neural Networks for IRIS Recognition
 
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION
 
Convolutional Neural Network.pdf
Convolutional Neural Network.pdfConvolutional Neural Network.pdf
Convolutional Neural Network.pdf
 
Classification of Images Using CNN Model and its Variants
Classification of Images Using CNN Model and its VariantsClassification of Images Using CNN Model and its Variants
Classification of Images Using CNN Model and its Variants
 
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkRunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
 
Transformer models for FER
Transformer models for FERTransformer models for FER
Transformer models for FER
 
Dl
DlDl
Dl
 
Image Segmentation and Classification using Neural Network
Image Segmentation and Classification using Neural NetworkImage Segmentation and Classification using Neural Network
Image Segmentation and Classification using Neural Network
 

Recently uploaded

WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...WSO2
 
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyviewmasabamasaba
 
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...Shane Coughlan
 
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...WSO2
 
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...masabamasaba
 
AI & Machine Learning Presentation Template
AI & Machine Learning Presentation TemplateAI & Machine Learning Presentation Template
AI & Machine Learning Presentation TemplatePresentation.STUDIO
 
%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in soweto%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in sowetomasabamasaba
 
%in ivory park+277-882-255-28 abortion pills for sale in ivory park
%in ivory park+277-882-255-28 abortion pills for sale in ivory park %in ivory park+277-882-255-28 abortion pills for sale in ivory park
%in ivory park+277-882-255-28 abortion pills for sale in ivory park masabamasaba
 
Direct Style Effect Systems - The Print[A] Example - A Comprehension Aid
Direct Style Effect Systems -The Print[A] Example- A Comprehension AidDirect Style Effect Systems -The Print[A] Example- A Comprehension Aid
Direct Style Effect Systems - The Print[A] Example - A Comprehension AidPhilip Schwarz
 
WSO2CON 2024 - How to Run a Security Program
WSO2CON 2024 - How to Run a Security ProgramWSO2CON 2024 - How to Run a Security Program
WSO2CON 2024 - How to Run a Security ProgramWSO2
 
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024VictoriaMetrics
 
Announcing Codolex 2.0 from GDK Software
Announcing Codolex 2.0 from GDK SoftwareAnnouncing Codolex 2.0 from GDK Software
Announcing Codolex 2.0 from GDK SoftwareJim McKeeth
 
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...Health
 
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...WSO2
 
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...WSO2
 
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...masabamasaba
 
WSO2CON 2024 Slides - Open Source to SaaS
WSO2CON 2024 Slides - Open Source to SaaSWSO2CON 2024 Slides - Open Source to SaaS
WSO2CON 2024 Slides - Open Source to SaaSWSO2
 
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfonteinmasabamasaba
 
tonesoftg
tonesoftgtonesoftg
tonesoftglanshi9
 

Recently uploaded (20)

WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
 
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview
%in Hazyview+277-882-255-28 abortion pills for sale in Hazyview
 
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...
OpenChain - The Ramifications of ISO/IEC 5230 and ISO/IEC 18974 for Legal Pro...
 
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...
WSO2CON 2024 - Navigating API Complexity: REST, GraphQL, gRPC, Websocket, Web...
 
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...
%+27788225528 love spells in Colorado Springs Psychic Readings, Attraction sp...
 
AI & Machine Learning Presentation Template
AI & Machine Learning Presentation TemplateAI & Machine Learning Presentation Template
AI & Machine Learning Presentation Template
 
%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in soweto%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in soweto
 
%in ivory park+277-882-255-28 abortion pills for sale in ivory park
%in ivory park+277-882-255-28 abortion pills for sale in ivory park %in ivory park+277-882-255-28 abortion pills for sale in ivory park
%in ivory park+277-882-255-28 abortion pills for sale in ivory park
 
Direct Style Effect Systems - The Print[A] Example - A Comprehension Aid
Direct Style Effect Systems -The Print[A] Example- A Comprehension AidDirect Style Effect Systems -The Print[A] Example- A Comprehension Aid
Direct Style Effect Systems - The Print[A] Example - A Comprehension Aid
 
WSO2CON 2024 - How to Run a Security Program
WSO2CON 2024 - How to Run a Security ProgramWSO2CON 2024 - How to Run a Security Program
WSO2CON 2024 - How to Run a Security Program
 
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
 
Abortion Pills In Pretoria ](+27832195400*)[ 🏥 Women's Abortion Clinic In Pre...
Abortion Pills In Pretoria ](+27832195400*)[ 🏥 Women's Abortion Clinic In Pre...Abortion Pills In Pretoria ](+27832195400*)[ 🏥 Women's Abortion Clinic In Pre...
Abortion Pills In Pretoria ](+27832195400*)[ 🏥 Women's Abortion Clinic In Pre...
 
Announcing Codolex 2.0 from GDK Software
Announcing Codolex 2.0 from GDK SoftwareAnnouncing Codolex 2.0 from GDK Software
Announcing Codolex 2.0 from GDK Software
 
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...
+971565801893>>SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHAB...
 
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...
WSO2CON 2024 - Building the API First Enterprise – Running an API Program, fr...
 
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...
WSO2CON 2024 - WSO2's Digital Transformation Journey with Choreo: A Platforml...
 
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...
%+27788225528 love spells in Toronto Psychic Readings, Attraction spells,Brin...
 
WSO2CON 2024 Slides - Open Source to SaaS
WSO2CON 2024 Slides - Open Source to SaaSWSO2CON 2024 Slides - Open Source to SaaS
WSO2CON 2024 Slides - Open Source to SaaS
 
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
 
tonesoftg
tonesoftgtonesoftg
tonesoftg
 

Lecture-7 Applied ML.pptx

  • 1. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Prof. Rizwan Ali Naqvi Department of Intelligent Mechatronics Engineering, Sejong University, Republic of Korea. Applied Machine Learning Spring 2023 Lecture 9: Convolutional Neural Networks (CNNs) & Recurrent Neural Networks (RNNs)
  • 2. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Agenda Convolutional Neural Network (CNN) Basic CNN Components Architectural Evolution of Deep CNNs Recurrent Neural Network (RNN) Summary
  • 3. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Convolutional Neural Network • Machine Learning (ML) algorithms are known to learn the underlying relationship in data and thus make decisions without requiring explicit instructions. • In 1989, a new class of Neural Networks (NN), called Convolutional Neural Networks (CNN) was reported, which has shown enormous potential in Machine Vision (MV) related tasks. • CNNs are one of the best learning algorithms for understanding image content and have shown exemplary performance in image segmentation, classification, detection, and retrieval-related tasks • The topology of CNN is divided into multiple learning stages composed of a combination of the convolutional layers, non-linear processing units, and subsampling layers.
  • 4. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Convolutional Neural Network • CNN, with the automatic feature extraction ability, reduces the need for a separate feature extractor. • Thus, CNN without exhaustive processing can learn good internal representation from raw pixels. • Deep Convolutional Neural Networks (CNN) is a special type of Neural Network, which has shown exemplary performance in several competitions related to Computer Vision and Image Processing. • The powerful learning ability of deep CNN is primarily due to the use of multiple feature extraction stages that can automatically learn representations from the data. • Several inspiring ideas to bring advancements in CNNs have been explored, such as the use of different activation and loss functions, parameter optimization, and architectural innovations.
  • 5. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Convolutional Neural Network • However, the significant improvement in the representational capacity of the deep CNN is achieved through architectural innovations. • CNN first came to the limelight through the work of LeCuN in 1989 for the processing of grid-like topological data (images and time series data) • The architectural design of CNN was inspired by Hubel and Wiesel’s work and thus mostly follows the basic structure of a primate’s visual cortex. • Nowadays, CNN is considered as one of the most widely used ML techniques, especially in vision-related applications. CNN can learn representations from grid-like data, and recently it has shown substantial performance improvement in various ML applications.
  • 6. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Convolutional Neural Network
  • 7. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Convolutional Neural Network • Since CNN possesses both good feature generation and discrimination ability, therefore in a typical ML system, CNN capabilities are exploited for feature generation and classification. • A typical CNN architecture generally comprises alternate layers of convolution and pooling followed by one or more fully connected layers at the end. • In some cases, a fully connected layer is replaced with a global average pooling layer. • In addition to different mapping functions, different regulatory units such as batch normalization and dropout are also incorporated to optimize CNN performance • The arrangement of CNN components plays a fundamental role in designing new architectures and thus achieving enhanced performance.
  • 8. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Basic CNN Components Convolutional Layer • The convolutional layer is composed of a set of convolutional kernels where each neuron acts as a kernel. • Convolutional kernel works by dividing the image into small slices, commonly known as receptive fields. • The division of an image into small blocks helps in extracting feature patterns. • Kernel convolves with the images using a specific set of weights by multiplying its elements with the corresponding elements of the receptive field • Due to weight sharing ability of convolutional operation, different sets of features within an image can be extracted by sliding kernel with the same set of weights on the image
  • 9. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning
  • 10. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Basic CNN Components Pooling Layer • Pooling or down-sampling is an interesting local operation. • It sums up similar information in the neighborhood of the receptive field and outputs the dominant response within this local region. • The use of pooling operation helps to extract a combination of features, which are invariant to translational shifts and small distortions.
  • 11. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Basic CNN Components Activation Function • Activation function serves as a decision function and helps in the learning of intricate patterns.. • Activation functions are necessary to prevent linearity. Without them, the data would pass through the nodes and layers of the network only going through linear functions • The selection of an appropriate activation function can accelerate the learning process
  • 12. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Dropout • Dropout introduces regularization within the network, which ultimately improves generalization by randomly skipping some units or connections • In NNs, multiple connections that learn a non- linear relation are sometimes co-adapted, which causes overfitting. • This random dropping of some connections or units produces several thinned network architectures, and finally, one representative network is selected with small weights. • This selected architecture is then considered as an approximation of all of the proposed network Basic CNN Components
  • 13. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Fully Connected Layer • Fully Connected Layer is simply, feed forward neural networks. • Fully Connected Layers form the last few layers in the network. • The input to the fully connected layer is the output from the final Pooling or Convolutional Layer, which is flattened and then fed into the fully connected layer. • The flattened output is fed to a feed-forward neural network and backpropagation is applied to every iteration of training. • Over a series of epochs, the model is able to distinguish between dominating and certain low-level features in images and classify them using the softmax classification technique. Basic CNN Components
  • 14. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs • Nowadays, CNNs are considered as the most widely used algorithms among biologically inspired Artificial Intelligence (AI) techniques. • CNN history begins with the neurobiological experiments conducted by Hubel and Wiesel. • Their work provided a platform for many cognitive models, and CNN replaced almost all of these. • Over the decades, different efforts have been carried out to improve the performance of CNNs.
  • 15. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs Stagnation of CNN: early 2000 • In the late 1990s and early 2000, researchers had little insight into the internal working of CNN, and it was considered as a black box. • Complicated architecture design and heavy processing made it hard to train CNN. • It was widely presumed in early 2000 that the backpropagation algorithm used for training CNN was not effective in converging to the global minima of the error surface. • Thus, CNN was considered a less effective feature extractor compared to handcrafted features. • Moreover, no comprehensive dataset of diverse categories of images was available at that time. • Therefore, because of the insignificant improvement in CNN performance at the cost of high computational time, little attention was given to exploring its role in different applications such as object detection, video surveillance, etc.
  • 16. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs • At that time, other statistical methods and in particular, SVM became more popular than CNN due to their relatively high performance • Meanwhile, a few research groups kept on working with CNNs and tried to optimize their performance. • In 2003, Simard improved CNN architecture and showed good results compared to SVM on a hand-digit benchmark dataset. • This improvement in performance expedited the research in CNNs by extending their applications beyond optical character recognition to other script’s character recognition, deployment in image sensors for face detection in video conferencing, and regulation of street crimes, etc. • Likewise, CNN-based systems were industrialized in markets for customer tracking. • Moreover, CNN’s potential in other applications such as medical image segmentation, anomaly detection, and robot vision was also explored.
  • 17. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs Revival of CNN: 2006–2011 • Deep CNNs generally have complex architecture and time-intensive training phases that sometimes may span over weeks. • In early 2000, there were a few parallel processing techniques and limited hardware resources for the training of deep Networks. • Since 2006, significant efforts have been made to tackle the CNN optimization problem. • Hinton reported the concept of greedy layerwise pre-training in 2006, which revived the research in deep learning. • Bengio and other researchers proposed that the sigmoid activation function is not suitable for the training of deep architectures. • This observation started the use of activation functions other than sigmoid such as ReLU, tanh, etc.,
  • 18. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs • The revival of deep learning was one of the factors, which brought deep CNNs into the limelight. • In late 2006, researchers started using graphics processing units (GPUs) to accelerate the training of deep NN and CNN architectures. • In 2007, NVIDIA launched the CUDA programming platform, which allows the exploitation of parallel processing capabilities of GPU to a greater degree. • In essence, the use of GPUs for NN and CNN training and other hardware improvements were the main factors, which revived the research in CNN. • In 2010, Fei–Fei Li’s group at Stanford, established a large database of images known as ImageNet, containing millions of annotated images belonging to a large number of classes. • This database was coupled with the annual ImageNet Large Scale Visual Recognition Challenge (ILSVRC), where the performances of various models have been evaluated and scored. • Similarly, in the same year, Stanford released PASCAL 2010 VOC dataset for object detection.
  • 19. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs Rise of CNN: 2012–2014 • The availability of extensive training data and hardware advancements are the factors that contributed to the advancement in CNN research. • But the main driving forces that have accelerated the research and give rise to the use of CNNs in image classification and recognition tasks are parameter optimization strategies and new architectural ideas. • The main breakthrough in CNN performance was brought by AlexNet, which showed exemplary performance in 2012-ILSVRC (reduced error rate from 25.8 to 16.4) as compared to conventional CV techniques. • In this era, several attempts were made to improve the performance of CNN; depth and parameter optimization strategies were explored with a significant reduction in computational cost.
  • 20. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Architectural Evolution of Deep CNNs • Similarly, different architectural designs were proposed, whereby each new architecture tried to overcome the shortcomings of previously proposed architectures in combination with new structural reformulations. • With the trend of designing very deep CNNs, it generally becomes difficult to independently determine filter dimensions, stride, padding, and other hyper-parameters for each layer. • This problem is resolved by designing convolutional layers with a fixed topology that can be repeated multiple times. • This shifted the trend from custom layer design towards modular and uniform layer design. • It should be noted that in this era, two different types of architectures, deep and narrow, as well as deep and wide, were in use.
  • 21. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning Rapid Increase in Architectural Innovations and Applications of CNN: 2015-Present • The research in CNN is still going on and has a significant potential for improvement. • It is generally observed that significant improvements in CNN performance occurred from 2015 to 2019. • The representational capacity of a CNN usually depends on its depth. • Different experimental studies showed that state-of-the-art deep architectures such as VGG, ResNet, ResNext, etc. also showed good results for challenging recognition and localization problems like semantic and instance-based object segmentation, etc. • However, two main concerns observed with deep and wide architectures are the high computational cost and memory requirement. Architectural Evolution of Deep CNNs
  • 22. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning • As a result, it is very challenging to deploy state-of-the-art wide and deep CNN models in resource-constrained environments. • Conventional convolution operation requires a huge number of multiplications, which increases the inference time and restricts the applicability of CNN to low memory and time-constraint applications. • Many real-world applications, such as autonomous vehicles, robotics, healthcare, and mobile applications, perform the tasks that need to be carried out on computationally limited platforms in a timely manner. • Therefore, different modifications in CNN are performed to make them appropriate for resource-constrained environments. • GoogleNet exploited the idea of small networks, which replaces conventional convolution with point-wise group convolution operation to make it computationally efficient. Architectural Evolution of Deep CNNs
  • 23. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning • Similarly, ShuffleNet used point-wise group convolution but with a new idea of channel shuffle that significantly reduces the number of operations without affecting the accuracy. • In the same way, ANTNet proposed a novel architectural block known as ANT Block, which at low computational cost, achieved good performance on benchmark datasets. Architectural Evolution of Deep CNNs
  • 24. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning
  • 25. Sejong University Week 9: Deep Learning (Part II) Applied Machine Learning

Editor's Notes

  1. The visual cortex is the primary cortical region of the brain that receives, integrates, and processes visual information relayed from the retinas.
  2. Translation shifts are the changes that occur when a text is translated into another language due to either the differences between the two languages or the translator's choices.
  3. Cognitive modeling is an area of computer science that deals with simulating human problem-solving and mental processing in a computerized model. Such a model can be used to simulate or predict human behavior or performance on tasks like the ones modeled and improve human-computer interaction.
  4. Greedy algorithms break a problem into many components, then solve for the optimal version of each component in isolation.
  5. Stride is the number of pixels shifts over the input matrix. When the stride is 1 then we move the filters to 1 pixel at a time. When the stride is 2 then we move the filters to 2 pixels at a time and so on To maintain the dimension of output as in input , we use padding. Padding is a process of adding zeros to the input matrix symmetrically. In the following example,the extra grey blocks denote the padding. It is used to make the dimension of output same as input. S
  6. Suppose, you have an input image of a street view consisting of several people, cars, buildings etc. If you only want to group objects belonging to the same category, say distinguish all cars from all buildings, it is the task of semantic segmentation. Within each category say, people, if you want to distinguish each individual person, that will be the task of instance segmentation