SlideShare a Scribd company logo
Machine Learning and its Applications
Presented by
Ankita Tiwari
Definition by Tom Mitchell(1998):
Machine Learning is the study of algorithms that
• improve their performance P
• at some task T
• with experience E
A well-defined learning task is given by <P, T, E>.
What is Machine Learning?
“Learning is a process by which a system improves
its performance from experience.”
- Herbert Simon
When Do We Use Machine Learning
ML is used when:
• Human expertise does not exist (navigation on mars )
• Human can't explain their expertise (speech recognition)
• Models must be customized (personalized medicine)
• Model are based on huge amounts of data (genomics)
Learning isn't always useful:
• There is no need to "learn" to calculate payroll
1. What exactly is deep learning ?
2. Why is it generally better than other methods on
image, speech and certain other types of data?
The short answers
• ‘Deep Learning’ means using a neural network with
several layers of nodes between input and output,
• The combination of layers between input & output
perform feature identification and processing in
stages, just as our brains seem to.
Introduction
Types of Learning
Unsupervised Learning
Unsupervised learning is the training of a
machine using information that is neither
classified nor labeled and allowing the
algorithm to act on that information
without guidance. Here the task of the
machine is to group unsorted information
according to similarities, patterns, and
differences without any prior training of
data. Unlike supervised learning, no
teacher is provided that means no
training will be given to the machine.
Therefore the machine is restricted to
find the hidden structure in unlabeled
data by itself.
Supervised Learning
It has the presence of a supervisor as a
teacher. Basically supervised learning is
when we teach or train the machine using
data that is well labeled. Which means
some data is already tagged with the
correct answer. After that, the machine is
provided with a new set of
examples(data) so that the supervised
learning algorithm analyses the training
data(set of training examples) and
produces a correct outcome from labeled
data.
Unsupervised Learning
Unsupervised learning is classified into two categories of algorithms:
• Clustering: A clustering problem is where you want to discover the inherent groupings in the data, such
as grouping customers by purchasing behavior.
• Association: An association rule learning problem is where you want to discover rules that describe
large portions of your data, such as people that buy X also tend to buy Y.
Types of Unsupervised Learning:-
Clustering
• Exclusive (partitioning)
• Agglomerative
• Overlapping
• Probabilistic
Clustering Types:-
• Hierarchical clustering
• K-means clustering
• Principal Component Analysis
• Singular Value Decomposition
• Independent Component Analysis
Supervised Learning
Supervised learning is classified into two categories of algorithms:
• Classification: A classification problem is when the output variable is a category, such
as “Red” or “blue” or “disease” and “no disease”.
• Regression: A regression problem is when the output variable is a real value, such as
“dollars” or “weight”.
Supervised learning deals with or learns with “labeled” data. This implies that some
data is already tagged with the correct answer.
Types:-
• Regression
• Logistic Regression
• Classification
• Naive Bayes Classifiers
• K-NN (k nearest neighbors)
• Decision Trees
• Support Vector Machine
Artificial Neural Network (ANN)
• It is a type of supervised learning.
• An ANN is an information processing paradigm that
is inspired by the brain.
• It learns by examples. It is configured for a specific
application, such as pattern recognition or data
classification, through a learning process.
• Learning largely involves adjustments to the
synaptic connections that exist between the
neurons.
Analogy Between Biological and ANN
Biological Neuron Artificial Neuron
W1
W2
W3
f(x)
1.4
-2.5
-0.06
Working of Perceptron
Sigmoid function
2.7
-8.6
0.002
f(x)
1.4
-2.5
-0.06
x = -0.06×2.7 + 2.5×8.6 + 1.4×0.002 = 21.34
Working of Perceptron
A dataset
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Training and weights update
Training the neural network
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Initialise with random weights
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Present a training pattern
1.4
2.7
1.9
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Feed it through to get output
1.4
2.7 0.8
1.9
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Compare with target output
1.4
2.7 0.8
0
1.9 error 0.8
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Adjust weights based on error
1.4
2.7 0.8
0
1.9 error 0.8
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Present a training pattern
6.4
2.8
1.7
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Feed it through to get output
6.4
2.8 0.9
1.7
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Compare with target output
6.4
2.8 0.9
1
1.7 error -0.1
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
Adjust weights based on error
6.4
2.8 0.9
1
1.7 error -0.1
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
And so on...
6.4
2.8 0.9
1
1.7 error -0.1
Training and weights update
Training data
Fields class
1.4 2.7 1.9 0
3.8 3.4 3.2 0
6.4 2.8 1.7 1
4.1 0.1 0.2 0
etc …
And so on ….
6.4
2.8 0.9
1
1.7 error -0.1
• Repeat the process number of times, maybe millions of
times each time taking a random training instance, and
making slight weight adjustments
• Algorithms for weight adjustment are designed to make
changes that will reduce the error
Training and weights update
The decision boundary perspective
Initial random weights
The decision boundary perspective
Present a training instance / adjust the weights
The decision boundary perspective
Present a training instance / adjust the weights
The decision boundary perspective
Present a training instance / adjust the weights
The decision boundary perspective
Present a training instance / adjust the weights
Eventually ….
The decision boundary perspective
Summary
• In general, weight learning algorithms of ANNs are like GIGO.
• They work by making thousands and thousands of tiny adjustments,
each making the network do better at the most recent pattern, but
perhaps a little worse on many others
• But, by luck, eventually this tends to be good enough to learn
effective classifiers for many real life applications
Feature detectors
Hand-written data set
what is this
unit doing?
Hand-written data set
Hidden layer makes
connection of neuron
strong or weak, based
on inputs
…
1
63
1 5 10 15 20 25 …
strong +ve weight
low/zero weight
Processing of hidden layer
Processing of hidden layer
…
1
63
1 5 10 15 20 25 …
strong +ve weight
low/zero weight
it will send strong signal for a horizontal
line in the top row, ignoring everywhere else
…
1
63
1 5 10 15 20 25 …
strong +ve weight
low/zero weight
Processing of hidden layer
…
1
63
1 5 10 15 20 25 …
strong +ve weight
low/zero weight
Strong signal for a dark area in the top left
corner
Processing of hidden layer
So: multiple layers make sense
So: multiple layers make sense
Your brain works that way
So: multiple layers make sense
Multi-layer neural network architectures can learn the true underlying features and
‘feature logic’, and therefore generalise very well.
But, until very recently, our weight-learning
algorithms simply did not work on multi-layer
architectures
There are new methods are proposed to
train multi-layer NN
Train this layer first
The new way to train multi-layer NN
Train this layer first
then this layer
The new way to train multi-layer NN
Train this layer first
then this layer
then this layer
The new way to train multi-layer NN
Train this layer first
then this layer
then this layer
then this layer
The new way to train multi-layer NN
Train this layer first
then this layer
then this layer
then this layer
finally this layer
The new way to train multi-layer NN
EACH of the (non-output) layers is
trained to be an auto-encoder
Basically, it is forced to learn good
features that describe what comes from
the previous layer
The new way to train multi-layer NN
Conclusion
• There are many types of deep learning models,
• Different kinds of autoencoder, variations on
architectures and training algorithms, etc.
• It is a very fast-growing area.
Applications
Autonomous Cars
Autonomous Car
Sensors
Autonomous Car
Technology
Deep Learning in the Headlines
Scene Labeling via Deep Learning
Machine Learning in
Automatic Speech Recognition
A Typical Speech Recognition System
ML used to predict of phone states from the sound spectrogram
Deep learning has state-of-the-art results
Zeiler et al. “On rectified linear units for speech
recognition” ICASSP 2013
Impact of Deep Learning in Speech Technology
Continuous Observation
Continuous transfer
of data for
observation and
monitoring
Delay,
Power,
Area
Device that analysis
and compute at same
level like human
brain
Requirement of
Artificial
Intelligent
device
ANN CNN
SNN
ANN / CNN
1) Large computation
2) Power hungry
3) Large area
Alternative
SNN:
1) High parallel
computation
2) Power efficient
3) Event-driven
4) More brain inspired
Most remarkable neural
network
21-02-2023
Challenges
Proposed solutions
21-02-2023
• Reduce Data Movement
• Maximize Data Reuse within PE
• Maximize Data Reuse with Buffer
• Multicast Network Design
• Exploit Data Statistics
• Data Compression
• Data gating/ zero skipping
• Operations exhibit High Parallelism
2/21/2023
Proposed work
Thank you
2/21/2023

More Related Content

Similar to EssentialsOfMachineLearning.pdf

CSA 3702 machine learning module 1
CSA 3702 machine learning module 1CSA 3702 machine learning module 1
CSA 3702 machine learning module 1
Nandhini S
 

Similar to EssentialsOfMachineLearning.pdf (20)

CSA 3702 machine learning module 1
CSA 3702 machine learning module 1CSA 3702 machine learning module 1
CSA 3702 machine learning module 1
 
Machine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionMachine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis Introduction
 
Getting started with Machine Learning
Getting started with Machine LearningGetting started with Machine Learning
Getting started with Machine Learning
 
Lecture 1
Lecture 1Lecture 1
Lecture 1
 
lec1.ppt
lec1.pptlec1.ppt
lec1.ppt
 
in5490-classification (1).pptx
in5490-classification (1).pptxin5490-classification (1).pptx
in5490-classification (1).pptx
 
Artificial Neural Networks for data mining
Artificial Neural Networks for data miningArtificial Neural Networks for data mining
Artificial Neural Networks for data mining
 
Artificial Neural Networks for Data Mining
Artificial Neural Networks for Data MiningArtificial Neural Networks for Data Mining
Artificial Neural Networks for Data Mining
 
ML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptxML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptx
 
Machine_Learning_Co__
Machine_Learning_Co__Machine_Learning_Co__
Machine_Learning_Co__
 
ANN - UNIT 3.pptx
ANN - UNIT 3.pptxANN - UNIT 3.pptx
ANN - UNIT 3.pptx
 
ANN - UNIT 3.pptx
ANN - UNIT 3.pptxANN - UNIT 3.pptx
ANN - UNIT 3.pptx
 
Machine Learning Essentials Demystified part2 | Big Data Demystified
Machine Learning Essentials Demystified part2 | Big Data DemystifiedMachine Learning Essentials Demystified part2 | Big Data Demystified
Machine Learning Essentials Demystified part2 | Big Data Demystified
 
part3Module 3 ppt_with classification.pptx
part3Module 3 ppt_with classification.pptxpart3Module 3 ppt_with classification.pptx
part3Module 3 ppt_with classification.pptx
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Visualization of Deep Learning
Visualization of Deep LearningVisualization of Deep Learning
Visualization of Deep Learning
 
Introduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner'sIntroduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner's
 
Deep learning
Deep learningDeep learning
Deep learning
 
IRJET - A Survey on Machine Learning Algorithms, Techniques and Applications
IRJET - A Survey on Machine Learning Algorithms, Techniques and ApplicationsIRJET - A Survey on Machine Learning Algorithms, Techniques and Applications
IRJET - A Survey on Machine Learning Algorithms, Techniques and Applications
 
Lecture 5 machine learning updated
Lecture 5   machine learning updatedLecture 5   machine learning updated
Lecture 5 machine learning updated
 

More from Ankita Tiwari

More from Ankita Tiwari (20)

surveyofdnnlearning.pdf
surveyofdnnlearning.pdfsurveyofdnnlearning.pdf
surveyofdnnlearning.pdf
 
Basic_Digital_Circuits_Implementation_using_Virtuoso.pdf
Basic_Digital_Circuits_Implementation_using_Virtuoso.pdfBasic_Digital_Circuits_Implementation_using_Virtuoso.pdf
Basic_Digital_Circuits_Implementation_using_Virtuoso.pdf
 
Relation of Big Data and E-Commerce
Relation of Big Data and E-CommerceRelation of Big Data and E-Commerce
Relation of Big Data and E-Commerce
 
Study of various Data Compression Techniques used in Lossless Compression of ...
Study of various Data Compression Techniques used in Lossless Compression of ...Study of various Data Compression Techniques used in Lossless Compression of ...
Study of various Data Compression Techniques used in Lossless Compression of ...
 
PIC Introduction and explained in detailed
PIC Introduction and explained in detailedPIC Introduction and explained in detailed
PIC Introduction and explained in detailed
 
What is IEEE and why?
What is IEEE and why?What is IEEE and why?
What is IEEE and why?
 
To implement Water level control using LabVIEW and analog input signals from ...
To implement Water level control using LabVIEW and analog input signals from ...To implement Water level control using LabVIEW and analog input signals from ...
To implement Water level control using LabVIEW and analog input signals from ...
 
To count number of external events using LabVIEW
To count number of external events using LabVIEWTo count number of external events using LabVIEW
To count number of external events using LabVIEW
 
To control the dc motor speed using PWM from LabVIEW
To control the dc motor speed using PWM from LabVIEWTo control the dc motor speed using PWM from LabVIEW
To control the dc motor speed using PWM from LabVIEW
 
To measure the intensity of light using LDR sensor by calibrating voltage wit...
To measure the intensity of light using LDR sensor by calibrating voltage wit...To measure the intensity of light using LDR sensor by calibrating voltage wit...
To measure the intensity of light using LDR sensor by calibrating voltage wit...
 
To interface temperature sensor with microcontroller and perform closed loop ...
To interface temperature sensor with microcontroller and perform closed loop ...To interface temperature sensor with microcontroller and perform closed loop ...
To interface temperature sensor with microcontroller and perform closed loop ...
 
Interface stepper motor through Arduino using LABVIEW.
Interface stepper motor through Arduino using LABVIEW.Interface stepper motor through Arduino using LABVIEW.
Interface stepper motor through Arduino using LABVIEW.
 
To study the relay operation from digital control signal using LabVIEW.
To study the relay operation from digital control signal using LabVIEW.To study the relay operation from digital control signal using LabVIEW.
To study the relay operation from digital control signal using LabVIEW.
 
Linux operating systems and Bootable Pendrive
Linux operating systems and Bootable PendriveLinux operating systems and Bootable Pendrive
Linux operating systems and Bootable Pendrive
 
Design the implementation of Robotic Simulator: Goalkeeper.
Design the implementation of Robotic Simulator: Goalkeeper.Design the implementation of Robotic Simulator: Goalkeeper.
Design the implementation of Robotic Simulator: Goalkeeper.
 
Design the implementation of Forward Dynamic for PUMA560.
Design the implementation of Forward Dynamic for PUMA560.Design the implementation of Forward Dynamic for PUMA560.
Design the implementation of Forward Dynamic for PUMA560.
 
Design the implementation of Anytime D Star on an Occupancy Grid
Design the implementation of Anytime D Star on an Occupancy GridDesign the implementation of Anytime D Star on an Occupancy Grid
Design the implementation of Anytime D Star on an Occupancy Grid
 
Design the implementation of CDEx Robust DC Motor.
Design the implementation of CDEx Robust DC Motor.Design the implementation of CDEx Robust DC Motor.
Design the implementation of CDEx Robust DC Motor.
 
Design the implementation of CDEx PID with Constraints
Design the implementation of CDEx PID with ConstraintsDesign the implementation of CDEx PID with Constraints
Design the implementation of CDEx PID with Constraints
 
Design the implementation of CDEx flexible joints
Design the implementation of CDEx flexible jointsDesign the implementation of CDEx flexible joints
Design the implementation of CDEx flexible joints
 

Recently uploaded

Hall booking system project report .pdf
Hall booking system project report  .pdfHall booking system project report  .pdf
Hall booking system project report .pdf
Kamal Acharya
 
Digital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdfDigital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdf
AbrahamGadissa
 

Recently uploaded (20)

Hall booking system project report .pdf
Hall booking system project report  .pdfHall booking system project report  .pdf
Hall booking system project report .pdf
 
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptxCloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
 
Democratizing Fuzzing at Scale by Abhishek Arya
Democratizing Fuzzing at Scale by Abhishek AryaDemocratizing Fuzzing at Scale by Abhishek Arya
Democratizing Fuzzing at Scale by Abhishek Arya
 
Natalia Rutkowska - BIM School Course in Kraków
Natalia Rutkowska - BIM School Course in KrakówNatalia Rutkowska - BIM School Course in Kraków
Natalia Rutkowska - BIM School Course in Kraków
 
Water Industry Process Automation and Control Monthly - May 2024.pdf
Water Industry Process Automation and Control Monthly - May 2024.pdfWater Industry Process Automation and Control Monthly - May 2024.pdf
Water Industry Process Automation and Control Monthly - May 2024.pdf
 
Online resume builder management system project report.pdf
Online resume builder management system project report.pdfOnline resume builder management system project report.pdf
Online resume builder management system project report.pdf
 
fluid mechanics gate notes . gate all pyqs answer
fluid mechanics gate notes . gate all pyqs answerfluid mechanics gate notes . gate all pyqs answer
fluid mechanics gate notes . gate all pyqs answer
 
Digital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdfDigital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdf
 
NO1 Pandit Amil Baba In Bahawalpur, Sargodha, Sialkot, Sheikhupura, Rahim Yar...
NO1 Pandit Amil Baba In Bahawalpur, Sargodha, Sialkot, Sheikhupura, Rahim Yar...NO1 Pandit Amil Baba In Bahawalpur, Sargodha, Sialkot, Sheikhupura, Rahim Yar...
NO1 Pandit Amil Baba In Bahawalpur, Sargodha, Sialkot, Sheikhupura, Rahim Yar...
 
The Benefits and Techniques of Trenchless Pipe Repair.pdf
The Benefits and Techniques of Trenchless Pipe Repair.pdfThe Benefits and Techniques of Trenchless Pipe Repair.pdf
The Benefits and Techniques of Trenchless Pipe Repair.pdf
 
Introduction to Casting Processes in Manufacturing
Introduction to Casting Processes in ManufacturingIntroduction to Casting Processes in Manufacturing
Introduction to Casting Processes in Manufacturing
 
ENERGY STORAGE DEVICES INTRODUCTION UNIT-I
ENERGY STORAGE DEVICES  INTRODUCTION UNIT-IENERGY STORAGE DEVICES  INTRODUCTION UNIT-I
ENERGY STORAGE DEVICES INTRODUCTION UNIT-I
 
2024 DevOps Pro Europe - Growing at the edge
2024 DevOps Pro Europe - Growing at the edge2024 DevOps Pro Europe - Growing at the edge
2024 DevOps Pro Europe - Growing at the edge
 
shape functions of 1D and 2 D rectangular elements.pptx
shape functions of 1D and 2 D rectangular elements.pptxshape functions of 1D and 2 D rectangular elements.pptx
shape functions of 1D and 2 D rectangular elements.pptx
 
fundamentals of drawing and isometric and orthographic projection
fundamentals of drawing and isometric and orthographic projectionfundamentals of drawing and isometric and orthographic projection
fundamentals of drawing and isometric and orthographic projection
 
A case study of cinema management system project report..pdf
A case study of cinema management system project report..pdfA case study of cinema management system project report..pdf
A case study of cinema management system project report..pdf
 
CME397 Surface Engineering- Professional Elective
CME397 Surface Engineering- Professional ElectiveCME397 Surface Engineering- Professional Elective
CME397 Surface Engineering- Professional Elective
 
Scaling in conventional MOSFET for constant electric field and constant voltage
Scaling in conventional MOSFET for constant electric field and constant voltageScaling in conventional MOSFET for constant electric field and constant voltage
Scaling in conventional MOSFET for constant electric field and constant voltage
 
Architectural Portfolio Sean Lockwood
Architectural Portfolio Sean LockwoodArchitectural Portfolio Sean Lockwood
Architectural Portfolio Sean Lockwood
 
ASME IX(9) 2007 Full Version .pdf
ASME IX(9)  2007 Full Version       .pdfASME IX(9)  2007 Full Version       .pdf
ASME IX(9) 2007 Full Version .pdf
 

EssentialsOfMachineLearning.pdf

  • 1. Machine Learning and its Applications Presented by Ankita Tiwari
  • 2. Definition by Tom Mitchell(1998): Machine Learning is the study of algorithms that • improve their performance P • at some task T • with experience E A well-defined learning task is given by <P, T, E>. What is Machine Learning? “Learning is a process by which a system improves its performance from experience.” - Herbert Simon
  • 3.
  • 4. When Do We Use Machine Learning ML is used when: • Human expertise does not exist (navigation on mars ) • Human can't explain their expertise (speech recognition) • Models must be customized (personalized medicine) • Model are based on huge amounts of data (genomics) Learning isn't always useful: • There is no need to "learn" to calculate payroll
  • 5. 1. What exactly is deep learning ? 2. Why is it generally better than other methods on image, speech and certain other types of data? The short answers • ‘Deep Learning’ means using a neural network with several layers of nodes between input and output, • The combination of layers between input & output perform feature identification and processing in stages, just as our brains seem to.
  • 7. Types of Learning Unsupervised Learning Unsupervised learning is the training of a machine using information that is neither classified nor labeled and allowing the algorithm to act on that information without guidance. Here the task of the machine is to group unsorted information according to similarities, patterns, and differences without any prior training of data. Unlike supervised learning, no teacher is provided that means no training will be given to the machine. Therefore the machine is restricted to find the hidden structure in unlabeled data by itself. Supervised Learning It has the presence of a supervisor as a teacher. Basically supervised learning is when we teach or train the machine using data that is well labeled. Which means some data is already tagged with the correct answer. After that, the machine is provided with a new set of examples(data) so that the supervised learning algorithm analyses the training data(set of training examples) and produces a correct outcome from labeled data.
  • 8. Unsupervised Learning Unsupervised learning is classified into two categories of algorithms: • Clustering: A clustering problem is where you want to discover the inherent groupings in the data, such as grouping customers by purchasing behavior. • Association: An association rule learning problem is where you want to discover rules that describe large portions of your data, such as people that buy X also tend to buy Y. Types of Unsupervised Learning:- Clustering • Exclusive (partitioning) • Agglomerative • Overlapping • Probabilistic Clustering Types:- • Hierarchical clustering • K-means clustering • Principal Component Analysis • Singular Value Decomposition • Independent Component Analysis
  • 9. Supervised Learning Supervised learning is classified into two categories of algorithms: • Classification: A classification problem is when the output variable is a category, such as “Red” or “blue” or “disease” and “no disease”. • Regression: A regression problem is when the output variable is a real value, such as “dollars” or “weight”. Supervised learning deals with or learns with “labeled” data. This implies that some data is already tagged with the correct answer. Types:- • Regression • Logistic Regression • Classification • Naive Bayes Classifiers • K-NN (k nearest neighbors) • Decision Trees • Support Vector Machine
  • 10. Artificial Neural Network (ANN) • It is a type of supervised learning. • An ANN is an information processing paradigm that is inspired by the brain. • It learns by examples. It is configured for a specific application, such as pattern recognition or data classification, through a learning process. • Learning largely involves adjustments to the synaptic connections that exist between the neurons.
  • 11. Analogy Between Biological and ANN Biological Neuron Artificial Neuron
  • 13. 2.7 -8.6 0.002 f(x) 1.4 -2.5 -0.06 x = -0.06×2.7 + 2.5×8.6 + 1.4×0.002 = 21.34 Working of Perceptron
  • 14. A dataset Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Training and weights update
  • 15. Training the neural network Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Training and weights update
  • 16. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Initialise with random weights Training and weights update
  • 17. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Present a training pattern 1.4 2.7 1.9 Training and weights update
  • 18. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Feed it through to get output 1.4 2.7 0.8 1.9 Training and weights update
  • 19. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Compare with target output 1.4 2.7 0.8 0 1.9 error 0.8 Training and weights update
  • 20. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Adjust weights based on error 1.4 2.7 0.8 0 1.9 error 0.8 Training and weights update
  • 21. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Present a training pattern 6.4 2.8 1.7 Training and weights update
  • 22. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Feed it through to get output 6.4 2.8 0.9 1.7 Training and weights update
  • 23. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Compare with target output 6.4 2.8 0.9 1 1.7 error -0.1 Training and weights update
  • 24. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … Adjust weights based on error 6.4 2.8 0.9 1 1.7 error -0.1 Training and weights update
  • 25. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … And so on... 6.4 2.8 0.9 1 1.7 error -0.1 Training and weights update
  • 26. Training data Fields class 1.4 2.7 1.9 0 3.8 3.4 3.2 0 6.4 2.8 1.7 1 4.1 0.1 0.2 0 etc … And so on …. 6.4 2.8 0.9 1 1.7 error -0.1 • Repeat the process number of times, maybe millions of times each time taking a random training instance, and making slight weight adjustments • Algorithms for weight adjustment are designed to make changes that will reduce the error Training and weights update
  • 27. The decision boundary perspective Initial random weights
  • 28. The decision boundary perspective Present a training instance / adjust the weights
  • 29. The decision boundary perspective Present a training instance / adjust the weights
  • 30. The decision boundary perspective Present a training instance / adjust the weights
  • 31. The decision boundary perspective Present a training instance / adjust the weights
  • 32. Eventually …. The decision boundary perspective
  • 33. Summary • In general, weight learning algorithms of ANNs are like GIGO. • They work by making thousands and thousands of tiny adjustments, each making the network do better at the most recent pattern, but perhaps a little worse on many others • But, by luck, eventually this tends to be good enough to learn effective classifiers for many real life applications
  • 35. what is this unit doing? Hand-written data set
  • 36. Hidden layer makes connection of neuron strong or weak, based on inputs … 1 63 1 5 10 15 20 25 … strong +ve weight low/zero weight Processing of hidden layer
  • 37. Processing of hidden layer … 1 63 1 5 10 15 20 25 … strong +ve weight low/zero weight it will send strong signal for a horizontal line in the top row, ignoring everywhere else
  • 38. … 1 63 1 5 10 15 20 25 … strong +ve weight low/zero weight Processing of hidden layer
  • 39. … 1 63 1 5 10 15 20 25 … strong +ve weight low/zero weight Strong signal for a dark area in the top left corner Processing of hidden layer
  • 40. So: multiple layers make sense
  • 41. So: multiple layers make sense Your brain works that way
  • 42. So: multiple layers make sense Multi-layer neural network architectures can learn the true underlying features and ‘feature logic’, and therefore generalise very well.
  • 43. But, until very recently, our weight-learning algorithms simply did not work on multi-layer architectures
  • 44. There are new methods are proposed to train multi-layer NN
  • 45. Train this layer first The new way to train multi-layer NN
  • 46. Train this layer first then this layer The new way to train multi-layer NN
  • 47. Train this layer first then this layer then this layer The new way to train multi-layer NN
  • 48. Train this layer first then this layer then this layer then this layer The new way to train multi-layer NN
  • 49. Train this layer first then this layer then this layer then this layer finally this layer The new way to train multi-layer NN
  • 50. EACH of the (non-output) layers is trained to be an auto-encoder Basically, it is forced to learn good features that describe what comes from the previous layer The new way to train multi-layer NN
  • 51. Conclusion • There are many types of deep learning models, • Different kinds of autoencoder, variations on architectures and training algorithms, etc. • It is a very fast-growing area.
  • 56. Deep Learning in the Headlines
  • 57. Scene Labeling via Deep Learning
  • 58. Machine Learning in Automatic Speech Recognition A Typical Speech Recognition System ML used to predict of phone states from the sound spectrogram Deep learning has state-of-the-art results Zeiler et al. “On rectified linear units for speech recognition” ICASSP 2013
  • 59. Impact of Deep Learning in Speech Technology
  • 60. Continuous Observation Continuous transfer of data for observation and monitoring Delay, Power, Area
  • 61. Device that analysis and compute at same level like human brain Requirement of Artificial Intelligent device ANN CNN SNN ANN / CNN 1) Large computation 2) Power hungry 3) Large area Alternative SNN: 1) High parallel computation 2) Power efficient 3) Event-driven 4) More brain inspired Most remarkable neural network
  • 63. Proposed solutions 21-02-2023 • Reduce Data Movement • Maximize Data Reuse within PE • Maximize Data Reuse with Buffer • Multicast Network Design • Exploit Data Statistics • Data Compression • Data gating/ zero skipping • Operations exhibit High Parallelism