SlideShare a Scribd company logo
UNSUPERVISED LEARNING
1
2
We don’t have outputs defined for the given inputs, but we need to group data into different
clusters for classification, we use Unsupervised Learning
The flowchart:
3
Unlabelled Training Data
Possible Clusters
We are finding cohesive groups of points and points which are outliers
I need some form of bias: here I have assumed all clusters as ellipsoids
Not all points fall into clusters
Applications
• Customer data (targeted promotions for group of customers)
– Discover classes of customers
• Image pixels (segmentation)
– Discover regions
• Words
– Synonyms
• Documents
– Topics
7
8
9
10
11
12
13
14
15
16
17
18
Mathematical Formula
19
20
21
How to get optimal solution
22
23
I X1 X2
A 1 1
B 1.5 2
C 3 4
D 4.5 5.5
E 3.5 5
F 3.5 4.5
Let’s take 2 points as cluster centers for the k=2
We can take point A and point B randomly
Now find distance of each point from these cluster centers
e.g. Distance of B from A = ((1.5-1)2 + (2-1)2))1/2 = 1.118
Distance of C from A = ((3-1)2 + (4-1)2))1/2 = 3.6 and so on…
0
1
2
3
4
5
6
0 1 2 3 4 5
24
i Cluster1 Cluster2
A 0 1.18
B 1.18 0
C
D
E
F
G
Example 2
R CODE
32
Choosing number of clusters
33
Perform k-means clustering on a data matrix.
34
• kmeans(x, centers, iter.max = 10, nstart = 1, algorithm = c("Hartigan-Wong",
"Lloyd", "Forgy", "MacQueen"), trace=FALSE) ## S3 method for class 'kmeans'
fitted
• (object, method = c("centers", "classes"), ...)
• Arguments
• x: numeric matrix of data, or an object that can be coerced to such a matrix (such
as a numeric vector or a data frame with all numeric columns).
• centers: either the number of clusters, say k, or a set of initial (distinct) cluster
centres.
• iter.max : the maximum number of iterations allowed.
• nstart: if centers is a number, how many random sets should be chosen?
• algorithm character: may be abbreviated. Note that "Lloyd" and "Forgy" and
alternative names for one algorithm.
• object: an R object of class "kmeans", typically the result ob of ob <- kmeans(..).
• Method: character: may be abbreviated. "centers" causes fitted to return cluster
centers (one for each input point) and"classes" causes fitted to return a vector of
class assignments.
Value
35
• kmeans returns an object of class "kmeans" which has a print and
a fitted method. It is a list with at least the following components:
• Cluster: A vector of integers (from 1:k) indicating the cluster to which each point
is allocated.
• Centers: A matrix of cluster centres.
• Totss: The total sum of squares.
• Withinss: Vector of within-cluster sum of squares, one component per cluster.
• tot.withinss: Total within-cluster sum of squares, i.e., sum(withinss).
• Betweenss: The between-cluster sum of squares, i.e. totss-tot.withinss.
• Size: The number of points in each cluster.
• Iter: The number of (outer) iterations.
• ifaultinteger: indicator of a possible algorithm problem – for experts.
R code
36
i<-iris
View(i)
i1<- i[,1]
View(i1)
i2<- i[,c(1,2)]
i3<- i[,c(1,2,3)]
i4<- i[,c(1,2,3,4)]
View(i2)
View(i3)
View(i4)
c1<-kmeans(i1,3)
c2<-kmeans(i2,3)
c3<-kmeans(i3,3)
c4<-kmeans(i4,3)
ic1<- data.frame(i1,i$Species,c1$cluster)
View(ic1)
ic2<- data.frame(i2,i$Species,c2$cluster)
View(ic2)
ic3<- data.frame(i3,i$Species,c3$cluster)
View(ic3)
ic4<-data.frame(i4,i$Species,c4$cluster)
View(ic4)
Example output: c2
37
• K-means clustering with 3 clusters of sizes 50, 47, 53 Cluster means:
• Sepal.Length Sepal.Width
• 1 5.006000 3.428000
• 2 6.812766 3.074468
• 3 5.773585 2.692453
• Clustering vector: [1] 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 [47] 1 1 1 1 2 2 2 3 2 3 2 3 2 3 3
3 3 3 3 2 3 3 3 3 3 3 3 3 2 2 2 2 3 3 3 3 3 3 3 3 2 3 3 3 3 3 [93] 3 3 3 3
3 3 3 3 2 3 2 2 2 2 3 2 2 2 2 2 2 3 3 2 2 2 2 3 2 3 2 3 2 2 3 3 2 2 2 2 2
3 3 2 2 2 [139] 3 2 2 2 3 2 2 2 3 2 2 3
• Within cluster sum of squares by cluster:
• [1] 13.1290 12.6217 11.3000
• (between_SS / total_SS = 71.6 %)
• Available components: [1] "cluster" "centers" "totss" "withinss"
"tot.withinss" "betweenss" [7] "size" "iter" "ifault”
Errors
38
• table(ic1[,2],ic1[,3])
table(ic2[,3],ic2[,4])
CODE for Elbow Rule
39
• ir<- iris
• irt<-ir[,-5]
• head(ir)
• head(irt)
• #Elbow Rule
• a<- numeric()
• for(i in 1:8){
•
• irk1<- kmeans(irt,i)
• a<- append(a,irk1$tot.withinss,i)
• }
• b<-1:8
• plot(b,a)
NAÏVE BAYE’S
40
41
42
43
44
Consider a school with a total population of 100 persons. These
100 persons can be seen either as ‘Students’ and ‘Teachers’ or as
a population of ‘Males’ and ‘Females’.
With below tabulation of the 100 people, what is the conditional
probability that a certain member of the school is a ‘Teacher’ given
that he is a ‘Man’?
Problem
• Say you have 1000 fruits which could be either ‘banana’,
‘orange’ or ‘other’. These are the 3 possible classes of the
Y variable. We have data for the following X variables, all
of which are binary (1 or 0).
• Long
• Sweet
• Yellow
So the objective of the classifier is to predict if a given fruit is a
‘Banana’ or ‘Orange’ or ‘Other’ when only the 3 features (long, sweet
and yellow) are known.
• Let’s say you are given a fruit that is: Long, Sweet and
Yellow, can you predict what fruit it is?
• This is the same of predicting the Y when only the X variables
in testing data are known.
• Let’s solve it by hand using Naive Bayes. The idea is to
compute the 3 probabilities, that is the probability of the fruit
being a banana, orange or other. Whichever fruit type gets
the highest probability wins.
Step 1: Compute the ‘Prior’ probabilities for each of the class of
fruits.
That is, the proportion of each fruit class out of all the fruits from the
population.
You can provide the ‘Priors’ from prior information about the
population. Otherwise, it can be computed from the training data.
For this case, let’s compute from the training data. Out of 1000
records in training data, you have 500 Bananas, 300 Oranges and
200 Others.
So the respective priors are 0.5, 0.3 and 0.2.
P(Y=Banana) = 500 / 1000 = 0.50
Step 2: Compute the probability of evidence that goes in the
denominator.
This is nothing but the product of P of Xs for all X.
This is an optional step because the denominator is the same for
all the classes and so will not affect the probabilities.
P(x1=Long) = 500 / 1000 = 0.50
P(x2=Sweet) = 650 / 1000 = 0.65
P(x3=Yellow) = 800 / 1000 = 0.80
Step 3: Compute the probability of likelihood of evidences
that goes in the numerator.
It is the product of conditional probabilities of the 3 features.
If you refer back to the formula, it says P(X1 |Y=k).
Here X1 is ‘Long’ and k is ‘Banana’.
That means the probability the fruit is ‘Long’ given that it is a
Banana.
In the above table, you have 500 Bananas. Out of that 400 is
long.
So, P(Long | Banana) = 400/500 = 0.8.
Here, I have done it for Banana alone.
Probability of Likelihood for Banana
P(x1=Long | Y=Banana) = 400 / 500 = 0.80
P(x2=Sweet | Y=Banana) = 350 / 500 = 0.70
P(x3=Yellow | Y=Banana) = 450 / 500 = 0.90.
So, the overall probability of Likelihood of evidence for
Banana = 0.8 * 0.7 * 0.9 = 0.504
For example, probability of playing golf given
that the temperature is cool,
i.e P(temp. = cool | play golf = Yes) = 3/9.
Also, we need to find class probabilities
(P(y)) which has been calculated in the table
5.
For example, P(play golf = Yes) = 9/14.
today = (Sunny, Hot, Normal, False)
kmean_naivebayes.pptx

More Related Content

Similar to kmean_naivebayes.pptx

K Means Clustering in ML.pptx
K Means Clustering in ML.pptxK Means Clustering in ML.pptx
K Means Clustering in ML.pptx
Ramakrishna Reddy Bijjam
 
Mean (Girja Pd. Patel)
Mean (Girja Pd. Patel)Mean (Girja Pd. Patel)
Mean (Girja Pd. Patel)
GirjaPrasad
 
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik   statistik deskriptif 2 numerikEe184405 statistika dan stokastik   statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
yusufbf
 
Clustering for Beginners
Clustering for BeginnersClustering for Beginners
Clustering for Beginners
Sayeed Mahmud
 
MEASURE OF CENTRAL TENDENCY
MEASURE OF CENTRAL TENDENCY  MEASURE OF CENTRAL TENDENCY
MEASURE OF CENTRAL TENDENCY
AB Rajar
 
analytical representation of data
 analytical representation of data analytical representation of data
analytical representation of data
Unsa Shakir
 
Uta005 lecture3
Uta005 lecture3Uta005 lecture3
Uta005 lecture3
vinay arora
 
CSA 3702 machine learning module 2
CSA 3702 machine learning module 2CSA 3702 machine learning module 2
CSA 3702 machine learning module 2
Nandhini S
 
Statistics for math (English Version)
Statistics for math (English Version)Statistics for math (English Version)
Statistics for math (English Version)
Tito_14
 
Lec13 Clustering.pptx
Lec13 Clustering.pptxLec13 Clustering.pptx
Lec13 Clustering.pptx
Khalid Rabayah
 
Measures of central tendency and dispersion
Measures of central tendency and dispersionMeasures of central tendency and dispersion
Measures of central tendency and dispersion
Abhinav yadav
 
Mathematics online: some common algorithms
Mathematics online: some common algorithmsMathematics online: some common algorithms
Mathematics online: some common algorithms
Mark Moriarty
 
Statistics - Basics
Statistics - BasicsStatistics - Basics
Statistics - Basics
Andrew Ferlitsch
 
Lecture 3 & 4 Measure of Central Tendency.pdf
Lecture 3 & 4 Measure of Central Tendency.pdfLecture 3 & 4 Measure of Central Tendency.pdf
Lecture 3 & 4 Measure of Central Tendency.pdf
kelashraisal
 
Lecture4.pptx
Lecture4.pptxLecture4.pptx
Lecture4.pptx
yasir149288
 
Permutations and Combinations IIT JEE+Olympiad Lecture 1
Permutations and Combinations IIT JEE+Olympiad Lecture 1 Permutations and Combinations IIT JEE+Olympiad Lecture 1
Permutations and Combinations IIT JEE+Olympiad Lecture 1
Parth Nandedkar
 
Unsupervised Learning in Machine Learning
Unsupervised Learning in Machine LearningUnsupervised Learning in Machine Learning
Unsupervised Learning in Machine Learning
Pyingkodi Maran
 
Basics of Stats (2).pptx
Basics of Stats (2).pptxBasics of Stats (2).pptx
Basics of Stats (2).pptx
madihamaqbool6
 
Unit 2 in daa
Unit 2 in daaUnit 2 in daa
Unit 2 in daa
Nv Thejaswini
 
algorithm Unit 2
algorithm Unit 2 algorithm Unit 2
algorithm Unit 2
Monika Choudhery
 

Similar to kmean_naivebayes.pptx (20)

K Means Clustering in ML.pptx
K Means Clustering in ML.pptxK Means Clustering in ML.pptx
K Means Clustering in ML.pptx
 
Mean (Girja Pd. Patel)
Mean (Girja Pd. Patel)Mean (Girja Pd. Patel)
Mean (Girja Pd. Patel)
 
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik   statistik deskriptif 2 numerikEe184405 statistika dan stokastik   statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
 
Clustering for Beginners
Clustering for BeginnersClustering for Beginners
Clustering for Beginners
 
MEASURE OF CENTRAL TENDENCY
MEASURE OF CENTRAL TENDENCY  MEASURE OF CENTRAL TENDENCY
MEASURE OF CENTRAL TENDENCY
 
analytical representation of data
 analytical representation of data analytical representation of data
analytical representation of data
 
Uta005 lecture3
Uta005 lecture3Uta005 lecture3
Uta005 lecture3
 
CSA 3702 machine learning module 2
CSA 3702 machine learning module 2CSA 3702 machine learning module 2
CSA 3702 machine learning module 2
 
Statistics for math (English Version)
Statistics for math (English Version)Statistics for math (English Version)
Statistics for math (English Version)
 
Lec13 Clustering.pptx
Lec13 Clustering.pptxLec13 Clustering.pptx
Lec13 Clustering.pptx
 
Measures of central tendency and dispersion
Measures of central tendency and dispersionMeasures of central tendency and dispersion
Measures of central tendency and dispersion
 
Mathematics online: some common algorithms
Mathematics online: some common algorithmsMathematics online: some common algorithms
Mathematics online: some common algorithms
 
Statistics - Basics
Statistics - BasicsStatistics - Basics
Statistics - Basics
 
Lecture 3 & 4 Measure of Central Tendency.pdf
Lecture 3 & 4 Measure of Central Tendency.pdfLecture 3 & 4 Measure of Central Tendency.pdf
Lecture 3 & 4 Measure of Central Tendency.pdf
 
Lecture4.pptx
Lecture4.pptxLecture4.pptx
Lecture4.pptx
 
Permutations and Combinations IIT JEE+Olympiad Lecture 1
Permutations and Combinations IIT JEE+Olympiad Lecture 1 Permutations and Combinations IIT JEE+Olympiad Lecture 1
Permutations and Combinations IIT JEE+Olympiad Lecture 1
 
Unsupervised Learning in Machine Learning
Unsupervised Learning in Machine LearningUnsupervised Learning in Machine Learning
Unsupervised Learning in Machine Learning
 
Basics of Stats (2).pptx
Basics of Stats (2).pptxBasics of Stats (2).pptx
Basics of Stats (2).pptx
 
Unit 2 in daa
Unit 2 in daaUnit 2 in daa
Unit 2 in daa
 
algorithm Unit 2
algorithm Unit 2 algorithm Unit 2
algorithm Unit 2
 

Recently uploaded

Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&BDesign and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Sreedhar Chowdam
 
Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
Aditya Rajan Patra
 
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
zwunae
 
introduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdfintroduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdf
ravindarpurohit26
 
Swimming pool mechanical components design.pptx
Swimming pool  mechanical components design.pptxSwimming pool  mechanical components design.pptx
Swimming pool mechanical components design.pptx
yokeleetan1
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
obonagu
 
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdfTutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
aqil azizi
 
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Soumen Santra
 
Building Electrical System Design & Installation
Building Electrical System Design & InstallationBuilding Electrical System Design & Installation
Building Electrical System Design & Installation
symbo111
 
basic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdfbasic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdf
NidhalKahouli2
 
ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024
Rahul
 
Adaptive synchronous sliding control for a robot manipulator based on neural ...
Adaptive synchronous sliding control for a robot manipulator based on neural ...Adaptive synchronous sliding control for a robot manipulator based on neural ...
Adaptive synchronous sliding control for a robot manipulator based on neural ...
IJECEIAES
 
6th International Conference on Machine Learning & Applications (CMLA 2024)
6th International Conference on Machine Learning & Applications (CMLA 2024)6th International Conference on Machine Learning & Applications (CMLA 2024)
6th International Conference on Machine Learning & Applications (CMLA 2024)
ClaraZara1
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
Ratnakar Mikkili
 
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
obonagu
 
This is my Environmental physics presentation
This is my Environmental physics presentationThis is my Environmental physics presentation
This is my Environmental physics presentation
ZainabHashmi17
 
Technical Drawings introduction to drawing of prisms
Technical Drawings introduction to drawing of prismsTechnical Drawings introduction to drawing of prisms
Technical Drawings introduction to drawing of prisms
heavyhaig
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
Madan Karki
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
KrishnaveniKrishnara1
 
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTCHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
jpsjournal1
 

Recently uploaded (20)

Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&BDesign and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
 
Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
 
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
一比一原版(IIT毕业证)伊利诺伊理工大学毕业证成绩单专业办理
 
introduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdfintroduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdf
 
Swimming pool mechanical components design.pptx
Swimming pool  mechanical components design.pptxSwimming pool  mechanical components design.pptx
Swimming pool mechanical components design.pptx
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
 
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdfTutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
 
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
 
Building Electrical System Design & Installation
Building Electrical System Design & InstallationBuilding Electrical System Design & Installation
Building Electrical System Design & Installation
 
basic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdfbasic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdf
 
ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024
 
Adaptive synchronous sliding control for a robot manipulator based on neural ...
Adaptive synchronous sliding control for a robot manipulator based on neural ...Adaptive synchronous sliding control for a robot manipulator based on neural ...
Adaptive synchronous sliding control for a robot manipulator based on neural ...
 
6th International Conference on Machine Learning & Applications (CMLA 2024)
6th International Conference on Machine Learning & Applications (CMLA 2024)6th International Conference on Machine Learning & Applications (CMLA 2024)
6th International Conference on Machine Learning & Applications (CMLA 2024)
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
 
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
 
This is my Environmental physics presentation
This is my Environmental physics presentationThis is my Environmental physics presentation
This is my Environmental physics presentation
 
Technical Drawings introduction to drawing of prisms
Technical Drawings introduction to drawing of prismsTechnical Drawings introduction to drawing of prisms
Technical Drawings introduction to drawing of prisms
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
 
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTCHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
 

kmean_naivebayes.pptx

  • 2. 2 We don’t have outputs defined for the given inputs, but we need to group data into different clusters for classification, we use Unsupervised Learning
  • 5. Possible Clusters We are finding cohesive groups of points and points which are outliers I need some form of bias: here I have assumed all clusters as ellipsoids Not all points fall into clusters
  • 6. Applications • Customer data (targeted promotions for group of customers) – Discover classes of customers • Image pixels (segmentation) – Discover regions • Words – Synonyms • Documents – Topics
  • 7. 7
  • 8. 8
  • 9. 9
  • 10. 10
  • 11. 11
  • 12. 12
  • 13. 13
  • 14. 14
  • 15. 15
  • 16. 16
  • 17. 17
  • 18. 18
  • 20. 20
  • 21. 21
  • 22. How to get optimal solution 22
  • 23. 23 I X1 X2 A 1 1 B 1.5 2 C 3 4 D 4.5 5.5 E 3.5 5 F 3.5 4.5 Let’s take 2 points as cluster centers for the k=2 We can take point A and point B randomly Now find distance of each point from these cluster centers e.g. Distance of B from A = ((1.5-1)2 + (2-1)2))1/2 = 1.118 Distance of C from A = ((3-1)2 + (4-1)2))1/2 = 3.6 and so on… 0 1 2 3 4 5 6 0 1 2 3 4 5
  • 24. 24 i Cluster1 Cluster2 A 0 1.18 B 1.18 0 C D E F G
  • 26.
  • 27.
  • 28.
  • 29.
  • 30.
  • 31.
  • 33. Choosing number of clusters 33
  • 34. Perform k-means clustering on a data matrix. 34 • kmeans(x, centers, iter.max = 10, nstart = 1, algorithm = c("Hartigan-Wong", "Lloyd", "Forgy", "MacQueen"), trace=FALSE) ## S3 method for class 'kmeans' fitted • (object, method = c("centers", "classes"), ...) • Arguments • x: numeric matrix of data, or an object that can be coerced to such a matrix (such as a numeric vector or a data frame with all numeric columns). • centers: either the number of clusters, say k, or a set of initial (distinct) cluster centres. • iter.max : the maximum number of iterations allowed. • nstart: if centers is a number, how many random sets should be chosen? • algorithm character: may be abbreviated. Note that "Lloyd" and "Forgy" and alternative names for one algorithm. • object: an R object of class "kmeans", typically the result ob of ob <- kmeans(..). • Method: character: may be abbreviated. "centers" causes fitted to return cluster centers (one for each input point) and"classes" causes fitted to return a vector of class assignments.
  • 35. Value 35 • kmeans returns an object of class "kmeans" which has a print and a fitted method. It is a list with at least the following components: • Cluster: A vector of integers (from 1:k) indicating the cluster to which each point is allocated. • Centers: A matrix of cluster centres. • Totss: The total sum of squares. • Withinss: Vector of within-cluster sum of squares, one component per cluster. • tot.withinss: Total within-cluster sum of squares, i.e., sum(withinss). • Betweenss: The between-cluster sum of squares, i.e. totss-tot.withinss. • Size: The number of points in each cluster. • Iter: The number of (outer) iterations. • ifaultinteger: indicator of a possible algorithm problem – for experts.
  • 36. R code 36 i<-iris View(i) i1<- i[,1] View(i1) i2<- i[,c(1,2)] i3<- i[,c(1,2,3)] i4<- i[,c(1,2,3,4)] View(i2) View(i3) View(i4) c1<-kmeans(i1,3) c2<-kmeans(i2,3) c3<-kmeans(i3,3) c4<-kmeans(i4,3) ic1<- data.frame(i1,i$Species,c1$cluster) View(ic1) ic2<- data.frame(i2,i$Species,c2$cluster) View(ic2) ic3<- data.frame(i3,i$Species,c3$cluster) View(ic3) ic4<-data.frame(i4,i$Species,c4$cluster) View(ic4)
  • 37. Example output: c2 37 • K-means clustering with 3 clusters of sizes 50, 47, 53 Cluster means: • Sepal.Length Sepal.Width • 1 5.006000 3.428000 • 2 6.812766 3.074468 • 3 5.773585 2.692453 • Clustering vector: [1] 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 [47] 1 1 1 1 2 2 2 3 2 3 2 3 2 3 3 3 3 3 3 2 3 3 3 3 3 3 3 3 2 2 2 2 3 3 3 3 3 3 3 3 2 3 3 3 3 3 [93] 3 3 3 3 3 3 3 3 2 3 2 2 2 2 3 2 2 2 2 2 2 3 3 2 2 2 2 3 2 3 2 3 2 2 3 3 2 2 2 2 2 3 3 2 2 2 [139] 3 2 2 2 3 2 2 2 3 2 2 3 • Within cluster sum of squares by cluster: • [1] 13.1290 12.6217 11.3000 • (between_SS / total_SS = 71.6 %) • Available components: [1] "cluster" "centers" "totss" "withinss" "tot.withinss" "betweenss" [7] "size" "iter" "ifault”
  • 39. CODE for Elbow Rule 39 • ir<- iris • irt<-ir[,-5] • head(ir) • head(irt) • #Elbow Rule • a<- numeric() • for(i in 1:8){ • • irk1<- kmeans(irt,i) • a<- append(a,irk1$tot.withinss,i) • } • b<-1:8 • plot(b,a)
  • 41. 41
  • 42. 42
  • 43. 43
  • 44. 44
  • 45. Consider a school with a total population of 100 persons. These 100 persons can be seen either as ‘Students’ and ‘Teachers’ or as a population of ‘Males’ and ‘Females’. With below tabulation of the 100 people, what is the conditional probability that a certain member of the school is a ‘Teacher’ given that he is a ‘Man’?
  • 46.
  • 47.
  • 48.
  • 49.
  • 50. Problem • Say you have 1000 fruits which could be either ‘banana’, ‘orange’ or ‘other’. These are the 3 possible classes of the Y variable. We have data for the following X variables, all of which are binary (1 or 0). • Long • Sweet • Yellow
  • 51. So the objective of the classifier is to predict if a given fruit is a ‘Banana’ or ‘Orange’ or ‘Other’ when only the 3 features (long, sweet and yellow) are known.
  • 52. • Let’s say you are given a fruit that is: Long, Sweet and Yellow, can you predict what fruit it is? • This is the same of predicting the Y when only the X variables in testing data are known. • Let’s solve it by hand using Naive Bayes. The idea is to compute the 3 probabilities, that is the probability of the fruit being a banana, orange or other. Whichever fruit type gets the highest probability wins.
  • 53. Step 1: Compute the ‘Prior’ probabilities for each of the class of fruits. That is, the proportion of each fruit class out of all the fruits from the population. You can provide the ‘Priors’ from prior information about the population. Otherwise, it can be computed from the training data. For this case, let’s compute from the training data. Out of 1000 records in training data, you have 500 Bananas, 300 Oranges and 200 Others. So the respective priors are 0.5, 0.3 and 0.2. P(Y=Banana) = 500 / 1000 = 0.50
  • 54. Step 2: Compute the probability of evidence that goes in the denominator. This is nothing but the product of P of Xs for all X. This is an optional step because the denominator is the same for all the classes and so will not affect the probabilities. P(x1=Long) = 500 / 1000 = 0.50 P(x2=Sweet) = 650 / 1000 = 0.65 P(x3=Yellow) = 800 / 1000 = 0.80
  • 55. Step 3: Compute the probability of likelihood of evidences that goes in the numerator. It is the product of conditional probabilities of the 3 features. If you refer back to the formula, it says P(X1 |Y=k). Here X1 is ‘Long’ and k is ‘Banana’. That means the probability the fruit is ‘Long’ given that it is a Banana. In the above table, you have 500 Bananas. Out of that 400 is long. So, P(Long | Banana) = 400/500 = 0.8. Here, I have done it for Banana alone.
  • 56. Probability of Likelihood for Banana P(x1=Long | Y=Banana) = 400 / 500 = 0.80 P(x2=Sweet | Y=Banana) = 350 / 500 = 0.70 P(x3=Yellow | Y=Banana) = 450 / 500 = 0.90. So, the overall probability of Likelihood of evidence for Banana = 0.8 * 0.7 * 0.9 = 0.504
  • 57.
  • 58.
  • 59.
  • 60. For example, probability of playing golf given that the temperature is cool, i.e P(temp. = cool | play golf = Yes) = 3/9. Also, we need to find class probabilities (P(y)) which has been calculated in the table 5. For example, P(play golf = Yes) = 9/14.
  • 61. today = (Sunny, Hot, Normal, False)