SlideShare a Scribd company logo
Introduction to
Unsupervised Learning
Legal Notices and Disclaimers
This presentation is for informational purposes only. INTEL MAKES NO WARRANTIES,
EXPRESS OR IMPLIED, IN THIS SUMMARY.
Intel technologies’ features and benefits depend on system configuration and may require
enabled hardware, software or service activation. Performance varies depending on system
configuration. Check with your system manufacturer or retailer or learn more at intel.com.
This sample source code is released under the Intel Sample Source Code License Agreement.
Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries.
*Other names and brands may be claimed as the property of others.
Copyright © 2017, Intel Corporation. All rights reserved.
Types of Machine Learning
data points have known outcomeSupervised
Unsupervised data points have unknown outcome
Types of Machine Learning
data points have known outcomeSupervised
Unsupervised data points have unknown outcome
Dimensionality
Reduction
Clustering
Types of Unsupervised Learning
identify unknown structure in data
use structural characteristics to simplify data
Dimensionality
Reduction
Types of Unsupervised Learning
identify unknown structure in data
use structural characteristics to simplify data
Clustering
Unsupervised Learning Overview
unlabeled data
(no answers)
map new
data to
structure
new
unlabeled
data
fit
+
structure
predict
model
text articles
of unknown
topics
model
predict similar
articles
text articles of
unknown
topics
fit
+
+
predict
Clustering: Finding Distinct Groups
model
model
high
resolution
images
model
compressed
images
high
resolution
images
fit
+
+
model
model
predict
Dimensionality Reduction: Simplifying Structure
Introduction to Unsupervised Learning
Users of a web application:
One feature (age)
Age
Introduction to Unsupervised Learning
Users of a web application:
One feature (age)
Two clusters
Age
Introduction to Unsupervised Learning
Users of a web application:
One feature (age)
Three clusters
Age
Introduction to Unsupervised Learning
Users of a web application:
One feature (age)
Five clusters
Age
Age
Income
K-Means Algorithm
K = 2 (find two clusters)
Age
Income
K-Means Algorithm
K = 2, Randomly assign cluster centers
Age
Income
K-Means Algorithm
K = 2, Each point belongs to closest center
Age
Income
K-Means Algorithm
K = 2, Move each center to cluster's mean
Age
Income
K-Means Algorithm
K = 2, Each point belongs to closest center
Age
Income
K-Means Algorithm
K = 2, Move each center to cluster's mean
Age
Income
K-Means Algorithm
K = 2, Points don't change  Converged
Age
Income
K-Means Algorithm
K = 2, Each point belongs to closest center
Age
Income
K-Means Algorithm
K = 3
Age
Income
K-Means Algorithm
K = 3, Results depend on initial cluster assignment
Age
Income
Which Model is the Right One?
Which Model is the Right One?
• Inertia: sum of squared distance from each
point (𝑥𝑖) to its cluster (𝐶 𝑘)
𝑖=1
𝑛
(𝑥𝑖 − 𝐶 𝑘)2
• Smaller value corresponds to tighter clusters
• Other metrics can also be used
Age
Income
Which Model is the Right One?
Initiate multiple times,
take model with the best score
Age
Income
Which Model is the Right One?
Inertia = 12.645
Age
Income
Which Model is the Right One?
Inertia = 12.943
Age
Income
Which Model is the Right One?
Inertia = 13.112
Age
Income
Smarter Initialization of K-Means Clusters
Age
Income
Smarter Initialization of K-Means Clusters
Pick one point at random as initial point
Age
Income
Smarter Initialization of K-Means Clusters
Pick next point with 1/distance2 probability
Age
Income
Smarter Initialization of K-Means Clusters
Pick next point with 1/distance2 probability
Age
Income
Smarter Initialization of K-Means Clusters
Pick next point with 1/distance2 probability
Age
Income
Smarter Initialization of K-Means Clusters
Assign clusters
Choosing the Right Number of Clusters
• Sometimes the question has a K
Choosing the Right Number of Clusters
• Sometimes the question has a K
• Clustering similar jobs on 4 CPU cores (K=4)
Choosing the Right Number of Clusters
• Sometimes the question has a K
• Clustering similar jobs on 4 CPU cores (K=4)
• A clothing design in 10 different sizes to cover
most people (K=10)
Choosing the Right Number of Clusters
• Sometimes the question has a K
• Clustering similar jobs on 4 CPU cores (K=4)
• A clothing design in 10 different sizes to cover
most people (K=10)
• A navigation interface for browsing scientific
papers with 20 disciplines (K=20)
Choosing the Right Number of Clusters
Inertia
K
Choosing the Right Number of Clusters
• Inertia measures distance
of point to cluster
• Value decreases with
increasing K as long as
cluster density increases
2 4 6 8 10
Inertia
K
Choosing the Right Number of Clusters
• Inertia measures distance
of point to cluster
• Value decreases with
increasing K as long as
cluster density increases
2 4 6 8 10
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then transform the data
X_trans = kmeans.fit_transform(X_sparse)
Can also be used in batch mode with MiniBatchKMeans.
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then transform the data
X_trans = kmeans.fit_transform(X_sparse)
Can also be used in batch mode with MiniBatchKMeans.
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then transform the data
X_trans = kmeans.fit_transform(X_sparse)
Can also be used in batch mode with MiniBatchKMeans.
final number
of clusters
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then transform the data
X_trans = kmeans.fit_transform(X_sparse)
Can also be used in batch mode with MiniBatchKMeans.
kmeans++
cluster
initiation
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then predict clusters for new data
kmeans = kmeans.fit(X1)
Can also be used in batch mode with MiniBatchKMeans.
K-Means: The Syntax
Import the class containing the clustering method
from sklearn.cluster import KMeans
Create an instance of the class
kmeans = KMeans(n_clusters=3,
init='k-means++')
Fit the instance on the data and then predict clusters for new data
kmeans = kmeans.fit(X1)
y_predict = kmeans.predict(X2)
Can also be used in batch mode with MiniBatchKMeans.
Distance Metrics
Distance Metric Choice
• Choice of distance metric is extremely
important to clustering success
• Each metric has strengths and most
appropriate use-cases…
• …but sometimes choice of distance
metric is also based on empirical
evaluation
Age
Income
Euclidean Distance
Age
Income
Euclidean Distance
Age
Income
Euclidean Distance (L2 Distance)
∆ Income
d
∆ Age
𝑑 = ∆𝐴𝑔𝑒2 + ∆𝐼𝑛𝑐𝑜𝑚𝑒2
Age
Income
Manhattan Distance (L1 or City Block Distance)
∆ Income
∆ Age
𝑑 = ∆𝐴𝑔𝑒 + ∆𝐼𝑛𝑐𝑜𝑚𝑒
Age
Income
Cosine Distance
cos(𝜃) =
𝐴 ∙ 𝐵
𝐴 𝐵
𝜃
Age
Income
Cosine Distance
cos(𝜃) =
𝐴 ∙ 𝐵
𝐴 𝐵
𝜃
Euclidean vs Cosine Distance
• Euclidean is useful for coordinate based
measurements
• Cosine is better for input data such as text
where location of occurrence is less important
• Euclidean distance is more sensitive to curse
of dimensionality (see lesson 12)
Euclidean vs Cosine Distance
• Euclidean is useful for coordinate based
measurements
• Cosine is better for data such as text where
location of occurrence is less important
• Euclidean distance is more sensitive to curse
of dimensionality (see lesson 12)
Euclidean vs Cosine Distance
• Euclidean is useful for coordinate based
measurements
• Cosine is better for data such as text where
location of occurrence is less important
• Euclidean distance is more sensitive to curse
of dimensionality (see lesson 12)
Jaccard Distance
1 −
𝐴 ∩ 𝐵
𝐴 ∪ 𝐵
= 1 −
𝑙𝑒𝑛(𝑠ℎ𝑎𝑟𝑒𝑑)
𝑙𝑒𝑛(𝑢𝑛𝑖𝑞𝑢𝑒)
Applies to sets (like word occurrence)
• Sentence A: “I like chocolate ice cream.”
• set A = {I, like, chocolate, ice, cream}
• Sentence B: “Do I want chocolate cream or vanilla cream?”
• set B = {Do, I, want, chocolate, cream, or, vanilla}
Jaccard Distance
1 −
𝐴 ∩ 𝐵
𝐴 ∪ 𝐵
= 1 −
3
9
• Sentence A: “I like chocolate ice cream.”
• set A = {I, like, chocolate, ice, cream}
• Sentence B: “Do I want chocolate cream or vanilla cream?”
• set B = {Do, I, want, chocolate, cream, or, vanilla}
Applies to sets (like word occurrence)
Distance Metrics: The Syntax
Import the general pairwise distance function
from sklearn.metrics import pairwise_distances
Calculate the distances
dist = pairwise_distances(X, Y,
metric='euclidean')
Other distance metric choices are: cosine, manhattan, jaccard, etc.
Distance metric methods can also be imported specifically, e.g.:
from sklearn.metrics import euclidean_distances
Distance Metrics: The Syntax
Import the general pairwise distance function
from sklearn.metrics import pairwise_distances
Calculate the distances
dist = pairwise_distances(X, Y,
metric='euclidean')
Other distance metric choices are: cosine, manhattan, jaccard, etc.
Distance metric methods can also be imported specifically, e.g.:
from sklearn.metrics import euclidean_distances
Distance Metrics: The Syntax
Import the general pairwise distance function
from sklearn.metrics import pairwise_distances
Calculate the distances
dist = pairwise_distances(X, Y,
metric='euclidean')
Other distance metric choices are: cosine, manhattan, jaccard, etc.
Distance metric methods can also be imported specifically, e.g.:
from sklearn.metrics import euclidean_distances
distance
metric choice
Distance Metrics: The Syntax
Import the general pairwise distance function
from sklearn.metrics import pairwise_distances
Calculate the distances
dist = pairwise_distances(X, Y,
metric='euclidean')
Other distance metric choices are: cosine, manhattan, jaccard, etc.
Distance metric methods can also be imported specifically, e.g.:
from sklearn.metrics import euclidean_distances
Distance Metrics: The Syntax
Import the general pairwise distance function
from sklearn.metrics import pairwise_distances
Calculate the distances
dist = pairwise_distances(X, Y,
metric='euclidean')
Other distance metric choices are: cosine, manhattan, jaccard, etc.
Distance metric methods can also be imported specifically, e.g.:
from sklearn.metrics import euclidean_distances
Hierarchical Agglomerative
Clustering
Age
Income
Hierarchical Agglomerative Clustering
Age
Income
Hierarchical Agglomerative Clustering
Find closest pair, merge into a cluster
Age
Income
Hierarchical Agglomerative ClusteringFind next closest pair and merge
Age
Income
Hierarchical Agglomerative Clustering
Find next closest pair and merge
Age
Income
Hierarchical Agglomerative Clustering
Keep merging closest pairs
Age
Income
Hierarchical Agglomerative Clustering
If the closest pair is two clusters, merge them
Age
Income
Hierarchical Agglomerative Clustering
Keep merging closest pairs and clusters
Age
Income
Hierarchical Agglomerative Clustering
Keep merging closest pairs and clusters
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 6
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 5
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 4
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 3
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 2
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 1
Agglomerative Clustering Stopping Conditions
the correct number of clusters is
reached
Condition 1
Condition 2
minimum cluster distance reaches a
set value
Agglomerative Clustering Stopping Conditions
the correct number of clusters is
reached
Condition 1
Condition 2
minimum average cluster distance
reaches a set value
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 5
Hierarchical Agglomerative Clustering
Current number of clusters = 5
Cluster
distance
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 4
Hierarchical Agglomerative Clustering
Current number of clusters = 4
Cluster
distance
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 3
Hierarchical Agglomerative Clustering
Current number of clusters = 3
Cluster
distance
Age
Income
Hierarchical Agglomerative Clustering
Current number of clusters = 2
Hierarchical Agglomerative Clustering
Current number of clusters = 2
Cluster
distance
Age
Income
Hierarchical Linkage Types
Single linkage: minimum pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Single linkage: minimum pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Complete linkage: maximum pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Complete linkage: maximum pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Average linkage: average pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Average linkage: average pairwise distance between clusters
Age
Income
Hierarchical Linkage Types
Ward linkage: merge based on best inertia
Age
Income
Hierarchical Linkage Types
Ward linkage: merge based on best inertia
Agglomerative Clustering: The Syntax
Import the class containing the clustering method
from sklearn.cluster import AgglomerativeClustering
Create an instance of the class
agg = AgglomerativeClustering(n_clusters=3,
affinity='euclidean',
linkage='ward')
Fit the instance on the data and then predict clusters for new data
agg = agg.fit(X1)
y_predict = agg.predict(X2)
Agglomerative Clustering: The Syntax
Import the class containing the clustering method
from sklearn.cluster import AgglomerativeClustering
Create an instance of the class
agg = AgglomerativeClustering(n_clusters=3,
affinity='euclidean',
linkage='ward')
Fit the instance on the data and then predict clusters for new data
agg = agg.fit(X1)
y_predict = agg.predict(X2)
final number
of clusters
Agglomerative Clustering: The Syntax
Import the class containing the clustering method
from sklearn.cluster import AgglomerativeClustering
Create an instance of the class
agg = AgglomerativeClustering(n_clusters=3,
affinity='euclidean',
linkage='ward')
Fit the instance on the data and then predict clusters for new data
agg = agg.fit(X1)
y_predict = agg.predict(X2)
cluster
affinity and
aggregation
Mini-
Batch
K-Means
Affinity
Propagatio
n
Mean
Shift
Spectral
Clustering
Ward DBSCAN
.01s 8.17s .02s .31s .21s .10s
0.1s 8.17s .03s .03s .26s .10s
.01s 8.45s .03s .04s .31s .11s
.02s 8.53s .06s .08s .21s .10s
Other Types of Clustering
Reference: http://scikit-learn.org/stable/auto_examples/cluster/plot_cluster_comparison.html
Ml9 introduction to-unsupervised_learning_and_clustering_methods

More Related Content

What's hot

Gbm.more GBM in H2O
Gbm.more GBM in H2OGbm.more GBM in H2O
Gbm.more GBM in H2O
Sri Ambati
 

What's hot (19)

Ml4 naive bayes
Ml4 naive bayesMl4 naive bayes
Ml4 naive bayes
 
Deep learning summary
Deep learning summaryDeep learning summary
Deep learning summary
 
Introduction to Boosted Trees by Tianqi Chen
Introduction to Boosted Trees by Tianqi ChenIntroduction to Boosted Trees by Tianqi Chen
Introduction to Boosted Trees by Tianqi Chen
 
Ppt shuai
Ppt shuaiPpt shuai
Ppt shuai
 
Boosted tree
Boosted treeBoosted tree
Boosted tree
 
GBM theory code and parameters
GBM theory code and parametersGBM theory code and parameters
GBM theory code and parameters
 
ICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged informationICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged information
 
Gradient Boosted Regression Trees in scikit-learn
Gradient Boosted Regression Trees in scikit-learnGradient Boosted Regression Trees in scikit-learn
Gradient Boosted Regression Trees in scikit-learn
 
Jan vitek distributedrandomforest_5-2-2013
Jan vitek distributedrandomforest_5-2-2013Jan vitek distributedrandomforest_5-2-2013
Jan vitek distributedrandomforest_5-2-2013
 
Neural Network Part-2
Neural Network Part-2Neural Network Part-2
Neural Network Part-2
 
GBM package in r
GBM package in rGBM package in r
GBM package in r
 
Nimrita deep learning
Nimrita deep learningNimrita deep learning
Nimrita deep learning
 
eam2
eam2eam2
eam2
 
Feature Engineering - Getting most out of data for predictive models - TDC 2017
Feature Engineering - Getting most out of data for predictive models - TDC 2017Feature Engineering - Getting most out of data for predictive models - TDC 2017
Feature Engineering - Getting most out of data for predictive models - TDC 2017
 
Hands-on Tutorial of Deep Learning
Hands-on Tutorial of Deep LearningHands-on Tutorial of Deep Learning
Hands-on Tutorial of Deep Learning
 
Gbm.more GBM in H2O
Gbm.more GBM in H2OGbm.more GBM in H2O
Gbm.more GBM in H2O
 
Neural Networks made easy
Neural Networks made easyNeural Networks made easy
Neural Networks made easy
 
Kaggle Higgs Boson Machine Learning Challenge
Kaggle Higgs Boson Machine Learning ChallengeKaggle Higgs Boson Machine Learning Challenge
Kaggle Higgs Boson Machine Learning Challenge
 
Tree models with Scikit-Learn: Great models with little assumptions
Tree models with Scikit-Learn: Great models with little assumptionsTree models with Scikit-Learn: Great models with little assumptions
Tree models with Scikit-Learn: Great models with little assumptions
 

Similar to Ml9 introduction to-unsupervised_learning_and_clustering_methods

Lec13 Clustering.pptx
Lec13 Clustering.pptxLec13 Clustering.pptx
Lec13 Clustering.pptx
Khalid Rabayah
 
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
Simplilearn
 
Unsupervised learning Algorithms and Assumptions
Unsupervised learning Algorithms and AssumptionsUnsupervised learning Algorithms and Assumptions
Unsupervised learning Algorithms and Assumptions
refedey275
 
Clustering (from Google)
Clustering (from Google)Clustering (from Google)
Clustering (from Google)
Sri Prasanna
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
Slideshare
 

Similar to Ml9 introduction to-unsupervised_learning_and_clustering_methods (20)

Mathematics online: some common algorithms
Mathematics online: some common algorithmsMathematics online: some common algorithms
Mathematics online: some common algorithms
 
machine learning - Clustering in R
machine learning - Clustering in Rmachine learning - Clustering in R
machine learning - Clustering in R
 
K-means Clustering
K-means ClusteringK-means Clustering
K-means Clustering
 
New Approach for K-mean and K-medoids Algorithm
New Approach for K-mean and K-medoids AlgorithmNew Approach for K-mean and K-medoids Algorithm
New Approach for K-mean and K-medoids Algorithm
 
Premeditated Initial Points for K-Means Clustering
Premeditated Initial Points for K-Means ClusteringPremeditated Initial Points for K-Means Clustering
Premeditated Initial Points for K-Means Clustering
 
CLUSTER ANALYSIS ALGORITHMS.pptx
CLUSTER ANALYSIS ALGORITHMS.pptxCLUSTER ANALYSIS ALGORITHMS.pptx
CLUSTER ANALYSIS ALGORITHMS.pptx
 
k-Means Clustering.pptx
k-Means Clustering.pptxk-Means Clustering.pptx
k-Means Clustering.pptx
 
Unsupervised learning clustering
Unsupervised learning clusteringUnsupervised learning clustering
Unsupervised learning clustering
 
Supervised and unsupervised learning
Supervised and unsupervised learningSupervised and unsupervised learning
Supervised and unsupervised learning
 
Lec4 Clustering
Lec4 ClusteringLec4 Clustering
Lec4 Clustering
 
Lec13 Clustering.pptx
Lec13 Clustering.pptxLec13 Clustering.pptx
Lec13 Clustering.pptx
 
ML basic & clustering
ML basic & clusteringML basic & clustering
ML basic & clustering
 
Cluster Analysis for Dummies
Cluster Analysis for DummiesCluster Analysis for Dummies
Cluster Analysis for Dummies
 
Lect4
Lect4Lect4
Lect4
 
Neural nw k means
Neural nw k meansNeural nw k means
Neural nw k means
 
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
K Means Clustering Algorithm | K Means Clustering Example | Machine Learning ...
 
Unsupervised learning Algorithms and Assumptions
Unsupervised learning Algorithms and AssumptionsUnsupervised learning Algorithms and Assumptions
Unsupervised learning Algorithms and Assumptions
 
Clustering (from Google)
Clustering (from Google)Clustering (from Google)
Clustering (from Google)
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
 
Log Analytics in Datacenter with Apache Spark and Machine Learning
Log Analytics in Datacenter with Apache Spark and Machine LearningLog Analytics in Datacenter with Apache Spark and Machine Learning
Log Analytics in Datacenter with Apache Spark and Machine Learning
 

More from ankit_ppt

More from ankit_ppt (20)

07 learning
07 learning07 learning
07 learning
 
06 image features
06 image features06 image features
06 image features
 
05 contours seg_matching
05 contours seg_matching05 contours seg_matching
05 contours seg_matching
 
04 image transformations_ii
04 image transformations_ii04 image transformations_ii
04 image transformations_ii
 
03 image transformations_i
03 image transformations_i03 image transformations_i
03 image transformations_i
 
02 image processing
02 image processing02 image processing
02 image processing
 
01 foundations
01 foundations01 foundations
01 foundations
 
Word2 vec
Word2 vecWord2 vec
Word2 vec
 
Text similarity measures
Text similarity measuresText similarity measures
Text similarity measures
 
Text generation and_advanced_topics
Text generation and_advanced_topicsText generation and_advanced_topics
Text generation and_advanced_topics
 
Nlp toolkits and_preprocessing_techniques
Nlp toolkits and_preprocessing_techniquesNlp toolkits and_preprocessing_techniques
Nlp toolkits and_preprocessing_techniques
 
Latent dirichlet allocation_and_topic_modeling
Latent dirichlet allocation_and_topic_modelingLatent dirichlet allocation_and_topic_modeling
Latent dirichlet allocation_and_topic_modeling
 
Intro to nlp
Intro to nlpIntro to nlp
Intro to nlp
 
Ml8 boosting and-stacking
Ml8 boosting and-stackingMl8 boosting and-stacking
Ml8 boosting and-stacking
 
Ml5 svm and-kernels
Ml5 svm and-kernelsMl5 svm and-kernels
Ml5 svm and-kernels
 
Lesson 3 ai in the enterprise
Lesson 3   ai in the enterpriseLesson 3   ai in the enterprise
Lesson 3 ai in the enterprise
 
Ml3 logistic regression-and_classification_error_metrics
Ml3 logistic regression-and_classification_error_metricsMl3 logistic regression-and_classification_error_metrics
Ml3 logistic regression-and_classification_error_metrics
 
Ml1 introduction to-supervised_learning_and_k_nearest_neighbors
Ml1 introduction to-supervised_learning_and_k_nearest_neighborsMl1 introduction to-supervised_learning_and_k_nearest_neighbors
Ml1 introduction to-supervised_learning_and_k_nearest_neighbors
 
Lesson 2 ai in industry
Lesson 2   ai in industryLesson 2   ai in industry
Lesson 2 ai in industry
 
Lesson 5 arima
Lesson 5 arimaLesson 5 arima
Lesson 5 arima
 

Recently uploaded

School management system project report.pdf
School management system project report.pdfSchool management system project report.pdf
School management system project report.pdf
Kamal Acharya
 
Hall booking system project report .pdf
Hall booking system project report  .pdfHall booking system project report  .pdf
Hall booking system project report .pdf
Kamal Acharya
 
Laundry management system project report.pdf
Laundry management system project report.pdfLaundry management system project report.pdf
Laundry management system project report.pdf
Kamal Acharya
 
Online blood donation management system project.pdf
Online blood donation management system project.pdfOnline blood donation management system project.pdf
Online blood donation management system project.pdf
Kamal Acharya
 
Fruit shop management system project report.pdf
Fruit shop management system project report.pdfFruit shop management system project report.pdf
Fruit shop management system project report.pdf
Kamal Acharya
 
Digital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdfDigital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdf
AbrahamGadissa
 

Recently uploaded (20)

2024 DevOps Pro Europe - Growing at the edge
2024 DevOps Pro Europe - Growing at the edge2024 DevOps Pro Europe - Growing at the edge
2024 DevOps Pro Europe - Growing at the edge
 
KIT-601 Lecture Notes-UNIT-5.pdf Frame Works and Visualization
KIT-601 Lecture Notes-UNIT-5.pdf Frame Works and VisualizationKIT-601 Lecture Notes-UNIT-5.pdf Frame Works and Visualization
KIT-601 Lecture Notes-UNIT-5.pdf Frame Works and Visualization
 
School management system project report.pdf
School management system project report.pdfSchool management system project report.pdf
School management system project report.pdf
 
ENERGY STORAGE DEVICES INTRODUCTION UNIT-I
ENERGY STORAGE DEVICES  INTRODUCTION UNIT-IENERGY STORAGE DEVICES  INTRODUCTION UNIT-I
ENERGY STORAGE DEVICES INTRODUCTION UNIT-I
 
Hall booking system project report .pdf
Hall booking system project report  .pdfHall booking system project report  .pdf
Hall booking system project report .pdf
 
Pharmacy management system project report..pdf
Pharmacy management system project report..pdfPharmacy management system project report..pdf
Pharmacy management system project report..pdf
 
Courier management system project report.pdf
Courier management system project report.pdfCourier management system project report.pdf
Courier management system project report.pdf
 
Natalia Rutkowska - BIM School Course in Kraków
Natalia Rutkowska - BIM School Course in KrakówNatalia Rutkowska - BIM School Course in Kraków
Natalia Rutkowska - BIM School Course in Kraków
 
Laundry management system project report.pdf
Laundry management system project report.pdfLaundry management system project report.pdf
Laundry management system project report.pdf
 
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptxCloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
Cloud-Computing_CSE311_Computer-Networking CSE GUB BD - Shahidul.pptx
 
Scaling in conventional MOSFET for constant electric field and constant voltage
Scaling in conventional MOSFET for constant electric field and constant voltageScaling in conventional MOSFET for constant electric field and constant voltage
Scaling in conventional MOSFET for constant electric field and constant voltage
 
Online blood donation management system project.pdf
Online blood donation management system project.pdfOnline blood donation management system project.pdf
Online blood donation management system project.pdf
 
Fruit shop management system project report.pdf
Fruit shop management system project report.pdfFruit shop management system project report.pdf
Fruit shop management system project report.pdf
 
shape functions of 1D and 2 D rectangular elements.pptx
shape functions of 1D and 2 D rectangular elements.pptxshape functions of 1D and 2 D rectangular elements.pptx
shape functions of 1D and 2 D rectangular elements.pptx
 
Digital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdfDigital Signal Processing Lecture notes n.pdf
Digital Signal Processing Lecture notes n.pdf
 
Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.
 
Top 13 Famous Civil Engineering Scientist
Top 13 Famous Civil Engineering ScientistTop 13 Famous Civil Engineering Scientist
Top 13 Famous Civil Engineering Scientist
 
The Benefits and Techniques of Trenchless Pipe Repair.pdf
The Benefits and Techniques of Trenchless Pipe Repair.pdfThe Benefits and Techniques of Trenchless Pipe Repair.pdf
The Benefits and Techniques of Trenchless Pipe Repair.pdf
 
İTÜ CAD and Reverse Engineering Workshop
İTÜ CAD and Reverse Engineering WorkshopİTÜ CAD and Reverse Engineering Workshop
İTÜ CAD and Reverse Engineering Workshop
 
Introduction to Machine Learning Unit-5 Notes for II-II Mechanical Engineering
Introduction to Machine Learning Unit-5 Notes for II-II Mechanical EngineeringIntroduction to Machine Learning Unit-5 Notes for II-II Mechanical Engineering
Introduction to Machine Learning Unit-5 Notes for II-II Mechanical Engineering
 

Ml9 introduction to-unsupervised_learning_and_clustering_methods