SlideShare a Scribd company logo
1 of 102
Machine Learning
Source Materials
• R. Duda, P. Hart & D. Stork, Pattern
Classification (2nd ed.), Wiley (Required)
• T. Mitchell, Machine Learning,
McGraw-Hill (Recommended)
• Papers
A Few Quotes
• “A breakthrough in machine learning would be worth
ten Microsofts” (Bill Gates, Chairman, Microsoft)
• “Machine learning is the next Internet”
(Tony Tether, Director, DARPA)
• Machine learning is the hot new thing”
(John Hennessy, President, Stanford)
• “Web rankings today are mostly a matter of machine
learning” (Prabhakar Raghavan, Dir. Research, Yahoo)
• “Machine learning is going to result in a real revolution”
(Greg Papadopoulos, CTO, Sun)
• “Machine learning is today’s discontinuity”
(Jerry Yang, CEO, Yahoo)
So, What Is Machine Learning?
• Automating automation
• Getting computers to program themselves
• Writing software is the bottleneck
• Let the data do the work instead!
Traditional Programming
Machine Learning
Computer
Data
Program
Output
Computer
Data
Output
Program
Magic?
No, more like gardening
• Seeds = Algorithms
• Nutrients = Data
• Gardener = You
• Plants = Programs
Sample Applications
• Web search
• Computational biology
• Finance
• E-commerce
• Space exploration
• Robotics
• Information extraction
• Social networks
• Debugging
• …[Your favorite area]
ML in a Nutshell
• Tens of thousands of machine learning
algorithms
• Hundreds new every year
• Every machine learning algorithm has
three components:
– Representation
– Evaluation
– Optimization
Representation
• Decision trees
• Sets of rules / Logic programs
• Instances
• Graphical models (Bayes/Markov nets)
• Neural networks
• Support vector machines
• Model ensembles
• Etc.
Evaluation
• Accuracy
• Precision and recall
• Squared error
• Likelihood
• Posterior probability
• Cost / Utility
• Margin
• Entropy
• K-L divergence
• Etc.
Optimization
• Combinatorial optimization
– E.g.: Greedy search
• Convex optimization
– E.g.: Gradient descent
• Constrained optimization
– E.g.: Linear programming
Types of Learning
• Supervised (inductive) learning
– Training data includes desired outputs
• Unsupervised learning
– Training data does not include desired outputs
• Semi-supervised learning
– Training data includes a few desired outputs
• Reinforcement learning
– Rewards from sequence of actions
Inductive Learning
• Given examples of a function (X, F(X))
• Predict function F(X) for new examples X
– Discrete F(X): Classification
– Continuous F(X): Regression
– F(X) = Probability(X): Probability estimation
What We’ll Cover
• Supervised learning
– Decision tree induction
– Rule induction
– Instance-based learning
– Bayesian learning
– Neural networks
– Support vector machines
– Model ensembles
– Learning theory
• Unsupervised learning
– Clustering
– Dimensionality reduction
ML in Practice
• Understanding domain, prior knowledge,
and goals
• Data integration, selection, cleaning,
pre-processing, etc.
• Learning models
• Interpreting results
• Consolidating and deploying discovered
knowledge
• Loop
16
Why “Learn”?
• Machine learning is programming computers to
optimize a performance criterion using example
data or past experience.
• There is no need to “learn” to calculate payroll
• Learning is used when:
– Human expertise does not exist (navigating
on Mars),
– Humans are unable to explain their expertise
(speech recognition)
– Solution changes in time (routing on a
computer network)
– Solution needs to be adapted to particular
cases (user biometrics)
17
What We Talk About When We
Talk About“Learning”
• Learning general models from a data of particular
examples
• Data is cheap and abundant (data warehouses, data
marts); knowledge is expensive and scarce.
• Example in retail: Customer transactions to consumer
behavior:
People who bought “Da Vinci Code” also bought “The
Five People You Meet in Heaven”
(www.amazon.com)
• Build a model that is a good and useful approximation to
the data.
18
Data Mining/KDD
• Retail: Market basket analysis, Customer relationship
management (CRM)
• Finance: Credit scoring, fraud detection
• Manufacturing: Optimization, troubleshooting
• Medicine: Medical diagnosis
• Telecommunications: Quality of service optimization
• Bioinformatics: Motifs, alignment
• Web mining: Search engines
• ...
Definition := “KDD is the non-trivial process of
identifying valid, novel, potentially useful, and
ultimately understandable patterns in data” (Fayyad)
Applications:
19
What is Machine Learning?
• Machine Learning
– Study of algorithms that
– improve their performance
– at some task
– with experience
• Optimize a performance criterion using example data or past
experience.
• Role of Statistics: Inference from a sample
• Role of Computer science: Efficient algorithms to
– Solve the optimization problem
– Representing and evaluating the model for inference
Growth of Machine Learning
• Machine learning is preferred approach to
– Speech recognition, Natural language processing
– Computer vision
– Medical outcomes analysis
– Robot control
– Computational biology
• This trend is accelerating
– Improved machine learning algorithms
– Improved data capture, networking, faster computers
– Software too complex to write by hand
– New sensors / IO devices
– Demand for self-customization to user, environment
– It turns out to be difficult to extract knowledge from human
expertsfailure of expert systems in the 1980’s.
20
21
Applications
• Association Analysis
• Supervised Learning
– Classification
– Regression/Prediction
• Unsupervised Learning
• Reinforcement Learning
Learning Associations
• Basket analysis:
P (Y | X ) probability that somebody who buys X
also buys Y where X and Y are
products/services.
Example: P ( chips | beer ) = 0.7
Market-Basket transactions
TID Items
1 Bread, Milk
2 Bread, Diaper, Beer, Eggs
3 Milk, Diaper, Beer, Coke
4 Bread, Milk, Diaper, Beer
5 Bread, Milk, Diaper, Coke
23
Classification
• Example:
Credit scoring
• Differentiating
between low-
risk and high-
risk customers
from their
income and
savings
Discriminant: IF income > θ1 AND savings >
θ2
THEN low-risk ELSE
high-risk
Model
24
Classification: Applications
• Aka Pattern recognition
• Face recognition: Pose, lighting, occlusion (glasses, beard), make-
up, hair style
• Character recognition: Different handwriting styles.
• Speech recognition: Temporal dependency.
– Use of a dictionary or the syntax of the language.
– Sensor fusion: Combine multiple modalities; eg, visual (lip
image) and acoustic for speech
• Medical diagnosis: From symptoms to illnesses
• Web Advertizing: Predict if a user clicks on an ad on the Internet.
25
Face Recognition
Training examples of a person
Test images
AT&T Laboratories, Cambridge UK
http://www.uk.research.att.com/facedatabase.html
26
Prediction: Regression
• Example: Price of
a used car
• x : car attributes
y : price
y = g (x | θ )
g ( ) model,
θ parameters
y = wx+w0
27
Regression Applications
• Navigating a car: Angle of the steering
wheel (CMU NavLab)
• Kinematics of a robot arm
α1= g1(x,y)
α2= g2(x,y)
α1
α2
(x,y)
28
Supervised Learning: Uses
• Prediction of future cases: Use the rule to predict
the output for future inputs
• Knowledge extraction: The rule is easy to
understand
• Compression: The rule is simpler than the data it
explains
• Outlier detection: Exceptions that are not
covered by the rule, e.g., fraud
Example: decision trees tools that create rules
29
Unsupervised Learning
• Learning “what normally happens”
• No output
• Clustering: Grouping similar instances
• Other applications: Summarization,
Association Analysis
• Example applications
– Customer segmentation in CRM
– Image compression: Color quantization
– Bioinformatics: Learning motifs
30
Reinforcement Learning
• Topics:
– Policies: what actions should an agent take in a
particular situation
– Utility estimation: how good is a state (used by
policy)
• No supervised output but delayed reward
• Credit assignment problem (what was responsible for the
outcome)
• Applications:
– Game playing
– Robot in a maze
– Multiple agents, partial observability, ...
31
Resources: Datasets
• UCI Repository:
http://www.ics.uci.edu/~mlearn/MLRepository.html
• UCI KDD Archive:
http://kdd.ics.uci.edu/summary.data.application.html
• Statlib: http://lib.stat.cmu.edu/
• Delve: http://www.cs.utoronto.ca/~delve/
32
Resources: Journals
• Journal of Machine Learning Research
www.jmlr.org
• Machine Learning
• IEEE Transactions on Neural Networks
• IEEE Transactions on Pattern Analysis
and Machine Intelligence
• Annals of Statistics
• Journal of the American Statistical
Association
33
Resources: Conferences
• International Conference on Machine Learning (ICML)
• European Conference on Machine Learning (ECML)
• Neural Information Processing Systems (NIPS)
• Computational Learning
• International Joint Conference on Artificial Intelligence (IJCAI)
• ACM SIGKDD Conference on Knowledge Discovery and Data Mining
(KDD)
• IEEE Int. Conf. on Data Mining (ICDM)
• Some Applications:
• Some Applications:
Applications of ML?
• Some Applications:
Applications of ML
• Some Applications:
Applications of ML
• Some Applications:
Applications of ML
• Some Applications:
An Example
• Suppose that:
– A fish packing plant
wants to automate the
process of sorting
incoming fish on a
conveyor belt
according to species,
– There are two
species:
• Sea bass,
• Salmon.
An Example
• How to distinguish one specie from
the other ? (length, width, weight,
number and shape of fins, tail
shape,etc.)
An Example
• Suppose somebody at the fish plant
say us that:
– Sea bass is generally longer than a
salmon
• Then our models for the fish:
– Sea bass have some typical length, and
this is greater than that for salmon.
An Example
• Then length becomes a feature,
• We might attempt to classify the fish
by seeing whether or not the length
of a fish exceeds some critical value
(threshold value) l*.
An Example
• How to desie on the critical value
(threshold value) ?
An Example
• How to desie on the critical value
(threshold value) ?
– We could obtain some training samples
of different types of fish,
– make length measurements,
– Inspect the results.
An Example
• Measurement results on the training
sample related to two species.
An Example
• Can we reliably seperate sea bass
from salmon by using length as a
feature ?
Remember our
model:
–Sea bass have
some typical
length, and
this is greater
than that for
salmon.
An Example
• From histogram we can see that
single criteria is quite poor.
An Example
• It is obvious that length is not a good
feature.
• What we can do to seperate sea
bass from salmon?
An Example
• What we can do to seperate sea
bass from salmon?
• Try another feature:
– average lightness of the fish scales.
An Example
• Can we reliably seperate sea bass
from salmon by using lightness as a
feature ?
An Example
• Lighness is better than length as a
feature but again there are some
problems.
An Example
• Suppose we also know that:
– Sea bass are typically wider than
salmon.
• We can use more than one feature
for our decision:
– Lightness (x1) and width (x2)
An Example
• Each fish is now a point in two
dimension.
– Lightness (x1) and width (x2)
An Example
• Each fish is now a point in two
dimension.
– Lightness (x1) and width (x2)
An Example
• Each fish is now a point in two
dimension.
– Lightness (x1) and width (x2)
Cost of error
• Cost of different errors must be
considered when making decisions,
• We try to make a decision rule so as
to minimize such a cost,
• This is the central task of decision
theory.
Cost of error
• For example, if the fish packing
company knows that:
– Customers who buy salmon will object
if they see sea bass in their cans.
– Customers who buy sea bass will not
be unhappy if they occasionally see
some expensive salmon in their cans.
Decision boundaries
• We can perform better if we use more
complex decision boundaries.
Decision boundaries
• There is a trade of between
complexity of the decision rules and
their performances to unknown
samples.
• Generalization: The ability of the
classifier to produce correct results
on novel patterns.
• Simplify the decision boundary!
The design cycle
The design cycle
• Collect data:
– Collect train and test data
• Choose features:
– Domain dependence and prior information,
– Computational cost and feasibility,
– Discriminative features,
– Invariant features with respect to translation, rotation and
scale,
– Robust features with respect to occlusion, distortion,
deformation, and variations in environment.
• Feature Selection & Extraction
• Selection v extraction
• How many and which subset of features to
use in constructing the decision boundary?
• Some features may be redundant
• Curse of dimensionality—Error rate may in
fact increase with too many features in the
case of small number of training samples
65
Learning
An example application
• An emergency room in a hospital
measures 17 variables (e.g., blood
pressure, age, etc) of newly admitted
patients.
• A decision is needed: whether to put a
new patient in an intensive-care unit.
• Due to the high cost of ICU, those
patients who may survive less than a
month are given higher priority.
• Problem: to predict high-risk patients and
discriminate them from low-risk patients.
66
Another application
• A credit card company receives thousands
of applications for new cards. Each
application contains information about an
applicant,
– age
– Marital status
– annual salary
– outstanding debts
– credit rating
– etc.
• Problem: to decide whether an application
should approved, or to classify
applications into two categories, approved
and not approved.
67
Machine learning and our focus
• Like human learning from past experiences.
• A computer does not have “experiences”.
• A computer system learns from data, which
represent some “past experiences” of an
application domain.
• Our focus: learn a target function that can be
used to predict the values of a discrete class
attribute, e.g., approve or not-approved, and
high-risk or low risk.
• The task is commonly called: Supervised
learning, classification, or inductive learning.
68
• Data: A set of data records (also called
examples, instances or cases) described
by
– k attributes: A1, A2, … Ak.
– a class: Each example is labelled with a pre-
defined class.
• Goal: To learn a classification model from
the data that can be used to predict the
classes of new (future, or test)
cases/instances.
The data and the goal
69
An example: data (loan application)
Approved or not
70
An example: the learning task
• Learn a classification model from the data
• Use the model to classify future loan
applications into
– Yes (approved) and
– No (not approved)
• What is the class for following
case/instance?
71
Supervised vs. unsupervised
Learning
• Supervised learning: classification is seen as
supervised learning from examples.
– Supervision: The data (observations,
measurements, etc.) are labeled with pre-
defined classes. It is like that a “teacher” gives
the classes (supervision).
– Test data are classified into these classes too.
• Unsupervised learning (clustering)
– Class labels of the data are unknown
– Given a set of data, the task is to establish the
existence of classes or clusters in the data
72
Supervised learning process:
two steps
Learning (training): Learn a model using the training data
Testing: Test the model using unseen test data to assess the model accuracy
,
cases
test
of
number
Total
tions
classifica
correct
of
Number

Accuracy
73
What do we mean by learning?
• Given
– a data set D,
– a task T, and
– a performance measure M,
a computer system is said to learn from D
to perform the task T if after learning the
system’s performance on T improves as
measured by M.
• In other words, the learned model helps
the system to perform T better as
compared to no learning.
74
An example
• Data: Loan application data
• Task: Predict whether a loan should be
approved or not.
• Performance measure: accuracy.
No learning: classify all future applications
(test data) to the majority class (i.e., Yes):
Accuracy = 9/15 = 60%.
• We can do better than 60% with learning.
75
Fundamental assumption of
learning
Assumption: The distribution of training examples
is identical to the distribution of test examples
(including future unseen examples).
• In practice, this assumption is often violated to
certain degree.
• Strong violations will clearly result in poor
classification accuracy.
• To achieve good accuracy on the test data,
training examples must be sufficiently
representative of the test data.
76
Evaluating classification methods
• Predictive accuracy
• Efficiency
– time to construct the model
– time to use the model
• Robustness: handling noise and missing values
• Scalability: efficiency in disk-resident databases
• Interpretability:
– understandable and insight provided by the model
• Compactness of the model: size of the tree, or the
number of rules.
77
Evaluation methods
• Holdout set: The available data set D is divided into two disjoint
subsets,
– the training set Dtrain (for learning a model)
– the test set Dtest (for testing the model)
• Important: training set should not be used in testing and the test set
should not be used in learning.
– Unseen test set provides a unbiased estimate of accuracy.
• The test set is also called the holdout set. (the examples in the
original data set D are all labeled with classes.)
• This method is mainly used when the data set D is large.
78
Evaluation methods (cont…)
• n-fold cross-validation: The available data is partitioned
into n equal-size disjoint subsets.
• Use each subset as the test set and combine the rest n-1
subsets as the training set to learn a classifier.
• The procedure is run n times, which give n accuracies.
• The final estimated accuracy of learning is the average of
the n accuracies.
• 10-fold and 5-fold cross-validations are commonly used.
• This method is used when the available data is not large.
79
Evaluation methods (cont…)
• Leave-one-out cross-validation: This method
is used when the data set is very small.
• It is a special case of cross-validation
• Each fold of the cross validation has only a
single test example and all the rest of the data is
used in training.
• If the original data has m examples, this is m-
fold cross-validation
80
Evaluation methods (cont…)
• Validation set: the available data is divided into three subsets,
– a training set,
– a validation set and
– a test set.
• A validation set is used frequently for estimating parameters in
learning algorithms.
• In such cases, the values that give the best accuracy on the
validation set are used as the final parameter values.
• Cross-validation can be used for parameter estimating as well.
81
Classification measures
• Accuracy is only one measure (error = 1-accuracy).
• Accuracy is not suitable in some applications.
• In text mining, we may only be interested in the documents of a
particular topic, which are only a small portion of a big document
collection.
• In classification involving skewed or highly imbalanced data, e.g.,
network intrusion and financial fraud detections, we are interested
only in the minority class.
– High accuracy does not mean any intrusion is detected.
– E.g., 1% intrusion. Achieve 99% accuracy by doing nothing.
• The class of interest is commonly called the positive class, and the
rest negative classes.
82
Precision and recall measures
• Used in information retrieval and text
classification.
• We use a confusion matrix to introduce
them.
83
Precision and recall measures
(cont…)
Precision p is the number of correctly classified positive
examples divided by the total number of examples
that are classified as positive.
Recall r is the number of correctly classified positive
examples divided by the total number of actual
positive examples in the test set.
.
.
FN
TP
TP
r
FP
TP
TP
p




84
An example
• This confusion matrix gives
– precision p = 100% and
– recall r = 1%
because we only classified one positive example correctly
and no negative examples wrongly.
• Note: precision and recall only measure
classification on the positive class.
85
F1-value (also called F1-score)
• It is hard to compare two classifiers using two measures. F1
score combines precision and recall into one measure
• The harmonic mean of two numbers tends to be closer to the
smaller of the two.
• For F1-value to be large, both p and r much be large.
Unsupervised Learning
• Definition of Unsupervised Learning:
Learning useful structure without labeled classes,
optimization criterion, feedback signal, or any
other information beyond the raw data
Unsupervised Learning
• Examples:
– Find natural groupings of Xs (X=human languages, stocks,
gene sequences, animal species,…)
Prelude to discovery of underlying properties
– Summarize the news for the past month
Cluster first, then report centroids.
– Sequence extrapolation: E.g. Predict cancer incidence next
decade; predict rise in antibiotic-resistant bacteria
• Methods
– Clustering (n-link, k-means, GAC,…)
– Taxonomy creation (hierarchical clustering)
– Novelty detection ("meaningful"outliers)
– Trend detection (extrapolation from multivariate partial
derivatives)
Similarity Measures in Data Analysis
• General Assumptions
– Each data item is a tuple (vector)
– Values of tuple are nominal, ordinal or
numerical
– Similarity = (Distance)-1
• Pure Numerical Tuples
– Sim(di,dj) = di,kdj,k
– sim (di,dj) = cos(didj)
– …and many more (slide after next)
Similarity Measures in Data Analysis
• For Ordinal Values
– E.g. "small," "medium," "large," "X-large"
– Convert to numerical assuming constant …on
a normalized [0,1] scale, where: max(v)=1,
min(v)=0, others interpolate
– E.g. "small"=0, "medium"=0.33, etc.
– Then, use numerical similarity measures
– Or, use similarity matrix (see next slide)
Similarity Measures (cont.)
• For Nominal Values
– E.g. "Boston", "LA", "Pittsburgh", or "male",
"female", or "diffuse", "globular", "spiral",
"pinwheel"
– Binary rule: If di,k=dj,k, then sim=1, else 0
– Use underlying sematic property: E.g.
Sim(Boston, LA)=dist(Boston, LA)-1, or
Sim(Boston, LA)=(|size(Boston) – size(LA)|
)-1
– Use similarity Matrix
Similarity Matrix
tiny little small medium large huge
tiny 1.0 0.8 0.7 0.5 0.2 0.0
little 1.0 0.9 0.7 0.3 0.1
small 1.0 0.7 0.3 0.2
medium 1.0 0.5 0.3
large 1.0 0.8
huge 1.0
– Diagonal must be 1.0
– Monotonicity property must hold
– Triangle inequality must hold
– Transitive property need *not* hold
Semi-Supervised Learning
Supervised Learning = learning from labeled
data. Dominant paradigm in Machine Learning.
• E.g, say you want to train an email classifier to
distinguish spam from important messages
Semi-Supervised Learning
Supervised Learning = learning from labeled
data. Dominant paradigm in Machine Learning.
• E.g, say you want to train an email classifier to
distinguish spam from important messages
• Take sample S of data, labeled according to
whether they were/weren’t spam.
Semi-Supervised Learning
Supervised Learning = learning from labeled
data. Dominant paradigm in Machine Learning.
• E.g, say you want to train an email classifier to
distinguish spam from important messages
• Take sample S of data, labeled according to
whether they were/weren’t spam.
• Train a classifier (like SVM, decision tree, etc)
on S. Make sure it’s not overfitting.
• Use to classify new emails.
Basic paradigm has many successes
• recognize speech,
• steer a car,
• classify documents
• classify proteins
• recognizing faces, objects in images
• ...
However, for many problems, labeled
data can be rare or expensive.
Unlabeled data is much cheaper.
Need to pay someone to do it, requires special testing,…
However, for many problems, labeled
data can be rare or expensive.
Unlabeled data is much cheaper.
Speech
Images
Medical outcomes
Customer modeling
Protein sequences
Web pages
Need to pay someone to do it, requires special testing,…
However, for many problems, labeled
data can be rare or expensive.
Unlabeled data is much cheaper.
[From Jerry Zhu]
Need to pay someone to do it, requires special testing,…
Need to pay someone to do it, requires special testing,…
However, for many problems, labeled
data can be rare or expensive.
Unlabeled data is much cheaper.
Can we make use of cheap
unlabeled data?
Semi-Supervised Learning
Can we use unlabeled data to augment a
small labeled sample to improve learning?
But unlabeled data
is missing the
most important
info!!
But maybe still has
useful regularities
that we can use.
But… But… But…
Semi-Supervised Learning
Substantial recent work in ML. A number of
interesting methods have been developed.
• Several diverse methods for taking advantage
of unlabeled data.
• General framework to understand when
unlabeled data can help, and make sense of
what’s going on.
References
• Srihari, S.N., Covindaraju, Pattern recognition, Chapman &Hall, London, 1034-1041, 1993,
• Sergios Theodoridis, Konstantinos Koutroumbas , pattern recognition , Pattern Recognition
,Elsevier(USA)) ,1982
• R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, New York: John Wiley, 2001,

More Related Content

Similar to intro to ML by the way m toh phasee movie Punjabi

Machine Learning: Applications, Process and Techniques
Machine Learning: Applications, Process and TechniquesMachine Learning: Applications, Process and Techniques
Machine Learning: Applications, Process and TechniquesRui Pedro Paiva
 
Intro to Data Science for Non-Data Scientists
Intro to Data Science for Non-Data ScientistsIntro to Data Science for Non-Data Scientists
Intro to Data Science for Non-Data ScientistsSri Ambati
 
Machine Learning in Modern Medicine with Erin LeDell at Stanford Med
Machine Learning in Modern Medicine with Erin LeDell at Stanford MedMachine Learning in Modern Medicine with Erin LeDell at Stanford Med
Machine Learning in Modern Medicine with Erin LeDell at Stanford MedSri Ambati
 
2016 03-16 digital energy luncheon
2016 03-16 digital energy luncheon2016 03-16 digital energy luncheon
2016 03-16 digital energy luncheonMark Reynolds
 
Essential concepts for machine learning
Essential concepts for machine learning Essential concepts for machine learning
Essential concepts for machine learning pyingkodi maran
 
Introduction to Machine Learning - An overview and first step for candidate d...
Introduction to Machine Learning - An overview and first step for candidate d...Introduction to Machine Learning - An overview and first step for candidate d...
Introduction to Machine Learning - An overview and first step for candidate d...Lucas Jellema
 
ML Basic Concepts.pdf
ML Basic Concepts.pdfML Basic Concepts.pdf
ML Basic Concepts.pdfManishaS49
 
Overview of Machine Learning and its Applications
Overview of Machine Learning and its ApplicationsOverview of Machine Learning and its Applications
Overview of Machine Learning and its ApplicationsDeepak Chawla
 
ML MODULE 1_slideshare.pdf
ML MODULE 1_slideshare.pdfML MODULE 1_slideshare.pdf
ML MODULE 1_slideshare.pdfShiwani Gupta
 
Data Mining - The Big Picture!
Data Mining - The Big Picture!Data Mining - The Big Picture!
Data Mining - The Big Picture!Khalid Salama
 
Introduction to machine_learning
Introduction to machine_learningIntroduction to machine_learning
Introduction to machine_learningKiran Lonikar
 
Introduction to Machine Learning.pptx
Introduction to Machine Learning.pptxIntroduction to Machine Learning.pptx
Introduction to Machine Learning.pptxDr. Amanpreet Kaur
 
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...The Art of Intelligence – A Practical Introduction Machine Learning for Orac...
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...Lucas Jellema
 
1 - Introduction to Machine Learning.pdf
1 - Introduction to Machine Learning.pdf1 - Introduction to Machine Learning.pdf
1 - Introduction to Machine Learning.pdfkhaiziliyahkhalid
 
Webinar: Machine Learning para Microcontroladores
Webinar: Machine Learning para MicrocontroladoresWebinar: Machine Learning para Microcontroladores
Webinar: Machine Learning para MicrocontroladoresEmbarcados
 
influence of AI in IS
influence of AI in ISinfluence of AI in IS
influence of AI in ISISACA Riyadh
 

Similar to intro to ML by the way m toh phasee movie Punjabi (20)

Machine Learning: Applications, Process and Techniques
Machine Learning: Applications, Process and TechniquesMachine Learning: Applications, Process and Techniques
Machine Learning: Applications, Process and Techniques
 
Intro to Data Science for Non-Data Scientists
Intro to Data Science for Non-Data ScientistsIntro to Data Science for Non-Data Scientists
Intro to Data Science for Non-Data Scientists
 
Machine Learning in Modern Medicine with Erin LeDell at Stanford Med
Machine Learning in Modern Medicine with Erin LeDell at Stanford MedMachine Learning in Modern Medicine with Erin LeDell at Stanford Med
Machine Learning in Modern Medicine with Erin LeDell at Stanford Med
 
2016 03-16 digital energy luncheon
2016 03-16 digital energy luncheon2016 03-16 digital energy luncheon
2016 03-16 digital energy luncheon
 
Essential concepts for machine learning
Essential concepts for machine learning Essential concepts for machine learning
Essential concepts for machine learning
 
Knowledge Discovery
Knowledge DiscoveryKnowledge Discovery
Knowledge Discovery
 
Introduction to Machine Learning - An overview and first step for candidate d...
Introduction to Machine Learning - An overview and first step for candidate d...Introduction to Machine Learning - An overview and first step for candidate d...
Introduction to Machine Learning - An overview and first step for candidate d...
 
ML Basic Concepts.pdf
ML Basic Concepts.pdfML Basic Concepts.pdf
ML Basic Concepts.pdf
 
Overview of Machine Learning and its Applications
Overview of Machine Learning and its ApplicationsOverview of Machine Learning and its Applications
Overview of Machine Learning and its Applications
 
ML MODULE 1_slideshare.pdf
ML MODULE 1_slideshare.pdfML MODULE 1_slideshare.pdf
ML MODULE 1_slideshare.pdf
 
Data Mining - The Big Picture!
Data Mining - The Big Picture!Data Mining - The Big Picture!
Data Mining - The Big Picture!
 
Week 2 lecture
Week 2 lectureWeek 2 lecture
Week 2 lecture
 
Introduction to machine_learning
Introduction to machine_learningIntroduction to machine_learning
Introduction to machine_learning
 
Introduction to Machine Learning.pptx
Introduction to Machine Learning.pptxIntroduction to Machine Learning.pptx
Introduction to Machine Learning.pptx
 
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...The Art of Intelligence – A Practical Introduction Machine Learning for Orac...
The Art of Intelligence – A Practical Introduction Machine Learning for Orac...
 
1 - Introduction to Machine Learning.pdf
1 - Introduction to Machine Learning.pdf1 - Introduction to Machine Learning.pdf
1 - Introduction to Machine Learning.pdf
 
Webinar: Machine Learning para Microcontroladores
Webinar: Machine Learning para MicrocontroladoresWebinar: Machine Learning para Microcontroladores
Webinar: Machine Learning para Microcontroladores
 
influence of AI in IS
influence of AI in ISinfluence of AI in IS
influence of AI in IS
 
Data-X-Sparse-v2
Data-X-Sparse-v2Data-X-Sparse-v2
Data-X-Sparse-v2
 
Data-X-v3.1
Data-X-v3.1Data-X-v3.1
Data-X-v3.1
 

Recently uploaded

Call Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceCall Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceSapana Sha
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...ThinkInnovation
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一F sss
 
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一F La
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Callshivangimorya083
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubaihf8803863
 
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfSocial Samosa
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingNeil Barnes
 
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Jack DiGiovanna
 
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...dajasot375
 
GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]📊 Markus Baersch
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样vhwb25kk
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...Suhani Kapoor
 
RadioAdProWritingCinderellabyButleri.pdf
RadioAdProWritingCinderellabyButleri.pdfRadioAdProWritingCinderellabyButleri.pdf
RadioAdProWritingCinderellabyButleri.pdfgstagge
 
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAmazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAbdelrhman abooda
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home ServiceSapana Sha
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFAAndrei Kaleshka
 

Recently uploaded (20)

Call Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceCall Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts Service
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
 
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...
Predictive Analysis - Using Insight-informed Data to Determine Factors Drivin...
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
 
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
 
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
 
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data Storytelling
 
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
 
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...
Indian Call Girls in Abu Dhabi O5286O24O8 Call Girls in Abu Dhabi By Independ...
 
GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
 
RadioAdProWritingCinderellabyButleri.pdf
RadioAdProWritingCinderellabyButleri.pdfRadioAdProWritingCinderellabyButleri.pdf
RadioAdProWritingCinderellabyButleri.pdf
 
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAmazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service
 
E-Commerce Order PredictionShraddha Kamble.pptx
E-Commerce Order PredictionShraddha Kamble.pptxE-Commerce Order PredictionShraddha Kamble.pptx
E-Commerce Order PredictionShraddha Kamble.pptx
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFA
 

intro to ML by the way m toh phasee movie Punjabi

  • 2. Source Materials • R. Duda, P. Hart & D. Stork, Pattern Classification (2nd ed.), Wiley (Required) • T. Mitchell, Machine Learning, McGraw-Hill (Recommended) • Papers
  • 3. A Few Quotes • “A breakthrough in machine learning would be worth ten Microsofts” (Bill Gates, Chairman, Microsoft) • “Machine learning is the next Internet” (Tony Tether, Director, DARPA) • Machine learning is the hot new thing” (John Hennessy, President, Stanford) • “Web rankings today are mostly a matter of machine learning” (Prabhakar Raghavan, Dir. Research, Yahoo) • “Machine learning is going to result in a real revolution” (Greg Papadopoulos, CTO, Sun) • “Machine learning is today’s discontinuity” (Jerry Yang, CEO, Yahoo)
  • 4. So, What Is Machine Learning? • Automating automation • Getting computers to program themselves • Writing software is the bottleneck • Let the data do the work instead!
  • 6. Magic? No, more like gardening • Seeds = Algorithms • Nutrients = Data • Gardener = You • Plants = Programs
  • 7. Sample Applications • Web search • Computational biology • Finance • E-commerce • Space exploration • Robotics • Information extraction • Social networks • Debugging • …[Your favorite area]
  • 8. ML in a Nutshell • Tens of thousands of machine learning algorithms • Hundreds new every year • Every machine learning algorithm has three components: – Representation – Evaluation – Optimization
  • 9. Representation • Decision trees • Sets of rules / Logic programs • Instances • Graphical models (Bayes/Markov nets) • Neural networks • Support vector machines • Model ensembles • Etc.
  • 10. Evaluation • Accuracy • Precision and recall • Squared error • Likelihood • Posterior probability • Cost / Utility • Margin • Entropy • K-L divergence • Etc.
  • 11. Optimization • Combinatorial optimization – E.g.: Greedy search • Convex optimization – E.g.: Gradient descent • Constrained optimization – E.g.: Linear programming
  • 12. Types of Learning • Supervised (inductive) learning – Training data includes desired outputs • Unsupervised learning – Training data does not include desired outputs • Semi-supervised learning – Training data includes a few desired outputs • Reinforcement learning – Rewards from sequence of actions
  • 13. Inductive Learning • Given examples of a function (X, F(X)) • Predict function F(X) for new examples X – Discrete F(X): Classification – Continuous F(X): Regression – F(X) = Probability(X): Probability estimation
  • 14. What We’ll Cover • Supervised learning – Decision tree induction – Rule induction – Instance-based learning – Bayesian learning – Neural networks – Support vector machines – Model ensembles – Learning theory • Unsupervised learning – Clustering – Dimensionality reduction
  • 15. ML in Practice • Understanding domain, prior knowledge, and goals • Data integration, selection, cleaning, pre-processing, etc. • Learning models • Interpreting results • Consolidating and deploying discovered knowledge • Loop
  • 16. 16 Why “Learn”? • Machine learning is programming computers to optimize a performance criterion using example data or past experience. • There is no need to “learn” to calculate payroll • Learning is used when: – Human expertise does not exist (navigating on Mars), – Humans are unable to explain their expertise (speech recognition) – Solution changes in time (routing on a computer network) – Solution needs to be adapted to particular cases (user biometrics)
  • 17. 17 What We Talk About When We Talk About“Learning” • Learning general models from a data of particular examples • Data is cheap and abundant (data warehouses, data marts); knowledge is expensive and scarce. • Example in retail: Customer transactions to consumer behavior: People who bought “Da Vinci Code” also bought “The Five People You Meet in Heaven” (www.amazon.com) • Build a model that is a good and useful approximation to the data.
  • 18. 18 Data Mining/KDD • Retail: Market basket analysis, Customer relationship management (CRM) • Finance: Credit scoring, fraud detection • Manufacturing: Optimization, troubleshooting • Medicine: Medical diagnosis • Telecommunications: Quality of service optimization • Bioinformatics: Motifs, alignment • Web mining: Search engines • ... Definition := “KDD is the non-trivial process of identifying valid, novel, potentially useful, and ultimately understandable patterns in data” (Fayyad) Applications:
  • 19. 19 What is Machine Learning? • Machine Learning – Study of algorithms that – improve their performance – at some task – with experience • Optimize a performance criterion using example data or past experience. • Role of Statistics: Inference from a sample • Role of Computer science: Efficient algorithms to – Solve the optimization problem – Representing and evaluating the model for inference
  • 20. Growth of Machine Learning • Machine learning is preferred approach to – Speech recognition, Natural language processing – Computer vision – Medical outcomes analysis – Robot control – Computational biology • This trend is accelerating – Improved machine learning algorithms – Improved data capture, networking, faster computers – Software too complex to write by hand – New sensors / IO devices – Demand for self-customization to user, environment – It turns out to be difficult to extract knowledge from human expertsfailure of expert systems in the 1980’s. 20
  • 21. 21 Applications • Association Analysis • Supervised Learning – Classification – Regression/Prediction • Unsupervised Learning • Reinforcement Learning
  • 22. Learning Associations • Basket analysis: P (Y | X ) probability that somebody who buys X also buys Y where X and Y are products/services. Example: P ( chips | beer ) = 0.7 Market-Basket transactions TID Items 1 Bread, Milk 2 Bread, Diaper, Beer, Eggs 3 Milk, Diaper, Beer, Coke 4 Bread, Milk, Diaper, Beer 5 Bread, Milk, Diaper, Coke
  • 23. 23 Classification • Example: Credit scoring • Differentiating between low- risk and high- risk customers from their income and savings Discriminant: IF income > θ1 AND savings > θ2 THEN low-risk ELSE high-risk Model
  • 24. 24 Classification: Applications • Aka Pattern recognition • Face recognition: Pose, lighting, occlusion (glasses, beard), make- up, hair style • Character recognition: Different handwriting styles. • Speech recognition: Temporal dependency. – Use of a dictionary or the syntax of the language. – Sensor fusion: Combine multiple modalities; eg, visual (lip image) and acoustic for speech • Medical diagnosis: From symptoms to illnesses • Web Advertizing: Predict if a user clicks on an ad on the Internet.
  • 25. 25 Face Recognition Training examples of a person Test images AT&T Laboratories, Cambridge UK http://www.uk.research.att.com/facedatabase.html
  • 26. 26 Prediction: Regression • Example: Price of a used car • x : car attributes y : price y = g (x | θ ) g ( ) model, θ parameters y = wx+w0
  • 27. 27 Regression Applications • Navigating a car: Angle of the steering wheel (CMU NavLab) • Kinematics of a robot arm α1= g1(x,y) α2= g2(x,y) α1 α2 (x,y)
  • 28. 28 Supervised Learning: Uses • Prediction of future cases: Use the rule to predict the output for future inputs • Knowledge extraction: The rule is easy to understand • Compression: The rule is simpler than the data it explains • Outlier detection: Exceptions that are not covered by the rule, e.g., fraud Example: decision trees tools that create rules
  • 29. 29 Unsupervised Learning • Learning “what normally happens” • No output • Clustering: Grouping similar instances • Other applications: Summarization, Association Analysis • Example applications – Customer segmentation in CRM – Image compression: Color quantization – Bioinformatics: Learning motifs
  • 30. 30 Reinforcement Learning • Topics: – Policies: what actions should an agent take in a particular situation – Utility estimation: how good is a state (used by policy) • No supervised output but delayed reward • Credit assignment problem (what was responsible for the outcome) • Applications: – Game playing – Robot in a maze – Multiple agents, partial observability, ...
  • 31. 31 Resources: Datasets • UCI Repository: http://www.ics.uci.edu/~mlearn/MLRepository.html • UCI KDD Archive: http://kdd.ics.uci.edu/summary.data.application.html • Statlib: http://lib.stat.cmu.edu/ • Delve: http://www.cs.utoronto.ca/~delve/
  • 32. 32 Resources: Journals • Journal of Machine Learning Research www.jmlr.org • Machine Learning • IEEE Transactions on Neural Networks • IEEE Transactions on Pattern Analysis and Machine Intelligence • Annals of Statistics • Journal of the American Statistical Association
  • 33. 33 Resources: Conferences • International Conference on Machine Learning (ICML) • European Conference on Machine Learning (ECML) • Neural Information Processing Systems (NIPS) • Computational Learning • International Joint Conference on Artificial Intelligence (IJCAI) • ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD) • IEEE Int. Conf. on Data Mining (ICDM)
  • 36. Applications of ML? • Some Applications:
  • 37. Applications of ML • Some Applications:
  • 38. Applications of ML • Some Applications:
  • 39. Applications of ML • Some Applications:
  • 40. An Example • Suppose that: – A fish packing plant wants to automate the process of sorting incoming fish on a conveyor belt according to species, – There are two species: • Sea bass, • Salmon.
  • 41. An Example • How to distinguish one specie from the other ? (length, width, weight, number and shape of fins, tail shape,etc.)
  • 42. An Example • Suppose somebody at the fish plant say us that: – Sea bass is generally longer than a salmon • Then our models for the fish: – Sea bass have some typical length, and this is greater than that for salmon.
  • 43. An Example • Then length becomes a feature, • We might attempt to classify the fish by seeing whether or not the length of a fish exceeds some critical value (threshold value) l*.
  • 44. An Example • How to desie on the critical value (threshold value) ?
  • 45. An Example • How to desie on the critical value (threshold value) ? – We could obtain some training samples of different types of fish, – make length measurements, – Inspect the results.
  • 46. An Example • Measurement results on the training sample related to two species.
  • 47. An Example • Can we reliably seperate sea bass from salmon by using length as a feature ? Remember our model: –Sea bass have some typical length, and this is greater than that for salmon.
  • 48. An Example • From histogram we can see that single criteria is quite poor.
  • 49. An Example • It is obvious that length is not a good feature. • What we can do to seperate sea bass from salmon?
  • 50. An Example • What we can do to seperate sea bass from salmon? • Try another feature: – average lightness of the fish scales.
  • 51. An Example • Can we reliably seperate sea bass from salmon by using lightness as a feature ?
  • 52. An Example • Lighness is better than length as a feature but again there are some problems.
  • 53. An Example • Suppose we also know that: – Sea bass are typically wider than salmon. • We can use more than one feature for our decision: – Lightness (x1) and width (x2)
  • 54. An Example • Each fish is now a point in two dimension. – Lightness (x1) and width (x2)
  • 55. An Example • Each fish is now a point in two dimension. – Lightness (x1) and width (x2)
  • 56. An Example • Each fish is now a point in two dimension. – Lightness (x1) and width (x2)
  • 57. Cost of error • Cost of different errors must be considered when making decisions, • We try to make a decision rule so as to minimize such a cost, • This is the central task of decision theory.
  • 58. Cost of error • For example, if the fish packing company knows that: – Customers who buy salmon will object if they see sea bass in their cans. – Customers who buy sea bass will not be unhappy if they occasionally see some expensive salmon in their cans.
  • 59. Decision boundaries • We can perform better if we use more complex decision boundaries.
  • 60. Decision boundaries • There is a trade of between complexity of the decision rules and their performances to unknown samples. • Generalization: The ability of the classifier to produce correct results on novel patterns. • Simplify the decision boundary!
  • 62. The design cycle • Collect data: – Collect train and test data • Choose features: – Domain dependence and prior information, – Computational cost and feasibility, – Discriminative features, – Invariant features with respect to translation, rotation and scale, – Robust features with respect to occlusion, distortion, deformation, and variations in environment.
  • 63. • Feature Selection & Extraction • Selection v extraction • How many and which subset of features to use in constructing the decision boundary? • Some features may be redundant • Curse of dimensionality—Error rate may in fact increase with too many features in the case of small number of training samples
  • 64.
  • 65. 65 Learning An example application • An emergency room in a hospital measures 17 variables (e.g., blood pressure, age, etc) of newly admitted patients. • A decision is needed: whether to put a new patient in an intensive-care unit. • Due to the high cost of ICU, those patients who may survive less than a month are given higher priority. • Problem: to predict high-risk patients and discriminate them from low-risk patients.
  • 66. 66 Another application • A credit card company receives thousands of applications for new cards. Each application contains information about an applicant, – age – Marital status – annual salary – outstanding debts – credit rating – etc. • Problem: to decide whether an application should approved, or to classify applications into two categories, approved and not approved.
  • 67. 67 Machine learning and our focus • Like human learning from past experiences. • A computer does not have “experiences”. • A computer system learns from data, which represent some “past experiences” of an application domain. • Our focus: learn a target function that can be used to predict the values of a discrete class attribute, e.g., approve or not-approved, and high-risk or low risk. • The task is commonly called: Supervised learning, classification, or inductive learning.
  • 68. 68 • Data: A set of data records (also called examples, instances or cases) described by – k attributes: A1, A2, … Ak. – a class: Each example is labelled with a pre- defined class. • Goal: To learn a classification model from the data that can be used to predict the classes of new (future, or test) cases/instances. The data and the goal
  • 69. 69 An example: data (loan application) Approved or not
  • 70. 70 An example: the learning task • Learn a classification model from the data • Use the model to classify future loan applications into – Yes (approved) and – No (not approved) • What is the class for following case/instance?
  • 71. 71 Supervised vs. unsupervised Learning • Supervised learning: classification is seen as supervised learning from examples. – Supervision: The data (observations, measurements, etc.) are labeled with pre- defined classes. It is like that a “teacher” gives the classes (supervision). – Test data are classified into these classes too. • Unsupervised learning (clustering) – Class labels of the data are unknown – Given a set of data, the task is to establish the existence of classes or clusters in the data
  • 72. 72 Supervised learning process: two steps Learning (training): Learn a model using the training data Testing: Test the model using unseen test data to assess the model accuracy , cases test of number Total tions classifica correct of Number  Accuracy
  • 73. 73 What do we mean by learning? • Given – a data set D, – a task T, and – a performance measure M, a computer system is said to learn from D to perform the task T if after learning the system’s performance on T improves as measured by M. • In other words, the learned model helps the system to perform T better as compared to no learning.
  • 74. 74 An example • Data: Loan application data • Task: Predict whether a loan should be approved or not. • Performance measure: accuracy. No learning: classify all future applications (test data) to the majority class (i.e., Yes): Accuracy = 9/15 = 60%. • We can do better than 60% with learning.
  • 75. 75 Fundamental assumption of learning Assumption: The distribution of training examples is identical to the distribution of test examples (including future unseen examples). • In practice, this assumption is often violated to certain degree. • Strong violations will clearly result in poor classification accuracy. • To achieve good accuracy on the test data, training examples must be sufficiently representative of the test data.
  • 76. 76 Evaluating classification methods • Predictive accuracy • Efficiency – time to construct the model – time to use the model • Robustness: handling noise and missing values • Scalability: efficiency in disk-resident databases • Interpretability: – understandable and insight provided by the model • Compactness of the model: size of the tree, or the number of rules.
  • 77. 77 Evaluation methods • Holdout set: The available data set D is divided into two disjoint subsets, – the training set Dtrain (for learning a model) – the test set Dtest (for testing the model) • Important: training set should not be used in testing and the test set should not be used in learning. – Unseen test set provides a unbiased estimate of accuracy. • The test set is also called the holdout set. (the examples in the original data set D are all labeled with classes.) • This method is mainly used when the data set D is large.
  • 78. 78 Evaluation methods (cont…) • n-fold cross-validation: The available data is partitioned into n equal-size disjoint subsets. • Use each subset as the test set and combine the rest n-1 subsets as the training set to learn a classifier. • The procedure is run n times, which give n accuracies. • The final estimated accuracy of learning is the average of the n accuracies. • 10-fold and 5-fold cross-validations are commonly used. • This method is used when the available data is not large.
  • 79. 79 Evaluation methods (cont…) • Leave-one-out cross-validation: This method is used when the data set is very small. • It is a special case of cross-validation • Each fold of the cross validation has only a single test example and all the rest of the data is used in training. • If the original data has m examples, this is m- fold cross-validation
  • 80. 80 Evaluation methods (cont…) • Validation set: the available data is divided into three subsets, – a training set, – a validation set and – a test set. • A validation set is used frequently for estimating parameters in learning algorithms. • In such cases, the values that give the best accuracy on the validation set are used as the final parameter values. • Cross-validation can be used for parameter estimating as well.
  • 81. 81 Classification measures • Accuracy is only one measure (error = 1-accuracy). • Accuracy is not suitable in some applications. • In text mining, we may only be interested in the documents of a particular topic, which are only a small portion of a big document collection. • In classification involving skewed or highly imbalanced data, e.g., network intrusion and financial fraud detections, we are interested only in the minority class. – High accuracy does not mean any intrusion is detected. – E.g., 1% intrusion. Achieve 99% accuracy by doing nothing. • The class of interest is commonly called the positive class, and the rest negative classes.
  • 82. 82 Precision and recall measures • Used in information retrieval and text classification. • We use a confusion matrix to introduce them.
  • 83. 83 Precision and recall measures (cont…) Precision p is the number of correctly classified positive examples divided by the total number of examples that are classified as positive. Recall r is the number of correctly classified positive examples divided by the total number of actual positive examples in the test set. . . FN TP TP r FP TP TP p    
  • 84. 84 An example • This confusion matrix gives – precision p = 100% and – recall r = 1% because we only classified one positive example correctly and no negative examples wrongly. • Note: precision and recall only measure classification on the positive class.
  • 85. 85 F1-value (also called F1-score) • It is hard to compare two classifiers using two measures. F1 score combines precision and recall into one measure • The harmonic mean of two numbers tends to be closer to the smaller of the two. • For F1-value to be large, both p and r much be large.
  • 86. Unsupervised Learning • Definition of Unsupervised Learning: Learning useful structure without labeled classes, optimization criterion, feedback signal, or any other information beyond the raw data
  • 87. Unsupervised Learning • Examples: – Find natural groupings of Xs (X=human languages, stocks, gene sequences, animal species,…) Prelude to discovery of underlying properties – Summarize the news for the past month Cluster first, then report centroids. – Sequence extrapolation: E.g. Predict cancer incidence next decade; predict rise in antibiotic-resistant bacteria • Methods – Clustering (n-link, k-means, GAC,…) – Taxonomy creation (hierarchical clustering) – Novelty detection ("meaningful"outliers) – Trend detection (extrapolation from multivariate partial derivatives)
  • 88. Similarity Measures in Data Analysis • General Assumptions – Each data item is a tuple (vector) – Values of tuple are nominal, ordinal or numerical – Similarity = (Distance)-1 • Pure Numerical Tuples – Sim(di,dj) = di,kdj,k – sim (di,dj) = cos(didj) – …and many more (slide after next)
  • 89. Similarity Measures in Data Analysis • For Ordinal Values – E.g. "small," "medium," "large," "X-large" – Convert to numerical assuming constant …on a normalized [0,1] scale, where: max(v)=1, min(v)=0, others interpolate – E.g. "small"=0, "medium"=0.33, etc. – Then, use numerical similarity measures – Or, use similarity matrix (see next slide)
  • 90. Similarity Measures (cont.) • For Nominal Values – E.g. "Boston", "LA", "Pittsburgh", or "male", "female", or "diffuse", "globular", "spiral", "pinwheel" – Binary rule: If di,k=dj,k, then sim=1, else 0 – Use underlying sematic property: E.g. Sim(Boston, LA)=dist(Boston, LA)-1, or Sim(Boston, LA)=(|size(Boston) – size(LA)| )-1 – Use similarity Matrix
  • 91. Similarity Matrix tiny little small medium large huge tiny 1.0 0.8 0.7 0.5 0.2 0.0 little 1.0 0.9 0.7 0.3 0.1 small 1.0 0.7 0.3 0.2 medium 1.0 0.5 0.3 large 1.0 0.8 huge 1.0 – Diagonal must be 1.0 – Monotonicity property must hold – Triangle inequality must hold – Transitive property need *not* hold
  • 92. Semi-Supervised Learning Supervised Learning = learning from labeled data. Dominant paradigm in Machine Learning. • E.g, say you want to train an email classifier to distinguish spam from important messages
  • 93. Semi-Supervised Learning Supervised Learning = learning from labeled data. Dominant paradigm in Machine Learning. • E.g, say you want to train an email classifier to distinguish spam from important messages • Take sample S of data, labeled according to whether they were/weren’t spam.
  • 94. Semi-Supervised Learning Supervised Learning = learning from labeled data. Dominant paradigm in Machine Learning. • E.g, say you want to train an email classifier to distinguish spam from important messages • Take sample S of data, labeled according to whether they were/weren’t spam. • Train a classifier (like SVM, decision tree, etc) on S. Make sure it’s not overfitting. • Use to classify new emails.
  • 95. Basic paradigm has many successes • recognize speech, • steer a car, • classify documents • classify proteins • recognizing faces, objects in images • ...
  • 96. However, for many problems, labeled data can be rare or expensive. Unlabeled data is much cheaper. Need to pay someone to do it, requires special testing,…
  • 97. However, for many problems, labeled data can be rare or expensive. Unlabeled data is much cheaper. Speech Images Medical outcomes Customer modeling Protein sequences Web pages Need to pay someone to do it, requires special testing,…
  • 98. However, for many problems, labeled data can be rare or expensive. Unlabeled data is much cheaper. [From Jerry Zhu] Need to pay someone to do it, requires special testing,…
  • 99. Need to pay someone to do it, requires special testing,… However, for many problems, labeled data can be rare or expensive. Unlabeled data is much cheaper. Can we make use of cheap unlabeled data?
  • 100. Semi-Supervised Learning Can we use unlabeled data to augment a small labeled sample to improve learning? But unlabeled data is missing the most important info!! But maybe still has useful regularities that we can use. But… But… But…
  • 101. Semi-Supervised Learning Substantial recent work in ML. A number of interesting methods have been developed. • Several diverse methods for taking advantage of unlabeled data. • General framework to understand when unlabeled data can help, and make sense of what’s going on.
  • 102. References • Srihari, S.N., Covindaraju, Pattern recognition, Chapman &Hall, London, 1034-1041, 1993, • Sergios Theodoridis, Konstantinos Koutroumbas , pattern recognition , Pattern Recognition ,Elsevier(USA)) ,1982 • R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, New York: John Wiley, 2001,