This document provides an overview of Linear Discriminant Analysis (LDA) for dimensionality reduction. LDA seeks to perform dimensionality reduction while preserving class discriminatory information as much as possible, unlike PCA which does not consider class labels. LDA finds a linear combination of features that separates classes best by maximizing the between-class variance while minimizing the within-class variance. This is achieved by solving the generalized eigenvalue problem involving the within-class and between-class scatter matrices. The document provides mathematical details and an example to illustrate LDA for a two-class problem.
Linear Regression vs Logistic Regression | EdurekaEdureka!
YouTube: https://youtu.be/OCwZyYH14uw
** Data Science Certification using R: https://www.edureka.co/data-science **
This Edureka PPT on Linear Regression Vs Logistic Regression covers the basic concepts of linear and logistic models. The following topics are covered in this session:
Types of Machine Learning
Regression Vs Classification
What is Linear Regression?
What is Logistic Regression?
Linear Regression Use Case
Logistic Regression Use Case
Linear Regression Vs Logistic Regression
Blog Series: http://bit.ly/data-science-blogs
Data Science Training Playlist: http://bit.ly/data-science-playlist
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Logistic Regression in Python | Logistic Regression Example | Machine Learnin...Edureka!
** Python Data Science Training : https://www.edureka.co/python **
This Edureka Video on Logistic Regression in Python will give you basic understanding of Logistic Regression Machine Learning Algorithm with examples. In this video, you will also get to see demo on Logistic Regression using Python. Below are the topics covered in this tutorial:
1. What is Regression?
2. What is Logistic Regression?
3. Why use Logistic Regression?
4. Linear vs Logistic Regression
5. Logistic Regression Use Cases
6. Logistic Regression Example Demo in Python
Subscribe to our channel to get video updates. Hit the subscribe button above.
Machine Learning Tutorial Playlist: https://goo.gl/UxjTxm
Logistic regression : Use Case | Background | Advantages | DisadvantagesRajat Sharma
This slide will help you to understand the working of logistic regression which is a type of machine learning model along with use cases, pros and cons.
Presentation at Advanced Intelligent Systems for Sustainable Development (AISSD 2021) 20-22 August 2021 organized by the scientific research group in Egypt with Collaboration with Faculty of Computers and AI, Cairo University and the Chinese University in Egypt
Linear Regression vs Logistic Regression | EdurekaEdureka!
YouTube: https://youtu.be/OCwZyYH14uw
** Data Science Certification using R: https://www.edureka.co/data-science **
This Edureka PPT on Linear Regression Vs Logistic Regression covers the basic concepts of linear and logistic models. The following topics are covered in this session:
Types of Machine Learning
Regression Vs Classification
What is Linear Regression?
What is Logistic Regression?
Linear Regression Use Case
Logistic Regression Use Case
Linear Regression Vs Logistic Regression
Blog Series: http://bit.ly/data-science-blogs
Data Science Training Playlist: http://bit.ly/data-science-playlist
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Logistic Regression in Python | Logistic Regression Example | Machine Learnin...Edureka!
** Python Data Science Training : https://www.edureka.co/python **
This Edureka Video on Logistic Regression in Python will give you basic understanding of Logistic Regression Machine Learning Algorithm with examples. In this video, you will also get to see demo on Logistic Regression using Python. Below are the topics covered in this tutorial:
1. What is Regression?
2. What is Logistic Regression?
3. Why use Logistic Regression?
4. Linear vs Logistic Regression
5. Logistic Regression Use Cases
6. Logistic Regression Example Demo in Python
Subscribe to our channel to get video updates. Hit the subscribe button above.
Machine Learning Tutorial Playlist: https://goo.gl/UxjTxm
Logistic regression : Use Case | Background | Advantages | DisadvantagesRajat Sharma
This slide will help you to understand the working of logistic regression which is a type of machine learning model along with use cases, pros and cons.
Presentation at Advanced Intelligent Systems for Sustainable Development (AISSD 2021) 20-22 August 2021 organized by the scientific research group in Egypt with Collaboration with Faculty of Computers and AI, Cairo University and the Chinese University in Egypt
In machine learning, support vector machines (SVMs, also support vector networks[1]) are supervised learning models with associated learning algorithms that analyze data and recognize patterns, used for classification and regression analysis. The basic SVM takes a set of input data and predicts, for each given input, which of two possible classes forms the output, making it a non-probabilistic binary linear classifier.
DBScan stands for Density-Based Spatial Clustering of Applications with Noise.
DBScan Concepts
DBScan Parameters
DBScan Connectivity and Reachability
DBScan Algorithm , Flowchart and Example
Advantages and Disadvantages of DBScan
DBScan Complexity
Outliers related question and its solution.
Abstract: This PDSG workshop introduces basic concepts of multiple linear regression in machine learning. Concepts covered are Feature Elimination and Backward Elimination, with examples in Python.
Level: Fundamental
Requirements: Should have some experience with Python programming.
Data Science - Part VII - Cluster AnalysisDerek Kane
This lecture provides an overview of clustering techniques, including K-Means, Hierarchical Clustering, and Gaussian Mixed Models. We will go through some methods of calibration and diagnostics and then apply the technique on a recognizable dataset.
This Logistic Regression Presentation will help you understand how a Logistic Regression algorithm works in Machine Learning. In this tutorial video, you will learn what is Supervised Learning, what is Classification problem and some associated algorithms, what is Logistic Regression, how it works with simple examples, the maths behind Logistic Regression, how it is different from Linear Regression and Logistic Regression applications. At the end, you will also see an interesting demo in Python on how to predict the number present in an image using Logistic Regression.
Below topics are covered in this Machine Learning Algorithms Presentation:
1. What is supervised learning?
2. What is classification? what are some of its solutions?
3. What is logistic regression?
4. Comparing linear and logistic regression
5. Logistic regression applications
6. Use case - Predicting the number in an image
What is Machine Learning: Machine Learning is an application of Artificial Intelligence (AI) that provides systems the ability to automatically learn and improve from experience without being explicitly programmed.
- - - - - - - -
About Simplilearn Machine Learning course:
A form of artificial intelligence, Machine Learning is revolutionizing the world of computing as well as all people’s digital interactions. Machine Learning powers such innovative automated technologies as recommendation engines, facial recognition, fraud protection and even self-driving cars.This Machine Learning course prepares engineers, data scientists and other professionals with knowledge and hands-on skills required for certification and job competency in Machine Learning.
- - - - - - -
Why learn Machine Learning?
Machine Learning is taking over the world- and with that, there is a growing need among companies for professionals to know the ins and outs of Machine Learning
The Machine Learning market size is expected to grow from USD 1.03 Billion in 2016 to USD 8.81 Billion by 2022, at a Compound Annual Growth Rate (CAGR) of 44.1% during the forecast period.
- - - - - -
What skills will you learn from this Machine Learning course?
By the end of this Machine Learning course, you will be able to:
1. Master the concepts of supervised, unsupervised and reinforcement learning concepts and modeling.
2. Gain practical mastery over principles, algorithms, and applications of Machine Learning through a hands-on approach which includes working on 28 projects and one capstone project.
3. Acquire thorough knowledge of the mathematical and heuristic aspects of Machine Learning.
4. Understand the concepts and operation of support vector machines, kernel SVM, naive bayes, decision tree classifier, random forest classifier, logistic regression, K-nearest neighbors, K-means clustering and more.
5. Be able to model a wide variety of robust Machine Learning algorithms including deep learning, clustering, and recommendation systems
- - - - - - -
One of the most important, yet often overlooked, aspects of predictive modeling is the transformation of data to create model inputs, better known as feature engineering (FE). This talk will go into the theoretical background behind FE, showing how it leverages existing data to produce better modeling results. It will then detail some important FE techniques that should be in every data scientist’s tool kit.
In machine learning, support vector machines (SVMs, also support vector networks[1]) are supervised learning models with associated learning algorithms that analyze data and recognize patterns, used for classification and regression analysis. The basic SVM takes a set of input data and predicts, for each given input, which of two possible classes forms the output, making it a non-probabilistic binary linear classifier.
DBScan stands for Density-Based Spatial Clustering of Applications with Noise.
DBScan Concepts
DBScan Parameters
DBScan Connectivity and Reachability
DBScan Algorithm , Flowchart and Example
Advantages and Disadvantages of DBScan
DBScan Complexity
Outliers related question and its solution.
Abstract: This PDSG workshop introduces basic concepts of multiple linear regression in machine learning. Concepts covered are Feature Elimination and Backward Elimination, with examples in Python.
Level: Fundamental
Requirements: Should have some experience with Python programming.
Data Science - Part VII - Cluster AnalysisDerek Kane
This lecture provides an overview of clustering techniques, including K-Means, Hierarchical Clustering, and Gaussian Mixed Models. We will go through some methods of calibration and diagnostics and then apply the technique on a recognizable dataset.
This Logistic Regression Presentation will help you understand how a Logistic Regression algorithm works in Machine Learning. In this tutorial video, you will learn what is Supervised Learning, what is Classification problem and some associated algorithms, what is Logistic Regression, how it works with simple examples, the maths behind Logistic Regression, how it is different from Linear Regression and Logistic Regression applications. At the end, you will also see an interesting demo in Python on how to predict the number present in an image using Logistic Regression.
Below topics are covered in this Machine Learning Algorithms Presentation:
1. What is supervised learning?
2. What is classification? what are some of its solutions?
3. What is logistic regression?
4. Comparing linear and logistic regression
5. Logistic regression applications
6. Use case - Predicting the number in an image
What is Machine Learning: Machine Learning is an application of Artificial Intelligence (AI) that provides systems the ability to automatically learn and improve from experience without being explicitly programmed.
- - - - - - - -
About Simplilearn Machine Learning course:
A form of artificial intelligence, Machine Learning is revolutionizing the world of computing as well as all people’s digital interactions. Machine Learning powers such innovative automated technologies as recommendation engines, facial recognition, fraud protection and even self-driving cars.This Machine Learning course prepares engineers, data scientists and other professionals with knowledge and hands-on skills required for certification and job competency in Machine Learning.
- - - - - - -
Why learn Machine Learning?
Machine Learning is taking over the world- and with that, there is a growing need among companies for professionals to know the ins and outs of Machine Learning
The Machine Learning market size is expected to grow from USD 1.03 Billion in 2016 to USD 8.81 Billion by 2022, at a Compound Annual Growth Rate (CAGR) of 44.1% during the forecast period.
- - - - - -
What skills will you learn from this Machine Learning course?
By the end of this Machine Learning course, you will be able to:
1. Master the concepts of supervised, unsupervised and reinforcement learning concepts and modeling.
2. Gain practical mastery over principles, algorithms, and applications of Machine Learning through a hands-on approach which includes working on 28 projects and one capstone project.
3. Acquire thorough knowledge of the mathematical and heuristic aspects of Machine Learning.
4. Understand the concepts and operation of support vector machines, kernel SVM, naive bayes, decision tree classifier, random forest classifier, logistic regression, K-nearest neighbors, K-means clustering and more.
5. Be able to model a wide variety of robust Machine Learning algorithms including deep learning, clustering, and recommendation systems
- - - - - - -
One of the most important, yet often overlooked, aspects of predictive modeling is the transformation of data to create model inputs, better known as feature engineering (FE). This talk will go into the theoretical background behind FE, showing how it leverages existing data to produce better modeling results. It will then detail some important FE techniques that should be in every data scientist’s tool kit.
Lecture 10b: Classification. k-Nearest Neighbor classifier, Logistic Regression, Support Vector Machines (SVM), Naive Bayes (ppt,pdf)
Chapters 4,5 from the book “Introduction to Data Mining” by Tan, Steinbach, Kumar.
Introduction:
Introduction to basic techniques. x
• Quantitative descriptors (e.g. area, length…).
• Patterns arranged in vectors.
– Structural techniques.
• Qualitative descriptors (relational descriptors for repetitive structures, e.g. staircase).
• Patterns arranged in strings or trees.
• Central idea: Learning from sample patterns
Distributed Architecture of Subspace Clustering and RelatedPei-Che Chang
Distributed Architecture of Subspace Clustering and Related
Sparse Subspace Clustering
Low-Rank Representation
Least Squares Regression
Multiview Subspace Clustering
[Paper reading] L-SHAPLEY AND C-SHAPLEY: EFFICIENT MODEL INTERPRETATION FOR S...Daiki Tanaka
paper at ICML 2019; "L-SHAPLEY AND C-SHAPLEY: EFFICIENT MODEL INTERPRETATION FOR STRUCTURED DATA"
openr eview link : https://openreview.net/forum?id=S1E3Ko09F7
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
Automobile Management System Project Report.pdfKamal Acharya
The proposed project is developed to manage the automobile in the automobile dealer company. The main module in this project is login, automobile management, customer management, sales, complaints and reports. The first module is the login. The automobile showroom owner should login to the project for usage. The username and password are verified and if it is correct, next form opens. If the username and password are not correct, it shows the error message.
When a customer search for a automobile, if the automobile is available, they will be taken to a page that shows the details of the automobile including automobile name, automobile ID, quantity, price etc. “Automobile Management System” is useful for maintaining automobiles, customers effectively and hence helps for establishing good relation between customer and automobile organization. It contains various customized modules for effectively maintaining automobiles and stock information accurately and safely.
When the automobile is sold to the customer, stock will be reduced automatically. When a new purchase is made, stock will be increased automatically. While selecting automobiles for sale, the proposed software will automatically check for total number of available stock of that particular item, if the total stock of that particular item is less than 5, software will notify the user to purchase the particular item.
Also when the user tries to sale items which are not in stock, the system will prompt the user that the stock is not enough. Customers of this system can search for a automobile; can purchase a automobile easily by selecting fast. On the other hand the stock of automobiles can be maintained perfectly by the automobile shop manager overcoming the drawbacks of existing system.
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSEDuvanRamosGarzon1
AIRCRAFT GENERAL
The Single Aisle is the most advanced family aircraft in service today, with fly-by-wire flight controls.
The A318, A319, A320 and A321 are twin-engine subsonic medium range aircraft.
The family offers a choice of engines
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
Courier management system project report.pdfKamal Acharya
It is now-a-days very important for the people to send or receive articles like imported furniture, electronic items, gifts, business goods and the like. People depend vastly on different transport systems which mostly use the manual way of receiving and delivering the articles. There is no way to track the articles till they are received and there is no way to let the customer know what happened in transit, once he booked some articles. In such a situation, we need a system which completely computerizes the cargo activities including time to time tracking of the articles sent. This need is fulfilled by Courier Management System software which is online software for the cargo management people that enables them to receive the goods from a source and send them to a required destination and track their status from time to time.
1. A Tutorial on Data
Reduction
Linear Discriminant
Analysis (LDA)
Shireen Elhabian and Aly A. Farag
University of Louisville, CVIP Lab
September 2009
2. Outline
• LDA objective
• Recall … PCA
• Now … LDA
• LDA … Two Classes
– Counter example
• LDA … C Classes
– Illustrative Example
• LDA vs PCA Example
• Limitations of LDA
3. LDA Objective
• The objective of LDA is to perform
dimensionality reduction …
– So what, PCA does this L…
• However, we want to preserve as much of the
class discriminatory information as possible.
– OK, that’s new, let dwell deeper J …
4. Recall … PCA
• In PCA, the main idea to re-express the available dataset to
extract the relevant information by reducing the redundancy
and minimize the noise.
• We didn’t care about whether this dataset represent features from one
or more classes, i.e. the discrimination power was not taken into
consideration while we were talking about PCA.
• In PCA, we had a dataset matrix X with dimensions mxn, where
columns represent different data samples.
• We first started by subtracting the mean to have a zero mean dataset,
then we computed the covariance matrix Sx = XXT.
• Eigen values and eigen vectors were then computed for Sx. Hence the
new basis vectors are those eigen vectors with highest eigen values,
where the number of those vectors was our choice.
• Thus, using the new basis, we can project the dataset onto a less
dimensional space with more powerful data representation.
n – feature vectors
(data samples)
m-dimensionaldatavector
5. Now … LDA
• Consider a pattern classification problem, where we have C-
classes, e.g. seabass, tuna, salmon …
• Each class has Ni m-dimensional samples, where i = 1,2, …, C.
• Hence we have a set of m-dimensional samples {x1, x2,…, xNi}
belong to class ωi.
• Stacking these samples from different classes into one big fat
matrix X such that each column represents one sample.
• We seek to obtain a transformation of X to Y through
projecting the samples in X onto a hyperplane with
dimension C-1.
• Let’s see what does this mean?
6. LDA … Two Classes
• Assume we have m-dimensional samples {x1,
x2,…, xN}, N1 of which belong to ω1 and
N2 belong to ω2.
• We seek to obtain a scalar y by projecting
the samples x onto a line (C-1 space, C = 2).
– where w is the projection vectors used to project x
to y.
• Of all the possible lines we would like to
select the one that maximizes the
separability of the scalars.
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ë
é
=
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ë
é
==
mm
T
w
w
wand
x
x
xwherexwy
.
.
.
.
11
The two classes are not well
separated when projected onto
this line
This line succeeded in separating
the two classes and in the
meantime reducing the
dimensionality of our problem
from two features (x1,x2) to only a
scalar value y.
7. LDA … Two Classes
• In order to find a good projection vector, we need to define a
measure of separation between the projections.
• The mean vector of each class in x and y feature space is:
– i.e. projecting x to y will lead to projecting the mean of x to the mean of
y.
• We could then choose the distance between the projected means
as our objective function
i
i
ii
i
i
i
T
xi
T
x
T
iyixi
wx
N
w
xw
N
y
N
andx
N
µ
µµ
w
www
==
===
å
ååå
Î
ÎÎÎ
1
11~1
( )222 111
~~)( µµµµµµ -=-=-= TTT
wwwwJ
8. LDA … Two Classes
• However, the distance between the projected means is not a very
good measure since it does not take into account the standard
deviation within the classes.
This axis has a larger distance between means
This axis yields better class separability
9. LDA … Two Classes
• The solution proposed by Fisher is to maximize a function that
represents the difference between the means, normalized by a
measure of the within-class variability, or the so-called scatter.
• For each class we define the scatter, an equivalent of the
variance, as; (sum of square differences between the projected samples and their class
mean).
• measures the variability within class ωi after projecting it on
the y-space.
• Thus measures the variability within the two
classes at hand after projection, hence it is called within-class scatter
of the projected samples.
( )åÎ
-=
iy
ii ys
w
µ 22 ~~
2~
is
2
2
2
1
~~ ss +
10. LDA … Two Classes
• The Fisher linear discriminant is defined as
the linear function wTx that maximizes the
criterion function: (the distance between the
projected means normalized by the within-
class scatter of the projected samples.
• Therefore, we will be looking for a projection
where examples from the same class are
projected very close to each other and, at the
same time, the projected means are as farther
apart as possible
2
2
2
1
2
2
~~
~~
)( 1
ss
wJ
+
-
=
µµ
11. LDA … Two Classes
• In order to find the optimum projection w*, we need to express
J(w) as an explicit function of w.
• We will define a measure of the scatter in multivariate feature
space x which are denoted as scatter matrices;
• Where Si is the covariance matrix of class ωi, and Sw is called the
within-class scatter matrix.
( )( )
21 SSS
xxS
w
T
i
x
ii
i
+=
--= åÎ
µµ
w
2
2
2
1
2
2
~~
~~
)( 1
ss
wJ
+
-
=
µµ
12. LDA … Two Classes
• Now, the scatter of the projection y can then be expressed as a function of
the scatter matrix in feature space x.
Where is the within-class scatter matrix of the projected samples y.
( ) ( )
( )( )
( )( )
( ) WW
TTTT
i
T
x
T
ii
T
x
T
ii
T
x
i
TT
y
ii
SwSwwSSwwSwwSwss
wSwwxxw
wxxw
wxwys
i
i
ii
~~~
~~
2121
2
2
2
1
222
==+=+=+
=÷
÷
ø
ö
ç
ç
è
æ
--=
--=
-=-=
å
å
åå
Î
Î
ÎÎ
w
w
ww
µµ
µµ
µµ
WS
~
2
2
2
1
2
2
~~
~~
)( 1
ss
wJ
+
-
=
µµ
13. LDA … Two Classes
• Similarly, the difference between the projected means (in y-space) can be
expressed in terms of the means in the original feature space (x-space).
• The matrix SB is called the between-class scatter of the original samples/feature
vectors, while is the between-class scatter of the projected samples y.
• Since SB is the outer product of two vectors, its rank is at most one.
( ) ( )
( )( )
BB
T
S
TT
TT
SwSw
ww
ww
B
~
~~
2121
2
21
2
21
==
--=
-=-
!!! "!!! #$
µµµµ
µµµµ
BS
~
2
2
2
1
2
2
~~
~~
)( 1
ss
wJ
+
-
=
µµ
14. LDA … Two Classes
• We can finally express the Fisher criterion in terms of SW and SB
as:
• Hence J(w) is a measure of the difference between class
means (encoded in the between-class scatter matrix)
normalized by a measure of the within-class scatter matrix.
wSw
wSw
ss
wJ
W
T
B
T
=
+
-
= 2
2
2
1
2
2
~~
~~
)( 1
µµ
15. LDA … Two Classes
• To find the maximum of J(w), we differentiate and equate to
zero.
( ) ( ) ( ) ( )
( ) ( )
0)(
0)(
0
:2
022
0
0)(
1
=-Þ
=-Þ
=÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
Þ
=-Þ
=-Þ
=÷÷
ø
ö
çç
è
æ
=
-
wwJwSS
wSwJwS
wS
wSw
wSw
wS
wSw
wSw
wSwbyDividing
wSwSwwSwSw
wSw
dw
d
wSwwSw
dw
d
wSw
wSw
wSw
dw
d
wJ
dw
d
BW
WB
W
W
T
B
T
B
W
T
W
T
W
T
WB
T
BW
T
W
T
B
T
B
T
W
T
W
T
B
T
16. LDA … Two Classes
• Solving the generalized eigen value problem
yields
• This is known as Fisher’s Linear Discriminant, although it is not a
discriminant but rather a specific choice of direction for the projection
of the data down to one dimension.
• Using the same notation as PCA, the solution will be the eigen
vector(s) of
scalarwJwherewwSS BW ===-
)(1
ll
( )21
1*
maxarg)(maxarg µµ -=÷÷
ø
ö
çç
è
æ
== -
W
W
T
B
T
ww
S
wSw
wSw
wJw
BWX SSS 1-
=
17. LDA … Two Classes - Example
• Compute the Linear Discriminant projection for the following two-
dimensional dataset.
– Samples for class ω1 : X1=(x1,x2)={(4,2),(2,4),(2,3),(3,6),(4,4)}
– Sample for class ω2 : X2=(x1,x2)={(9,10),(6,8),(9,5),(8,7),(10,8)}
0 1 2 3 4 5 6 7 8 9 10
0
1
2
3
4
5
6
7
8
9
10
x1
x2
18. LDA … Two Classes - Example
• The classes mean are :
÷÷
ø
ö
çç
è
æ
=ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
==
÷÷
ø
ö
çç
è
æ
=ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
+÷÷
ø
ö
çç
è
æ
==
å
å
Î
Î
6.7
4.8
8
10
7
8
5
9
8
6
10
9
5
11
8.3
3
4
4
6
3
3
2
4
2
2
4
5
11
2
2
1
1
2
1
w
w
µ
µ
x
x
x
N
x
N
19. LDA … Two Classes - Example
• Covariance matrix of the first class:
( )( )
÷÷
ø
ö
çç
è
æ
-
-
=
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
=--= åÎ
2.225.0
25.01
8.3
3
4
4
8.3
3
6
3
8.3
3
3
2
8.3
3
4
2
8.3
3
2
4
222
22
111
1
T
x
xxS µµ
w
20. LDA … Two Classes - Example
• Covariance matrix of the second class:
( )( )
÷÷
ø
ö
çç
è
æ
-
-
=
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
+ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
=--= åÎ
3.305.0
05.03.2
6.7
4.8
8
10
6.7
4.8
7
8
6.7
4.8
5
9
6.7
4.8
8
6
6.7
4.8
10
9
222
22
222
2
T
x
xxS µµ
w
21. LDA … Two Classes - Example
• Within-class scatter matrix:
÷÷
ø
ö
çç
è
æ
-
-
=
÷÷
ø
ö
çç
è
æ
-
-
+÷÷
ø
ö
çç
è
æ
-
-
=+=
5.53.0
3.03.3
3.305.0
05.03.2
2.225.0
25.01
21 SSSw
22. LDA … Two Classes - Example
• Between-class scatter matrix:
( )( )
( )
÷÷
ø
ö
çç
è
æ
=
--÷÷
ø
ö
çç
è
æ
-
-
=
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
=
--=
44.1452.20
52.2016.29
8.34.5
8.3
4.5
6.7
4.8
8.3
3
6.7
4.8
8.3
3
2121
T
T
BS µµµµ
23. LDA … Two Classes - Example
• The LDA projection is then obtained as the solution of the generalized eigen
value problem
( )( )
( )
2007.12,0
02007.1202007.12
02339.4489.69794.22213.9
9794.22339.4
489.62213.9
0
10
01
44.1452.20
52.2016.29
1827.00166.0
0166.03045.0
0
10
01
44.1452.20
52.2016.29
5.53.0
3.03.3
0
21
2
1
1
1
==Þ
=-Þ=-Þ
=´---=
÷÷
ø
ö
çç
è
æ
-
-
Þ
=÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
÷÷
ø
ö
çç
è
æ
Þ
=÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
÷÷
ø
ö
çç
è
æ
-
-
Þ
=-Þ
=
-
-
-
ll
llll
ll
l
l
l
l
l
l
ISS
wwSS
BW
BW
24. LDA … Two Classes - Example
• Hence
• The optimal projection is the one that given maximum λ = J(w)
!
*
21
2
1
2
2
1
1
4173.0
9088.0
8178.0
5755.0
;
2007.12
9794.22339.4
489.62213.9
0
9794.22339.4
489.62213.9
2
1
wwandw
Thus
w
w
w
and
w
w
w
=÷÷
ø
ö
çç
è
æ
=÷÷
ø
ö
çç
è
æ-
=
÷÷
ø
ö
çç
è
æ
=÷÷
ø
ö
çç
è
æ
÷÷
ø
ö
çç
è
æ
=÷÷
ø
ö
çç
è
æ
"#"$%
l
l
25. LDA … Two Classes - Example
( )
÷÷
ø
ö
çç
è
æ
=
÷÷
ø
ö
çç
è
æ
-
-
÷÷
ø
ö
çç
è
æ
=
ú
û
ù
ê
ë
é
÷÷
ø
ö
çç
è
æ
-÷÷
ø
ö
çç
è
æ
÷÷
ø
ö
çç
è
æ
-
-
=-=
-
-
4173.0
9088.0
8.3
4.5
1827.00166.0
0166.03045.0
6.7
4.8
8.3
3
5.53.0
3.03.3
1
21
1*
µµWSw
Or directly;
26. LDA - Projection
-4 -3 -2 -1 0 1 2 3 4 5 6
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
y
p(y|wi
)
Classes PDF : using the LDA projection vector with the other eigen value = 8.8818e-016
-7 -6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6 7 8 9 10
0
1
2
3
4
5
6
7
8
9
10
x1
x
2
LDA projection vector with the other eigen value = 8.8818e-016
The projection vector
corresponding to the
smallest eigen value
Using this vector leads to
bad separability
between the two classes
27. LDA - Projection
0 5 10 15
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
y
p(y|wi
)
Classes PDF : using the LDA projection vector with highest eigen value = 12.2007
0 1 2 3 4 5 6 7 8 9 10
0
1
2
3
4
5
6
7
8
9
10
x1
x
2
LDA projection vector with the highest eigen value = 12.2007
The projection vector
corresponding to the
highest eigen value
Using this vector leads to
good separability
between the two classes
28. LDA … C-Classes
• Now, we have C-classes instead of just two.
• We are now seeking (C-1) projections [y1, y2, …, yC-1] by means
of (C-1) projection vectors wi.
• wi can be arranged by columns into a projection matrix W =
[w1|w2|…|wC-1] such that:
[ ]1211
1
1
11
1
1
|...||
.
.
,
.
.
--´
-
´-´
=
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ë
é
=
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ë
é
=
CCm
C
C
m
m
wwwWand
y
y
y
x
x
xwhere
xWyxwy TT
ii =Þ=
29. LDA … C-Classes
• If we have n-feature vectors, we can stack them into one matrix
as follows;
[ ]1211
1
2
1
1
1
1
2
1
1
1
1
21
1
2
1
1
1
|...||
.
....
....
.
,
.
....
....
.
--´
---
´-´
=
ú
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ê
ë
é
=
ú
ú
ú
ú
ú
û
ù
ê
ê
ê
ê
ê
ë
é
=
CCm
n
CCC
n
nC
n
mmm
n
nm
wwwWand
yyy
yyy
Y
xxx
xxx
Xwhere
XWY T
=
30. LDA – C-Classes
• Recall the two classes case, the
within-class scatter was computed as:
• This can be generalized in the C-
classes case as:
21 SSSw +=
( )( )
å
å
å
Î
Î
=
=
--=
=
i
i
xi
i
T
i
x
ii
C
i
iW
x
N
and
xxSwhere
SS
w
w
µ
µµ
1
1
x1
x2
µ1
µ2
µ3
Sw2
Example of two-dimensional
features (m = 2), with three
classes C = 3.
Ni : number of data samples
in class ωi.
31. LDA – C-Classes
• Recall the two classes case, the between-
class scatter was computed as:
• For C-classes case, we will measure the
between-class scatter with respect to the
mean of all classes as follows:
( )( )
å
åå
å
Î
""
=
=
==
--=
ixi
i
x
ii
x
C
i
T
iiiB
x
N
and
N
N
x
N
where
NS
w
µ
µµ
µµµµ
1
11
1
( )( )T
BS 2121 µµµµ --=
x1
x2
µ1
µ2
µ3
Sw2
µ
Example of two-dimensional
features (m = 2), with three
classes C = 3.
Ni : number of data samples
in class ωi.
N: number of all data .
32. LDA – C-Classes
• Similarly,
– We can define the mean vectors for the projected samples y as:
– While the scatter matrices for the projected samples y will be:
åå "Î
==
yyi
i y
N
andy
N i
1~1~ µµ
w
( )( )ååå = Î=
--==
C
i
T
i
y
i
C
i
iW yySS
i11
~~~~
µµ
w
( )( )å=
--=
C
i
T
iiiB NS
1
~~~~~
µµµµ
33. LDA – C-Classes
• Recall in two-classes case, we have expressed the scatter matrices of the
projected samples in terms of those of the original samples as:
This still hold in C-classes case.
• Recall that we are looking for a projection that maximizes the ratio of
between-class to within-class scatter.
• Since the projection is no longer a scalar (it has C-1 dimensions), we then use
the determinant of the scatter matrices to obtain a scalar objective function:
• And we will seek the projection W* that maximizes this ratio.
WSWS
WSWS
B
T
B
W
T
W
=
=
~
~
WSW
WSW
S
S
WJ
W
T
B
T
W
B
== ~
~
)(
34. LDA – C-Classes
• To find the maximum of J(W), we differentiate with respect to W and equate
to zero.
• Recall in two-classes case, we solved the eigen value problem.
• For C-classes case, we have C-1 projection vectors, hence the eigen value
problem can be generalized to the C-classes case as:
• Thus, It can be shown that the optimal projection matrix W* is the one whose
columns are the eigenvectors corresponding to the largest eigen values of the
following generalized eigen value problem:
1,...2,1)(1
-====-
CiandscalarwJwherewwSS iiiiiBW ll
scalarwJwherewwSS BW ===-
)(1
ll
[ ]*
1
*
2
*
1
**
**1
|...||)( -
-
===
=
C
BW
wwwWandscalarWJwhere
WWSS
l
l
35. Illustration – 3 Classes
• Let’s generate a dataset for each
class to simulate the three classes
shown
• For each class do the following,
– Use the random number generator
to generate a uniform stream of
500 samples that follows U(0,1).
– Using the Box-Muller approach,
convert the generated uniform
stream to N(0,1).
x1
x2
µ1
µ2
µ3
µ
– Then use the method of eigen values
and eigen vectors to manipulate the
standard normal to have the required
mean vector and covariance matrix .
– Estimate the mean and covariance
matrix of the resulted dataset.
36. • By visual inspection of the figure,
classes parameters (means and
covariance matrices can be given as
follows:
Dataset Generation
x1
x2
µ1
µ2
µ3
µ
÷÷
ø
ö
çç
è
æ
=
÷÷
ø
ö
çç
è
æ
=
÷÷
ø
ö
çç
è
æ
-
-
=
ú
û
ù
ê
ë
é
+=ú
û
ù
ê
ë
é
-
-
+=ú
û
ù
ê
ë
é-
+=
ú
û
ù
ê
ë
é
=
5.23
15.3
40
04
33
15
5
7
,
5.3
5.2
,
7
3
5
5
meanOverall
3
2
1
321
S
S
S
µµµµµµ
µ
Zero covariance to lead to data samples distributed horizontally.
Positive covariance to lead to data samples distributed along the y = x line.
Negative covariance to lead to data samples distributed along the y = -x line.
38. It’s Working … J
-5 0 5 10 15 20
-5
0
5
10
15
20
X1
- the first feature
X2
-thesecondfeature
x1
x2
µ1
µ2
µ3
µ
39. Computing LDA Projection Vectors
( )( )
å
åå
å
Î
""
=
=
==
--=
ixi
i
x
ii
x
C
i
T
iiiB
x
N
and
N
N
x
N
where
NS
w
µ
µµ
µµµµ
1
11
1
( )( )
å
å
å
Î
Î
=
=
--=
=
i
i
xi
i
T
i
x
ii
C
i
iW
x
N
and
xxSwhere
SS
w
w
µ
µµ
1
1
Recall …
BW SS 1-
40. Let’s visualize the projection vectors W
-15 -10 -5 0 5 10 15 20 25
-10
-5
0
5
10
15
20
25
X1
- the first feature
X2
-thesecondfeature
41. Projection … y = WTx
Along first projection vector
-5 0 5 10 15 20 25
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
y
p(y|wi
)
Classes PDF : using the first projection vector with eigen value = 4508.2089
42. Projection … y = WTx
Along second projection vector
-10 -5 0 5 10 15 20
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
y
p(y|wi
)
Classes PDF : using the second projection vector with eigen value = 1878.8511
43. Which is Better?!!!
• Apparently, the projection vector that has the highest eigen
value provides higher discrimination power between classes
-10 -5 0 5 10 15 20
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
y
p(y|wi
)
Classes PDF : using the second projection vector with eigen value = 1878.8511
-5 0 5 10 15 20 25
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
y
p(y|wi
)
Classes PDF : using the first projection vector with eigen value = 4508.2089
45. Limitations of LDA L
• LDA produces at most C-1 feature projections
– If the classification error estimates establish that more features are needed, some other method must be
employed to provide those additional features
• LDA is a parametric method since it assumes unimodal Gaussian likelihoods
– If the distributions are significantly non-Gaussian, the LDA projections will not be able to preserve any
complex structure of the data, which may be needed for classification.
46. Limitations of LDA L
• LDA will fail when the discriminatory information is not in the mean but
rather in the variance of the data