SlideShare a Scribd company logo
1 of 28
Download to read offline
Jeonghun Yoon
Machine Learning? ๊ธฐ๊ณ„ ํ•™์Šต?
๏ฌ ํ•™์Šต(learning)
โ—‹ ํ•˜๋‚˜์˜ ๋ฌธ์ œ๋ฅผ ์ˆ˜ํ–‰ํ•œ ํ›„์— ๊ทธ ์ถ”๋ก ๊ณผ์ •์—์„œ ์–ป์€ ๊ฒฝํ—˜์„ ๋ฐ”ํƒ•์œผ๋กœ ์‹œ์Šคํ…œ์˜ ์ง€์‹์„
์ˆ˜์ • ๋ฐ ๋ณด์™„ํ•˜์—ฌ, ๋‹ค์Œ์— ๊ทธ ๋ฌธ์ œ๋‚˜ ๋˜๋Š” ๋น„์Šทํ•œ ๋ฌธ์ œ๋ฅผ ์ˆ˜ํ–‰ํ•  ๋•Œ์—๋Š” ์ฒ˜์Œ๋ณด๋‹ค ๋”
ํšจ์œจ์ ์ด๊ณ  ํšจ๊ณผ์ ์œผ๋กœ ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•  ์ˆ˜ ์žˆ๋Š” ์ ์‘์„ฑ
โ—‹ ์ƒˆ๋กœ์šด ์„ ์–ธ์  ์ง€์‹์˜ ์Šต๋“, ์ง€๋„ ๋ฐ ์‹ค์Šต์„ ํ†ตํ•œ ์ธ์ง€์ ์ธ ๊ธฐ์ˆ ์˜ ๊ฐœ๋ฐœ, ์ƒˆ๋กœ์šด ์ง€์‹์˜
์ผ๋ฐ˜์ ์ด๊ณ  ํšจ๊ณผ์ ์ธ ํ‘œํ˜„์œผ๋กœ์˜ ์กฐ์งํ™”, ๊ด€์ฐฐ ์ด๋‚˜ ์‹คํ—˜์„ ํ†ตํ•œ ์ƒˆ๋กœ์šด ์‚ฌ์‹ค์ด๋‚˜ ์ด๋ก 
์˜ ๋ฐœ๊ฒฌ ๋“ฑ๊ณผ ๊ฐ™์€ ๋‹ค์–‘ํ•œ ๊ณผ์ •๋“ค์„ ํฌํ•จํ•œ๋‹ค.
โ€ป Wikipedia ์ฐธ์กฐ
๏ฌ ๊ธฐ๊ณ„ ํ•™์Šต(Machine Learning)
โ—‹ ๊ธฐ๊ณ„ ํ•™์Šต(machine learning)์€ ์ธ๊ณต ์ง€๋Šฅ์˜ ํ•œ ๋ถ„์•ผ๋กœ, ์ปดํ“จํ„ฐ๊ฐ€ ํ•™์Šตํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜
๋Š” ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ๊ธฐ์ˆ ์„ ๊ฐœ๋ฐœํ•˜๋Š” ๋ถ„์•ผ๋ฅผ ๋งํ•œ๋‹ค. ๊ฐ€๋ น, ๊ธฐ๊ณ„ ํ•™์Šต์„ ํ†ตํ•ด์„œ ์ˆ˜์‹ ํ•œ ์ด๋ฉ”
์ผ์ด ์ŠคํŒธ์ธ์ง€ ์•„๋‹Œ์ง€๋ฅผ ๊ตฌ๋ถ„ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ›ˆ๋ จํ•  ์ˆ˜ ์žˆ๋‹ค.
โ—‹ "์ปดํ“จํ„ฐ์—๊ฒŒ ๋ฐฐ์šธ ์ˆ˜ ์žˆ๋Š” ๋Šฅ๋ ฅ, ์ฆ‰ ์ฝ”๋“œ๋กœ ์ •์˜ํ•˜์ง€ ์•Š์€ ๋™์ž‘์„ ์‹คํ–‰ํ•˜๋Š” ๋Šฅ๋ ฅ์— ๋Œ€
ํ•œ ์—ฐ๊ตฌ ๋ถ„์•ผ"
โ—‹ ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ๋ฅผ ํ†ตํ•ด ์ด๋ฏธ ์•Œ๋ ค์ง„ ์†์„ฑ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ์˜ˆ์ธกํ•˜๋Š” ๋Šฅ๋ ฅ์„ ์˜๋ฏธํ•œ๋‹ค.
โ€ป Wikipedia ์ฐธ์กฐ
๏ฌ ๋ฐ์ดํ„ฐ ๋งˆ์ด๋‹(data mining)
โ—‹ ๋Œ€๊ทœ๋ชจ๋กœ ์ €์žฅ๋œ ๋ฐ์ดํ„ฐ ์•ˆ์—์„œ ์ฒด๊ณ„์ ์ด๊ณ  ์ž๋™์ ์œผ๋กœ ํ†ต๊ณ„์  ๊ทœ์น™์ด๋‚˜ ํŒจํ„ด์„ ์ฐพ์•„
๋‚ด๋Š” ๊ฒƒ์ด๋‹ค.
โ—‹ ๋‹ค๋ฅธ ๋ง๋กœ๋Š” KDD(๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค ์†์˜ ์ง€์‹ ๋ฐœ๊ฒฌ, knowledge-discovery in databases)๋ผ
๊ณ ๋„ ์ผ์ปซ๋Š”๋‹ค.
โ€ป Wikipedia ์ฐธ์กฐ
1. Introduction of Machine Learning
2. Naive Bayesian Classifier
3. Linear Regression
4. Logistic Regression
5. K-means Clsutering
6. Graph Mining
7. Dimensional Reduction (PCA)
8. Spectral Clustering
9. Association Rule Mining
10. Bayesian Network 1 & 2
11. Decision tree
12. Support Vector Machine (SVM) 1 & 2
13. Hidden Markov Model (HMM)
14. Markov chain Monte Carlo(MCMC)
15. Gibbs Sampling
16. Latent Dirichlet allocation (LDA)
17. Neural Networks
Definition( Machine Learning / ๊ธฐ๊ณ„ํ•™์Šต )
A set of methods that can automatically detect patters in data, and then use
the uncovered patterns to predict patterns to predict future data, or to
perform other kinds of decision making under uncertainty
"Machine Learning-A Probabilistic Perspective"
Kevin P. Murphy
Supervised Learning
Classification
Bayesian Classifier
Logistic Regression
KNN Classifier
Support Vector Machine
(SVM)
Regression Linear Regression
Unsupervised Learning
Clustering
K-means Clustering
Spectral Clustering
Dimensional Reduction PCA
Reinforce Learning
A feature vector is an ๐‘›-dimensional vector of numerical features
that represent some object.
For example ,
a document , ๐‘ฅ๐‘– : ๋ฌธ์„œ ์•ˆ์—์„œ์˜ ๐‘– ๋ฒˆ์งธ ๋‹จ์–ด
I love you. ๐•ฉ = ๐ผ, ๐‘™๐‘œ๐‘ฃ๐‘’, ๐‘ฆ๐‘œ๐‘ข
๐•ฉ ๐‘ฆ
๐‘“mapping
๋ถˆ์—ฐ์† ๊ฐ’ ๋˜๋Š”
์—ฐ์† ๊ฐ’
E-mail (Words) Spam or Not(๋ถˆ)
Web Site (Words) Sports or Science or News(๋ถˆ)
ํŠน์„ฑ ๋ฒกํ„ฐ
๊ฝƒ (๊ฝƒ์˜ ์ƒ๊น€์ƒˆ) Line-flower or Mass-flower(๋ถˆ)
์•„์ด์˜ ํ‚ค ์•„๋ฒ„์ง€์˜์˜ ํ‚ค(์—ฐ)
๋ฐฉ์˜ ๊ฐœ์ˆ˜, ๋ฐฉ์˜ ๋„“์ด ์ง‘ ๊ฐ’(์—ฐ)
(visit, money, buy, girl, Viagra)
For example,
Spam mail
๐‘“
Goal of Supervised Learning (predictive learning)
To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, given a labeled set of
input-output pairs ๐ท = ๐•ฉ๐‘–, ๐‘ฆ๐‘– ๐‘–=1
๐‘
, where ๐ท is called the training set, and ๐‘
is the number of training examples.
- ๐•ฉ๐‘– : ๐ท-dimensional vector of numbers โ‡’ feature vector
- ๐‘ฆ๐‘– : response variable โ‡’ categories or real-values
๐‘ฆ๐‘– is categorical, the problem is classification,
๐‘ฆ๐‘– is real-valued, the problem is regression.
What is natural grouping among these objects?
Simpson's Family School Employees Females males
Goal of Unsupervised Learning (descriptive learning)
To find "interesting pattern" in the data, given ๐ท = ๐•ฉ๐‘– ๐‘–=1
๐‘
, where ๐ท is called
the training set, and ๐‘ is the number of training examples.
- ๐•ฉ๐‘– : ๐ท-dimensional vector of numbers โ‡’ feature vector
It is known knowledge discovery.
Goal of Reinforcement Learning
To learn how to act or behave when given occasional reward or punishment
sinnals.
Supervised Learning
Classification
Bayesian Classifier
Logistic Regression
KNN Classifier
Support Vector Machine
(SVM)
Regression Linear Regression
Unsupervised Learning
Clustering
K-means Clustering
Spectral Clustering
Dimensional Reduction PCA
Reinforce Learning
๐•ฉ ๐‘ฆ
๐‘“mapping
๋ถˆ์—ฐ์† ๊ฐ’ ๋˜๋Š”
์—ฐ์† ๊ฐ’
E-mail (Words) Spam or Not(๋ถˆ)
Web Site (Words) Sports or Science or News(๋ถˆ)
ํŠน์„ฑ ๋ฒกํ„ฐ
๊ฝƒ (๊ฝƒ์˜ ์ƒ๊น€์ƒˆ) Line-flower or Mass-flower(๋ถˆ)
Goal of Classification
To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, where ๐‘ฆ โˆˆ {1, โ€ฆ , ๐ถ},
with ๐ถ being the number of classes.
function approximation
Assume that ๐‘ฆ = ๐‘“(๐•ฉ) for unknown function โ„Ž, the goal of learning is to
estimate function ๐‘“ given a labeled training set, and then to make predictions
using ๐‘ฆโˆ— = ๐‘“โˆ—(๐•ฉ). Then we can make predictions on novel input.
way to formalize the problem
Compute!! our "best guess" using
๐‘ฆโˆ— = ๐‘“โˆ— ๐•ฉ = ๐‘Ž๐‘Ÿ๐‘” max
๐‘=1โ€ฆ๐ถ
๐‘(๐‘ฆ = ๐‘|๐•ฉ, ๐ท)
This corresponds to the most probable class label.
It is known as a MAP estimate (Maximum a Posterior).
Ex) mail spam filtering
๐‘ ๐‘ฆ = 1 ๐•ฉ, ๐ท)
๐‘ ๐‘ฆ = 2 ๐•ฉ, ๐ท)
...
๐‘ ๐‘ฆ = ๐ถ ๐•ฉ, ๐ท)
Goal of Regression
To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, where ๐‘ฆ is continuous.
๐‘ฆ = ๐œ–1 + ๐œ–2 ๐‘ฅ ๐‘ฆ = ๐œ–1 + ๐œ–2 ๐‘ฅ + ๐œ€3 ๐‘ฅ2
Supervised Learning
Classification
Bayesian Classifier
Logistic Regression
KNN Classifier
Support Vector Machine
(SVM)
Regression Linear Regression
Unsupervised Learning
Clustering
K-means Clustering
Spectral Clustering
Dimensional Reduction PCA
Reinforce Learning
What is natural grouping among these objects?
Simpson's Family School Employees Females males
Goal of Clustering
To estimate the distribution over the number of cluster, ๐‘ ๐พ ๐ท ; this tells us if
there are subpopulations within the data
To estimate which cluster each point belong to.
๐พโˆ—
= ๐‘Ž๐‘Ÿ๐‘” max
๐พ
๐‘(๐พ|๐ท)
We often approximate the distribution
We can infer which cluster each data point belongs to by computing
๐‘ง๐‘–
โˆ—
= ๐‘Ž๐‘Ÿ๐‘” max
๐‘˜
๐‘(๐‘ง๐‘– = ๐‘˜|๐•ฉ๐‘–, ๐ท)
๐‘ ๐‘ง๐‘– = 1 ๐•ฉ, ๐ท)
๐‘ ๐‘ง๐‘– = 2 ๐•ฉ, ๐ท)
...
๐‘ ๐‘ง๐‘– = ๐ถ ๐•ฉ, ๐ท)
Goal of Dimensional Reduction
To reduce the dimensionality by projecting the data to a lower dimensional
subspace which captures the "essence" of the data.
Motivation : Although the data may appear high dimensional, there are
only be a small number of degrees of variability, corresponding
to latent factors.
latent factor : which describe most of the variability
๏ฌ "Machine Learning-A Probabilistic Perspective" Kevin P. Murphy
๏ฌ http://ko.wikipedia.org/wiki/๊ธฐ๊ณ„_ํ•™์Šต

More Related Content

What's hot

What's hot (20)

06. graph mining
06. graph mining06. graph mining
06. graph mining
ย 
SVM
SVMSVM
SVM
ย 
04. logistic regression ( ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€ )
04. logistic regression ( ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€ )04. logistic regression ( ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€ )
04. logistic regression ( ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€ )
ย 
0207 1 gradient
0207 1 gradient0207 1 gradient
0207 1 gradient
ย 
Neural network (perceptron)
Neural network (perceptron)Neural network (perceptron)
Neural network (perceptron)
ย 
0131 2 spectral_theorem_eigenvalue
0131 2 spectral_theorem_eigenvalue0131 2 spectral_theorem_eigenvalue
0131 2 spectral_theorem_eigenvalue
ย 
0124 1 linear_algebra_basic_vector
0124 1 linear_algebra_basic_vector0124 1 linear_algebra_basic_vector
0124 1 linear_algebra_basic_vector
ย 
0124 2 linear_algebra_basic_matrix
0124 2 linear_algebra_basic_matrix0124 2 linear_algebra_basic_matrix
0124 2 linear_algebra_basic_matrix
ย 
0307 1 estimation_theory
0307 1 estimation_theory0307 1 estimation_theory
0307 1 estimation_theory
ย 
0228 2 sample_distribution
0228 2 sample_distribution0228 2 sample_distribution
0228 2 sample_distribution
ย 
0314 1 anova
0314 1 anova0314 1 anova
0314 1 anova
ย 
Ensemble Model (Hybrid model)
Ensemble Model (Hybrid model)Ensemble Model (Hybrid model)
Ensemble Model (Hybrid model)
ย 
Variational AutoEncoder(VAE)
Variational AutoEncoder(VAE)Variational AutoEncoder(VAE)
Variational AutoEncoder(VAE)
ย 
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 07-๊ณ ๊ธ‰ ๋ถ„๋ฅ˜ ๊ธฐ๋ฒ•-์ปค๋„ ๊ธฐ๋ฒ•๊ณผ svm-01
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 07-๊ณ ๊ธ‰ ๋ถ„๋ฅ˜ ๊ธฐ๋ฒ•-์ปค๋„ ๊ธฐ๋ฒ•๊ณผ svm-01์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 07-๊ณ ๊ธ‰ ๋ถ„๋ฅ˜ ๊ธฐ๋ฒ•-์ปค๋„ ๊ธฐ๋ฒ•๊ณผ svm-01
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 07-๊ณ ๊ธ‰ ๋ถ„๋ฅ˜ ๊ธฐ๋ฒ•-์ปค๋„ ๊ธฐ๋ฒ•๊ณผ svm-01
ย 
์ฐจ์›์ถ•์†Œ ํ›‘์–ด๋ณด๊ธฐ (PCA, SVD, NMF)
์ฐจ์›์ถ•์†Œ ํ›‘์–ด๋ณด๊ธฐ (PCA, SVD, NMF)์ฐจ์›์ถ•์†Œ ํ›‘์–ด๋ณด๊ธฐ (PCA, SVD, NMF)
์ฐจ์›์ถ•์†Œ ํ›‘์–ด๋ณด๊ธฐ (PCA, SVD, NMF)
ย 
๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ณธ ์›๋ฆฌ์˜ ์ดํ•ด
๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ณธ ์›๋ฆฌ์˜ ์ดํ•ด๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ณธ ์›๋ฆฌ์˜ ์ดํ•ด
๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ณธ ์›๋ฆฌ์˜ ์ดํ•ด
ย 
3.unsupervised learing
3.unsupervised learing3.unsupervised learing
3.unsupervised learing
ย 
0221 basic probability theory
0221 basic probability theory0221 basic probability theory
0221 basic probability theory
ย 
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 06-์˜์‚ฌ๊ฒฐ์ •ํŠธ๋ฆฌ-01
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 06-์˜์‚ฌ๊ฒฐ์ •ํŠธ๋ฆฌ-01์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 06-์˜์‚ฌ๊ฒฐ์ •ํŠธ๋ฆฌ-01
์ง‘๋‹จ์ง€์„ฑ ํ”„๋กœ๊ทธ๋ž˜๋ฐ 06-์˜์‚ฌ๊ฒฐ์ •ํŠธ๋ฆฌ-01
ย 
Chapter 19 Variational Inference
Chapter 19 Variational InferenceChapter 19 Variational Inference
Chapter 19 Variational Inference
ย 

Similar to 01. introduction

แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
ssusercdf17c
ย 

Similar to 01. introduction (20)

แ„€แ…ตแ„€แ…จแ„’แ…กแ†จแ„‰แ…ณแ†ธ Overview
แ„€แ…ตแ„€แ…จแ„’แ…กแ†จแ„‰แ…ณแ†ธ  Overviewแ„€แ…ตแ„€แ…จแ„’แ…กแ†จแ„‰แ…ณแ†ธ  Overview
แ„€แ…ตแ„€แ…จแ„’แ…กแ†จแ„‰แ…ณแ†ธ Overview
ย 
์บ๋นˆ๋จธํ”ผ ๋จธ์‹ ๋Ÿฌ๋‹ Kevin Murphy Machine Learning Statistic
์บ๋นˆ๋จธํ”ผ ๋จธ์‹ ๋Ÿฌ๋‹ Kevin Murphy Machine Learning Statistic์บ๋นˆ๋จธํ”ผ ๋จธ์‹ ๋Ÿฌ๋‹ Kevin Murphy Machine Learning Statistic
์บ๋นˆ๋จธํ”ผ ๋จธ์‹ ๋Ÿฌ๋‹ Kevin Murphy Machine Learning Statistic
ย 
แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
แ„†แ…ฅแ„‰แ…ตแ†ซ แ„…แ…ฅแ„‚แ…ตแ†ผแ„‹แ…ณแ†ฏ แ„’แ…ขแ„‡แ…ฉแ„Œแ…ก 1แ„Œแ…กแ†ผ (2022แ„‚แ…งแ†ซ แ„‰แ…ณแ„แ…ฅแ„ƒแ…ต)
ย 
Naive ML Overview
Naive ML OverviewNaive ML Overview
Naive ML Overview
ย 
Machine learning bysogood
Machine learning bysogoodMachine learning bysogood
Machine learning bysogood
ย 
DSAC M2 #1
DSAC M2 #1DSAC M2 #1
DSAC M2 #1
ย 
Lecture 4: Neural Networks I
Lecture 4: Neural Networks ILecture 4: Neural Networks I
Lecture 4: Neural Networks I
ย 
๋”ฅ๋Ÿฌ๋‹์„ ์ด์šฉํ•œ ์ž์—ฐ์–ด์ฒ˜๋ฆฌ์˜ ์—ฐ๊ตฌ๋™ํ–ฅ
๋”ฅ๋Ÿฌ๋‹์„ ์ด์šฉํ•œ ์ž์—ฐ์–ด์ฒ˜๋ฆฌ์˜ ์—ฐ๊ตฌ๋™ํ–ฅ๋”ฅ๋Ÿฌ๋‹์„ ์ด์šฉํ•œ ์ž์—ฐ์–ด์ฒ˜๋ฆฌ์˜ ์—ฐ๊ตฌ๋™ํ–ฅ
๋”ฅ๋Ÿฌ๋‹์„ ์ด์šฉํ•œ ์ž์—ฐ์–ด์ฒ˜๋ฆฌ์˜ ์—ฐ๊ตฌ๋™ํ–ฅ
ย 
Deep learning study 1
Deep learning study 1Deep learning study 1
Deep learning study 1
ย 
Ml
MlMl
Ml
ย 
2.supervised learning(epoch#2)-1
2.supervised learning(epoch#2)-12.supervised learning(epoch#2)-1
2.supervised learning(epoch#2)-1
ย 
ํŒŒ์ด์ฌ(Python) ์œผ๋กœ ๋‚˜๋งŒ์˜ ๋”ฅ๋Ÿฌ๋‹ API ๋งŒ๋“ค๊ธฐ ๊ฐ•์ขŒ (Feat. AutoAI )
ํŒŒ์ด์ฌ(Python) ์œผ๋กœ ๋‚˜๋งŒ์˜ ๋”ฅ๋Ÿฌ๋‹ API ๋งŒ๋“ค๊ธฐ ๊ฐ•์ขŒ (Feat. AutoAI ) ํŒŒ์ด์ฌ(Python) ์œผ๋กœ ๋‚˜๋งŒ์˜ ๋”ฅ๋Ÿฌ๋‹ API ๋งŒ๋“ค๊ธฐ ๊ฐ•์ขŒ (Feat. AutoAI )
ํŒŒ์ด์ฌ(Python) ์œผ๋กœ ๋‚˜๋งŒ์˜ ๋”ฅ๋Ÿฌ๋‹ API ๋งŒ๋“ค๊ธฐ ๊ฐ•์ขŒ (Feat. AutoAI )
ย 
Rแ„‹แ…ณแ†ฏ แ„‹แ…ตแ„‹แ…ญแ†ผแ„’แ…กแ†ซ แ„ƒแ…ฆแ„‹แ…ตแ„แ…ฅ แ„‡แ…ฎแ†ซแ„‰แ…ฅแ†จ
Rแ„‹แ…ณแ†ฏ แ„‹แ…ตแ„‹แ…ญแ†ผแ„’แ…กแ†ซ แ„ƒแ…ฆแ„‹แ…ตแ„แ…ฅ แ„‡แ…ฎแ†ซแ„‰แ…ฅแ†จRแ„‹แ…ณแ†ฏ แ„‹แ…ตแ„‹แ…ญแ†ผแ„’แ…กแ†ซ แ„ƒแ…ฆแ„‹แ…ตแ„แ…ฅ แ„‡แ…ฎแ†ซแ„‰แ…ฅแ†จ
Rแ„‹แ…ณแ†ฏ แ„‹แ…ตแ„‹แ…ญแ†ผแ„’แ…กแ†ซ แ„ƒแ…ฆแ„‹แ…ตแ„แ…ฅ แ„‡แ…ฎแ†ซแ„‰แ…ฅแ†จ
ย 
์ธ๊ณต์‹ ๊ฒฝ๋ง
์ธ๊ณต์‹ ๊ฒฝ๋ง์ธ๊ณต์‹ ๊ฒฝ๋ง
์ธ๊ณต์‹ ๊ฒฝ๋ง
ย 
Learning by association
Learning by associationLearning by association
Learning by association
ย 
2.supervised learning
2.supervised learning2.supervised learning
2.supervised learning
ย 
2.linear regression and logistic regression
2.linear regression and logistic regression2.linear regression and logistic regression
2.linear regression and logistic regression
ย 
From maching learning to deep learning
From maching learning to deep learningFrom maching learning to deep learning
From maching learning to deep learning
ย 
Recurrent Neural Net์˜ ์ด๋ก ๊ณผ ์„ค๋ช…
Recurrent Neural Net์˜ ์ด๋ก ๊ณผ ์„ค๋ช…Recurrent Neural Net์˜ ์ด๋ก ๊ณผ ์„ค๋ช…
Recurrent Neural Net์˜ ์ด๋ก ๊ณผ ์„ค๋ช…
ย 
Mahout
MahoutMahout
Mahout
ย 

01. introduction

  • 3. ๏ฌ ํ•™์Šต(learning) โ—‹ ํ•˜๋‚˜์˜ ๋ฌธ์ œ๋ฅผ ์ˆ˜ํ–‰ํ•œ ํ›„์— ๊ทธ ์ถ”๋ก ๊ณผ์ •์—์„œ ์–ป์€ ๊ฒฝํ—˜์„ ๋ฐ”ํƒ•์œผ๋กœ ์‹œ์Šคํ…œ์˜ ์ง€์‹์„ ์ˆ˜์ • ๋ฐ ๋ณด์™„ํ•˜์—ฌ, ๋‹ค์Œ์— ๊ทธ ๋ฌธ์ œ๋‚˜ ๋˜๋Š” ๋น„์Šทํ•œ ๋ฌธ์ œ๋ฅผ ์ˆ˜ํ–‰ํ•  ๋•Œ์—๋Š” ์ฒ˜์Œ๋ณด๋‹ค ๋” ํšจ์œจ์ ์ด๊ณ  ํšจ๊ณผ์ ์œผ๋กœ ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•  ์ˆ˜ ์žˆ๋Š” ์ ์‘์„ฑ โ—‹ ์ƒˆ๋กœ์šด ์„ ์–ธ์  ์ง€์‹์˜ ์Šต๋“, ์ง€๋„ ๋ฐ ์‹ค์Šต์„ ํ†ตํ•œ ์ธ์ง€์ ์ธ ๊ธฐ์ˆ ์˜ ๊ฐœ๋ฐœ, ์ƒˆ๋กœ์šด ์ง€์‹์˜ ์ผ๋ฐ˜์ ์ด๊ณ  ํšจ๊ณผ์ ์ธ ํ‘œํ˜„์œผ๋กœ์˜ ์กฐ์งํ™”, ๊ด€์ฐฐ ์ด๋‚˜ ์‹คํ—˜์„ ํ†ตํ•œ ์ƒˆ๋กœ์šด ์‚ฌ์‹ค์ด๋‚˜ ์ด๋ก  ์˜ ๋ฐœ๊ฒฌ ๋“ฑ๊ณผ ๊ฐ™์€ ๋‹ค์–‘ํ•œ ๊ณผ์ •๋“ค์„ ํฌํ•จํ•œ๋‹ค. โ€ป Wikipedia ์ฐธ์กฐ
  • 4. ๏ฌ ๊ธฐ๊ณ„ ํ•™์Šต(Machine Learning) โ—‹ ๊ธฐ๊ณ„ ํ•™์Šต(machine learning)์€ ์ธ๊ณต ์ง€๋Šฅ์˜ ํ•œ ๋ถ„์•ผ๋กœ, ์ปดํ“จํ„ฐ๊ฐ€ ํ•™์Šตํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜ ๋Š” ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ๊ธฐ์ˆ ์„ ๊ฐœ๋ฐœํ•˜๋Š” ๋ถ„์•ผ๋ฅผ ๋งํ•œ๋‹ค. ๊ฐ€๋ น, ๊ธฐ๊ณ„ ํ•™์Šต์„ ํ†ตํ•ด์„œ ์ˆ˜์‹ ํ•œ ์ด๋ฉ” ์ผ์ด ์ŠคํŒธ์ธ์ง€ ์•„๋‹Œ์ง€๋ฅผ ๊ตฌ๋ถ„ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ›ˆ๋ จํ•  ์ˆ˜ ์žˆ๋‹ค. โ—‹ "์ปดํ“จํ„ฐ์—๊ฒŒ ๋ฐฐ์šธ ์ˆ˜ ์žˆ๋Š” ๋Šฅ๋ ฅ, ์ฆ‰ ์ฝ”๋“œ๋กœ ์ •์˜ํ•˜์ง€ ์•Š์€ ๋™์ž‘์„ ์‹คํ–‰ํ•˜๋Š” ๋Šฅ๋ ฅ์— ๋Œ€ ํ•œ ์—ฐ๊ตฌ ๋ถ„์•ผ" โ—‹ ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ๋ฅผ ํ†ตํ•ด ์ด๋ฏธ ์•Œ๋ ค์ง„ ์†์„ฑ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ์˜ˆ์ธกํ•˜๋Š” ๋Šฅ๋ ฅ์„ ์˜๋ฏธํ•œ๋‹ค. โ€ป Wikipedia ์ฐธ์กฐ
  • 5. ๏ฌ ๋ฐ์ดํ„ฐ ๋งˆ์ด๋‹(data mining) โ—‹ ๋Œ€๊ทœ๋ชจ๋กœ ์ €์žฅ๋œ ๋ฐ์ดํ„ฐ ์•ˆ์—์„œ ์ฒด๊ณ„์ ์ด๊ณ  ์ž๋™์ ์œผ๋กœ ํ†ต๊ณ„์  ๊ทœ์น™์ด๋‚˜ ํŒจํ„ด์„ ์ฐพ์•„ ๋‚ด๋Š” ๊ฒƒ์ด๋‹ค. โ—‹ ๋‹ค๋ฅธ ๋ง๋กœ๋Š” KDD(๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค ์†์˜ ์ง€์‹ ๋ฐœ๊ฒฌ, knowledge-discovery in databases)๋ผ ๊ณ ๋„ ์ผ์ปซ๋Š”๋‹ค. โ€ป Wikipedia ์ฐธ์กฐ
  • 6. 1. Introduction of Machine Learning 2. Naive Bayesian Classifier 3. Linear Regression 4. Logistic Regression 5. K-means Clsutering 6. Graph Mining 7. Dimensional Reduction (PCA) 8. Spectral Clustering 9. Association Rule Mining 10. Bayesian Network 1 & 2 11. Decision tree 12. Support Vector Machine (SVM) 1 & 2 13. Hidden Markov Model (HMM) 14. Markov chain Monte Carlo(MCMC) 15. Gibbs Sampling 16. Latent Dirichlet allocation (LDA) 17. Neural Networks
  • 7. Definition( Machine Learning / ๊ธฐ๊ณ„ํ•™์Šต ) A set of methods that can automatically detect patters in data, and then use the uncovered patterns to predict patterns to predict future data, or to perform other kinds of decision making under uncertainty "Machine Learning-A Probabilistic Perspective" Kevin P. Murphy
  • 8. Supervised Learning Classification Bayesian Classifier Logistic Regression KNN Classifier Support Vector Machine (SVM) Regression Linear Regression Unsupervised Learning Clustering K-means Clustering Spectral Clustering Dimensional Reduction PCA Reinforce Learning
  • 9. A feature vector is an ๐‘›-dimensional vector of numerical features that represent some object. For example , a document , ๐‘ฅ๐‘– : ๋ฌธ์„œ ์•ˆ์—์„œ์˜ ๐‘– ๋ฒˆ์งธ ๋‹จ์–ด I love you. ๐•ฉ = ๐ผ, ๐‘™๐‘œ๐‘ฃ๐‘’, ๐‘ฆ๐‘œ๐‘ข
  • 10. ๐•ฉ ๐‘ฆ ๐‘“mapping ๋ถˆ์—ฐ์† ๊ฐ’ ๋˜๋Š” ์—ฐ์† ๊ฐ’ E-mail (Words) Spam or Not(๋ถˆ) Web Site (Words) Sports or Science or News(๋ถˆ) ํŠน์„ฑ ๋ฒกํ„ฐ ๊ฝƒ (๊ฝƒ์˜ ์ƒ๊น€์ƒˆ) Line-flower or Mass-flower(๋ถˆ) ์•„์ด์˜ ํ‚ค ์•„๋ฒ„์ง€์˜์˜ ํ‚ค(์—ฐ) ๋ฐฉ์˜ ๊ฐœ์ˆ˜, ๋ฐฉ์˜ ๋„“์ด ์ง‘ ๊ฐ’(์—ฐ) (visit, money, buy, girl, Viagra) For example, Spam mail ๐‘“
  • 11. Goal of Supervised Learning (predictive learning) To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, given a labeled set of input-output pairs ๐ท = ๐•ฉ๐‘–, ๐‘ฆ๐‘– ๐‘–=1 ๐‘ , where ๐ท is called the training set, and ๐‘ is the number of training examples. - ๐•ฉ๐‘– : ๐ท-dimensional vector of numbers โ‡’ feature vector - ๐‘ฆ๐‘– : response variable โ‡’ categories or real-values ๐‘ฆ๐‘– is categorical, the problem is classification, ๐‘ฆ๐‘– is real-valued, the problem is regression.
  • 12. What is natural grouping among these objects? Simpson's Family School Employees Females males
  • 13.
  • 14. Goal of Unsupervised Learning (descriptive learning) To find "interesting pattern" in the data, given ๐ท = ๐•ฉ๐‘– ๐‘–=1 ๐‘ , where ๐ท is called the training set, and ๐‘ is the number of training examples. - ๐•ฉ๐‘– : ๐ท-dimensional vector of numbers โ‡’ feature vector It is known knowledge discovery.
  • 15. Goal of Reinforcement Learning To learn how to act or behave when given occasional reward or punishment sinnals.
  • 16. Supervised Learning Classification Bayesian Classifier Logistic Regression KNN Classifier Support Vector Machine (SVM) Regression Linear Regression Unsupervised Learning Clustering K-means Clustering Spectral Clustering Dimensional Reduction PCA Reinforce Learning
  • 17. ๐•ฉ ๐‘ฆ ๐‘“mapping ๋ถˆ์—ฐ์† ๊ฐ’ ๋˜๋Š” ์—ฐ์† ๊ฐ’ E-mail (Words) Spam or Not(๋ถˆ) Web Site (Words) Sports or Science or News(๋ถˆ) ํŠน์„ฑ ๋ฒกํ„ฐ ๊ฝƒ (๊ฝƒ์˜ ์ƒ๊น€์ƒˆ) Line-flower or Mass-flower(๋ถˆ)
  • 18. Goal of Classification To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, where ๐‘ฆ โˆˆ {1, โ€ฆ , ๐ถ}, with ๐ถ being the number of classes. function approximation Assume that ๐‘ฆ = ๐‘“(๐•ฉ) for unknown function โ„Ž, the goal of learning is to estimate function ๐‘“ given a labeled training set, and then to make predictions using ๐‘ฆโˆ— = ๐‘“โˆ—(๐•ฉ). Then we can make predictions on novel input. way to formalize the problem
  • 19. Compute!! our "best guess" using ๐‘ฆโˆ— = ๐‘“โˆ— ๐•ฉ = ๐‘Ž๐‘Ÿ๐‘” max ๐‘=1โ€ฆ๐ถ ๐‘(๐‘ฆ = ๐‘|๐•ฉ, ๐ท) This corresponds to the most probable class label. It is known as a MAP estimate (Maximum a Posterior). Ex) mail spam filtering ๐‘ ๐‘ฆ = 1 ๐•ฉ, ๐ท) ๐‘ ๐‘ฆ = 2 ๐•ฉ, ๐ท) ... ๐‘ ๐‘ฆ = ๐ถ ๐•ฉ, ๐ท)
  • 20. Goal of Regression To learn a mapping(function) from input ๐•ฉ to output ๐‘ฆ, where ๐‘ฆ is continuous. ๐‘ฆ = ๐œ–1 + ๐œ–2 ๐‘ฅ ๐‘ฆ = ๐œ–1 + ๐œ–2 ๐‘ฅ + ๐œ€3 ๐‘ฅ2
  • 21. Supervised Learning Classification Bayesian Classifier Logistic Regression KNN Classifier Support Vector Machine (SVM) Regression Linear Regression Unsupervised Learning Clustering K-means Clustering Spectral Clustering Dimensional Reduction PCA Reinforce Learning
  • 22. What is natural grouping among these objects? Simpson's Family School Employees Females males
  • 23.
  • 24. Goal of Clustering To estimate the distribution over the number of cluster, ๐‘ ๐พ ๐ท ; this tells us if there are subpopulations within the data To estimate which cluster each point belong to.
  • 25. ๐พโˆ— = ๐‘Ž๐‘Ÿ๐‘” max ๐พ ๐‘(๐พ|๐ท) We often approximate the distribution We can infer which cluster each data point belongs to by computing ๐‘ง๐‘– โˆ— = ๐‘Ž๐‘Ÿ๐‘” max ๐‘˜ ๐‘(๐‘ง๐‘– = ๐‘˜|๐•ฉ๐‘–, ๐ท) ๐‘ ๐‘ง๐‘– = 1 ๐•ฉ, ๐ท) ๐‘ ๐‘ง๐‘– = 2 ๐•ฉ, ๐ท) ... ๐‘ ๐‘ง๐‘– = ๐ถ ๐•ฉ, ๐ท)
  • 26.
  • 27. Goal of Dimensional Reduction To reduce the dimensionality by projecting the data to a lower dimensional subspace which captures the "essence" of the data. Motivation : Although the data may appear high dimensional, there are only be a small number of degrees of variability, corresponding to latent factors. latent factor : which describe most of the variability
  • 28. ๏ฌ "Machine Learning-A Probabilistic Perspective" Kevin P. Murphy ๏ฌ http://ko.wikipedia.org/wiki/๊ธฐ๊ณ„_ํ•™์Šต