SlideShare a Scribd company logo
.
.
A Conjecture on Strongly Consistent Learning
Joe Suzuki
Osaka University
July 7, 2009
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 1 / 21
Outline
Outline
1 Stochastic Learning with Model Selection
2 Learning CP
3 Learning ARMA
4 Conjecture
5 Summary
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 2 / 21
Stochastic Learning with Model Selection
Probability Space (Ω, F, µ)
Ω: the entire set
F: the set of events over Ω
µ : F → [0, 1]: a probability measure (µ(Ω) = 1)
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 3 / 21
Stochastic Learning with Model Selection
Stochastic Learning
X : Ω → R: random variable
µX : the measure w.r.t. X
x1, · · · , xn ∈ X(Ω): n samples
 
Induction/Inference
.
µX −→ x1, · · · , xn (Random Number Generation)
x1, · · · , xn −→ µX (Stochastic Learning)
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 4 / 21
Stochastic Learning with Model Selection
Two Problems with Model Selection
Conditional Probability for Y given X
.
.
µY |X
Identify the equivalence relation in X from samples.
x ∼ x′
⇐⇒ µ(Y = y|X = x) = µ(Y = y|X = x′
) for y ∈ Y (Ω)
ARMA for {Xn}∞
n=−∞
.
Xn +
∑k
j=1 λj Xn−j ∼ N(0, σ2) with {λ}k
j=1 (0 ≤ k < ∞)
Identify the true order k from samples.
This paper
compares CP and ARMA to find how close they are.
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 5 / 21
Learning CP
CP
A ∈ F
G ⊆ F
CP of A given G (Radon-Nykodim)
.
.
∃G-measurable g : Ω → R s.t µ(A ∩ G) =
∫
G
g(ω)µ(dω) for G ∈ G.
CP µY |X
A := (Y = y), y ∈ Y (Ω)
G: generated by subsets of X.
Assumption
|Y (Ω)| < ∞
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 6 / 21
Learning CP
Applications for CP
Stochastic Decision Trees, Stochastic Horn Clauseses Y |X, etc.
.
.
Find G from samples.
G = {G1, G2, G3, G4, G5}
"!
#
"!
#
"!
#
"!
#
"!
#
"!
#
"!
#
"!
#
¡
¡¡
e
ee
¡
¡¡
e
ee
¨¨¨¨¨¨
rrrrrr
Weather
Temp
G3
Wind
G1 G2 G4 G5
Clear
Cloud
Rain
≤ 75 > 75 Yes No
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 7 / 21
Learning CP
Applications for CP (cont’d)
Finite Bayesian Networks Xi |Xj , j ∈ π(i)
.
.
Find π(i) ⊆ {1, · · · , i − 1} from samples.


X3


X4


X5


X2


X1
T
E E
d
d
ds
 
 
 ©
d
d
d‚
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 8 / 21
Learning CP
Learning CP
zi := (xi , yi ) ∈ Z(Ω) := X(Ω) × Y (Ω)
From zn := (z1, · · · , zn) ∈ Zn(Ω), find a minimal G.
G∗: true
ˆGn: estimated
Assumption
.
.
|G∗|  ∞
Two Types of Errors
G∗ ⊂ ˆGn (Over Estimation)
G∗ ̸⊆ ˆGn (Under Estimation)
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 9 / 21
Learning CP
Example: Quinlan’s Q4.5
(a) G∗
















¡¡ ee ¡¡ ee
¨¨¨¨
rrrr
Weather
Temp
3
Wind
1 2 4 5
Clear
Cloud
Rain
≤ 75  75 Yes No
(b) Overestimation




















¡¡ ee ¡¡ ee
¨¨¨¨
rrrr
Weather
Temp
Wind
Wind
1 2 5 6
3 4
Clear
Cloud
Rain
≤ 75  75 Yes No
Yes No
¡¡ ee
(c) Underestimation












¡¡ ee
¨¨¨¨
rrrr
Weather
Temp
3
4
1 2
Clear
Cloud
Rain
≤ 75  75
(d) Underestimation












¨¨¨¨
rrrr
Weather
1
Wind
Wind
4 5


¡¡ ee

¡¡ ee
2 3
Clear
Cloud
Rain
Yes No
¡¡ ee
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 10 / 21
Learning CP
Information Criteria for CP
Given zn ∈ Z(Ω), find G minimizing
I(G, zn
) := H(G, zn
) +
k(G)
2
dn
H(G, zn): Empirical Entropy (Fitness of zn to G)
k(G): the # of Parameters (Simplicity of G)
dn ≥ 0:
dn
n
→ 0
dn = log n BIC/MDL
dn = 2 AIC
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 11 / 21
Learning CP
Consistency
Consistency ( ˆGn −→ G∗ (n → ∞))
.
.
Weak Consistency Probability Convergence (O(1)  dn  o(n))
Strong Consistency Almost Surely Convergence (MDL/BIC etc.)
AIC (dn = 2) is not consistent because {dn} is too small !
Problem
What is the minimum {dn} for Strong Consistency ?
Answer (Suzuki, 2006)
dn = 2 log log n
(the Law of Iterated Logarithms)
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 12 / 21
Learning CP
Error Probability for CP
G∗ ⊂ G (Over Estimation)
.
.
µ{ω ∈ Ω|I(G, Zn
(ω))  I(G∗
, Zn
(ω))}
= µ{ω ∈ Ω|χ2
K(G)−K(G∗)(ω)  (K(G) − K(G∗
))dn}
χ2
K(G)−K(G∗) ∼ χ2 of freedom K(G) − K(G∗)
G∗ ̸⊆ G (Under Estimation)
µ{ω ∈ Ω|I(G, Zn
(ω))  I(G∗
, Zn
(ω))
diminishes exponentially with n
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 13 / 21
Learning CP
Applications for ARMA
Time Series Xi |Xi−k, · · · , Xi−1
.
.
Xi +
k∑
j=1
λj Xi−j ∼ N(0, σ2
)
Find k from samples.
Gaussian Bayesian Networks Xi |Xj , j ∈ π(i)
Xi +
∑
j∈π(i)
λj,i Xj ∼ N(0, σ2
i )
Find π(i) ⊆ {1, · · · , i − 1} from samples.
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 14 / 21
Learning ARMA
Learning ARMA
k ≥ 0
{λj }k
j=1: λi ∈ R
σ2 ∈ R0
{Xi }∞
i=−∞: Xi +
k∑
j=1
λj Xi−j = ϵi ∼ N(0, σ2
)
Given xn := (x1, · · · , xn) ∈ X1(Ω) × · · · × Xn(Ω), estimate
k: known {λj }k
j=1, σ2
k: Unknown k as well as {λj }k
j=1, σ2
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 15 / 21
Learning ARMA
Yule-Walker
If k is known, solve {ˆλj,k}k
j=1 and ˆσ2
k w.r.t.
¯x :=
1
n
n∑
i=1
xi
cj :=
1
n
n−j∑
i=1
(xi − ¯x)(xi+j − ¯x) , j = 0, · · · , k







−1 c1 c2 · · · ck
0 c0 c1 · · · ck−1
0 c1 c0 · · · ck−2
...
...
...
...
...
0 ck−1 ck−2 · · · c0














ˆσ2
k
ˆλ1,k
ˆλ2,k
...
ˆλk,k







=







−c0
−c1
−c2
...
−ck







Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 16 / 21
Learning ARMA
Information Criteria for ARMA
If k is unknown, given xn ∈ Xn(Ω), find k minimiziing
I(k, xn
) :=
1
2
log ˆσ2
k +
k
2
dn
ˆσ2
k: obtain using Yule-Walker
dn ≥ 0:
dn
n
→ 0
dn = log n BIC/MDL
dn = 2 AIC
dn = 2 log log n Hannan-Quinn (1979)
=⇒the minimum {dn} satisfying Strong Consistency
Suzuki (2006) was inspired by Hannan-Quinn (1979) !
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 17 / 21
Learning ARMA
Error Probability for ARMA
k∗: true Order
k∗  k (Under Estimation)
.
µ{ω ∈ Ω|I(k, Xn
(ω))  I(k∗, Xn
(ω))}
diminishes exponentially with n
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 18 / 21
Conjecture
Error Probability for ARMA (cont’d)
Conjecture: k∗  k (Over Estimation)
.
µ{ω ∈ Ω|I(k, Xn
(ω))  I(k∗, Xn
(ω))}
= µ{ω ∈ Ω|χ2
k−k∗
(ω)  (k − k∗)dn}
χ2
k−k∗
∼ χ2 of freedom k − k∗
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 19 / 21
Conjecture
In fact, ...
For k  k∗ and large n (use ˆσ2
k = (1 − ˆλ2
k,k)σ2
k−1),
1
2
log ˆσ2
k +
k
2
dn 
k∗
2
log σ2
k∗
+
k∗
2
dn
⇐⇒
k∑
j=k∗+1
log
ˆσ2
j
ˆσ2
j−1
 (k − k∗)dn
⇐⇒
k∑
j=k∗+1
log(1 − ˆλ2
k,k)  dn
⇐⇒
k∑
j=k∗+1
ˆλ2
k,k  (k − k∗)dn
It is very likely that
∑k
j=k∗+1
ˆλ2
k,k ∼ χ2
k−k∗
although ˆλk,k ̸∼ N(0, 1).
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 20 / 21
Summary
Summary
CP and ARMA are similar
.
.
1 Results in ARMA can be applied to CP
2 Results in Gausssian BN can be applied to finite BN.
 
The conjecture is likely enough ?
.
.
Answer: Perhaps. Several evidences.
{Zi }n
i=1: independent
Zi ∼ N(0, 1)
rankA = n − k, A2
= A =⇒ t
[Z1, · · · , Zn]A[Z1, · · · , Zn] ∼ χ2
n−k
Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 21 / 21

More Related Content

What's hot

WITMSE 2013
WITMSE 2013WITMSE 2013
WITMSE 2013
Joe Suzuki
 
Continuous and Discrete-Time Analysis of SGD
Continuous and Discrete-Time Analysis of SGDContinuous and Discrete-Time Analysis of SGD
Continuous and Discrete-Time Analysis of SGD
Valentin De Bortoli
 
Multilinear Twisted Paraproducts
Multilinear Twisted ParaproductsMultilinear Twisted Paraproducts
Multilinear Twisted Paraproducts
VjekoslavKovac1
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Bellman functions and Lp estimates for paraproducts
Bellman functions and Lp estimates for paraproductsBellman functions and Lp estimates for paraproducts
Bellman functions and Lp estimates for paraproducts
VjekoslavKovac1
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
RuFiDiM
RuFiDiMRuFiDiM
Scattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysisScattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysis
VjekoslavKovac1
 
A new axisymmetric finite element
A new axisymmetric finite elementA new axisymmetric finite element
A new axisymmetric finite element
Stefan Duprey
 
Thesis defense
Thesis defenseThesis defense
Thesis defense
Zheng Mengdi
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
On Twisted Paraproducts and some other Multilinear Singular Integrals
On Twisted Paraproducts and some other Multilinear Singular IntegralsOn Twisted Paraproducts and some other Multilinear Singular Integrals
On Twisted Paraproducts and some other Multilinear Singular Integrals
VjekoslavKovac1
 
Hyperfunction method for numerical integration and Fredholm integral equation...
Hyperfunction method for numerical integration and Fredholm integral equation...Hyperfunction method for numerical integration and Fredholm integral equation...
Hyperfunction method for numerical integration and Fredholm integral equation...
HidenoriOgata
 
Tales on two commuting transformations or flows
Tales on two commuting transformations or flowsTales on two commuting transformations or flows
Tales on two commuting transformations or flows
VjekoslavKovac1
 
Paraproducts with general dilations
Paraproducts with general dilationsParaproducts with general dilations
Paraproducts with general dilations
VjekoslavKovac1
 
Trilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operatorsTrilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operators
VjekoslavKovac1
 
Norm-variation of bilinear averages
Norm-variation of bilinear averagesNorm-variation of bilinear averages
Norm-variation of bilinear averages
VjekoslavKovac1
 
A Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cubeA Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cube
VjekoslavKovac1
 
Multilinear singular integrals with entangled structure
Multilinear singular integrals with entangled structureMultilinear singular integrals with entangled structure
Multilinear singular integrals with entangled structure
VjekoslavKovac1
 

What's hot (20)

WITMSE 2013
WITMSE 2013WITMSE 2013
WITMSE 2013
 
Continuous and Discrete-Time Analysis of SGD
Continuous and Discrete-Time Analysis of SGDContinuous and Discrete-Time Analysis of SGD
Continuous and Discrete-Time Analysis of SGD
 
A
AA
A
 
Multilinear Twisted Paraproducts
Multilinear Twisted ParaproductsMultilinear Twisted Paraproducts
Multilinear Twisted Paraproducts
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Bellman functions and Lp estimates for paraproducts
Bellman functions and Lp estimates for paraproductsBellman functions and Lp estimates for paraproducts
Bellman functions and Lp estimates for paraproducts
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
RuFiDiM
RuFiDiMRuFiDiM
RuFiDiM
 
Scattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysisScattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysis
 
A new axisymmetric finite element
A new axisymmetric finite elementA new axisymmetric finite element
A new axisymmetric finite element
 
Thesis defense
Thesis defenseThesis defense
Thesis defense
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
On Twisted Paraproducts and some other Multilinear Singular Integrals
On Twisted Paraproducts and some other Multilinear Singular IntegralsOn Twisted Paraproducts and some other Multilinear Singular Integrals
On Twisted Paraproducts and some other Multilinear Singular Integrals
 
Hyperfunction method for numerical integration and Fredholm integral equation...
Hyperfunction method for numerical integration and Fredholm integral equation...Hyperfunction method for numerical integration and Fredholm integral equation...
Hyperfunction method for numerical integration and Fredholm integral equation...
 
Tales on two commuting transformations or flows
Tales on two commuting transformations or flowsTales on two commuting transformations or flows
Tales on two commuting transformations or flows
 
Paraproducts with general dilations
Paraproducts with general dilationsParaproducts with general dilations
Paraproducts with general dilations
 
Trilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operatorsTrilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operators
 
Norm-variation of bilinear averages
Norm-variation of bilinear averagesNorm-variation of bilinear averages
Norm-variation of bilinear averages
 
A Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cubeA Szemerédi-type theorem for subsets of the unit cube
A Szemerédi-type theorem for subsets of the unit cube
 
Multilinear singular integrals with entangled structure
Multilinear singular integrals with entangled structureMultilinear singular integrals with entangled structure
Multilinear singular integrals with entangled structure
 

Similar to A Conjecture on Strongly Consistent Learning

2014 9-22
2014 9-222014 9-22
2014 9-22
Joe Suzuki
 
Bayesian network structure estimation based on the Bayesian/MDL criteria when...
Bayesian network structure estimation based on the Bayesian/MDL criteria when...Bayesian network structure estimation based on the Bayesian/MDL criteria when...
Bayesian network structure estimation based on the Bayesian/MDL criteria when...
Joe Suzuki
 
2014 9-16
2014 9-162014 9-16
2014 9-16
Joe Suzuki
 
Lecture9 xing
Lecture9 xingLecture9 xing
Lecture9 xing
Tianlu Wang
 
Properties of field induced Josephson junction(s)
Properties of field induced Josephson junction(s)Properties of field induced Josephson junction(s)
Properties of field induced Josephson junction(s)
Krzysztof Pomorski
 
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
Frank Nielsen
 
Lecture12 xing
Lecture12 xingLecture12 xing
Lecture12 xing
Tianlu Wang
 
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
Ali Ajouz
 
Stochastic Alternating Direction Method of Multipliers
Stochastic Alternating Direction Method of MultipliersStochastic Alternating Direction Method of Multipliers
Stochastic Alternating Direction Method of Multipliers
Taiji Suzuki
 
10.11648.j.pamj.20170601.11
10.11648.j.pamj.20170601.1110.11648.j.pamj.20170601.11
10.11648.j.pamj.20170601.11DAVID GIKUNJU
 
Strong convergence of an algorithm about strongly quasi nonexpansive mappings
Strong convergence of an algorithm about strongly quasi nonexpansive mappingsStrong convergence of an algorithm about strongly quasi nonexpansive mappings
Strong convergence of an algorithm about strongly quasi nonexpansive mappings
Alexander Decker
 
Free video lectures for mca
Free video lectures for mcaFree video lectures for mca
Free video lectures for mca
Edhole.com
 
Slides: Jeffreys centroids for a set of weighted histograms
Slides: Jeffreys centroids for a set of weighted histogramsSlides: Jeffreys centroids for a set of weighted histograms
Slides: Jeffreys centroids for a set of weighted histograms
Frank Nielsen
 
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier DetectionA Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
Konkuk University, Korea
 
Metrics for generativemodels
Metrics for generativemodelsMetrics for generativemodels
Metrics for generativemodels
Dai-Hai Nguyen
 
Slides_ICML v6.pdf
Slides_ICML v6.pdfSlides_ICML v6.pdf
Slides_ICML v6.pdf
QiuweiLi1
 
2.9 Cartesian products
2.9 Cartesian products2.9 Cartesian products
2.9 Cartesian products
Jan Plaza
 
zkStudyClub - cqlin: Efficient linear operations on KZG commitments
zkStudyClub - cqlin: Efficient linear operations on KZG commitments zkStudyClub - cqlin: Efficient linear operations on KZG commitments
zkStudyClub - cqlin: Efficient linear operations on KZG commitments
Alex Pruden
 
Fixed point theorems for random variables in complete metric spaces
Fixed point theorems for random variables in complete metric spacesFixed point theorems for random variables in complete metric spaces
Fixed point theorems for random variables in complete metric spaces
Alexander Decker
 
Cheatsheet probability
Cheatsheet probabilityCheatsheet probability
Cheatsheet probability
Ashish Patel
 

Similar to A Conjecture on Strongly Consistent Learning (20)

2014 9-22
2014 9-222014 9-22
2014 9-22
 
Bayesian network structure estimation based on the Bayesian/MDL criteria when...
Bayesian network structure estimation based on the Bayesian/MDL criteria when...Bayesian network structure estimation based on the Bayesian/MDL criteria when...
Bayesian network structure estimation based on the Bayesian/MDL criteria when...
 
2014 9-16
2014 9-162014 9-16
2014 9-16
 
Lecture9 xing
Lecture9 xingLecture9 xing
Lecture9 xing
 
Properties of field induced Josephson junction(s)
Properties of field induced Josephson junction(s)Properties of field induced Josephson junction(s)
Properties of field induced Josephson junction(s)
 
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
Slides: Total Jensen divergences: Definition, Properties and k-Means++ Cluste...
 
Lecture12 xing
Lecture12 xingLecture12 xing
Lecture12 xing
 
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
Hecke Operators on Jacobi Forms of Lattice Index and the Relation to Elliptic...
 
Stochastic Alternating Direction Method of Multipliers
Stochastic Alternating Direction Method of MultipliersStochastic Alternating Direction Method of Multipliers
Stochastic Alternating Direction Method of Multipliers
 
10.11648.j.pamj.20170601.11
10.11648.j.pamj.20170601.1110.11648.j.pamj.20170601.11
10.11648.j.pamj.20170601.11
 
Strong convergence of an algorithm about strongly quasi nonexpansive mappings
Strong convergence of an algorithm about strongly quasi nonexpansive mappingsStrong convergence of an algorithm about strongly quasi nonexpansive mappings
Strong convergence of an algorithm about strongly quasi nonexpansive mappings
 
Free video lectures for mca
Free video lectures for mcaFree video lectures for mca
Free video lectures for mca
 
Slides: Jeffreys centroids for a set of weighted histograms
Slides: Jeffreys centroids for a set of weighted histogramsSlides: Jeffreys centroids for a set of weighted histograms
Slides: Jeffreys centroids for a set of weighted histograms
 
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier DetectionA Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
 
Metrics for generativemodels
Metrics for generativemodelsMetrics for generativemodels
Metrics for generativemodels
 
Slides_ICML v6.pdf
Slides_ICML v6.pdfSlides_ICML v6.pdf
Slides_ICML v6.pdf
 
2.9 Cartesian products
2.9 Cartesian products2.9 Cartesian products
2.9 Cartesian products
 
zkStudyClub - cqlin: Efficient linear operations on KZG commitments
zkStudyClub - cqlin: Efficient linear operations on KZG commitments zkStudyClub - cqlin: Efficient linear operations on KZG commitments
zkStudyClub - cqlin: Efficient linear operations on KZG commitments
 
Fixed point theorems for random variables in complete metric spaces
Fixed point theorems for random variables in complete metric spacesFixed point theorems for random variables in complete metric spaces
Fixed point theorems for random variables in complete metric spaces
 
Cheatsheet probability
Cheatsheet probabilityCheatsheet probability
Cheatsheet probability
 

More from Joe Suzuki

RとPythonを比較する
RとPythonを比較するRとPythonを比較する
RとPythonを比較する
Joe Suzuki
 
R集会@統数研
R集会@統数研R集会@統数研
R集会@統数研
Joe Suzuki
 
E-learning Development of Statistics and in Duex: Practical Approaches and Th...
E-learning Development of Statistics and in Duex: Practical Approaches and Th...E-learning Development of Statistics and in Duex: Practical Approaches and Th...
E-learning Development of Statistics and in Duex: Practical Approaches and Th...
Joe Suzuki
 
分枝限定法でモデル選択の計算量を低減する
分枝限定法でモデル選択の計算量を低減する分枝限定法でモデル選択の計算量を低減する
分枝限定法でモデル選択の計算量を低減する
Joe Suzuki
 
連続変量を含む条件付相互情報量の推定
連続変量を含む条件付相互情報量の推定連続変量を含む条件付相互情報量の推定
連続変量を含む条件付相互情報量の推定
Joe Suzuki
 
E-learning Design and Development for Data Science in Osaka University
E-learning Design and Development for Data Science in Osaka UniversityE-learning Design and Development for Data Science in Osaka University
E-learning Design and Development for Data Science in Osaka University
Joe Suzuki
 
UAI 2017
UAI 2017UAI 2017
UAI 2017
Joe Suzuki
 
AMBN2017 サテライトワークショップ
AMBN2017 サテライトワークショップAMBN2017 サテライトワークショップ
AMBN2017 サテライトワークショップ
Joe Suzuki
 
CRAN Rパッケージ BNSLの概要
CRAN Rパッケージ BNSLの概要CRAN Rパッケージ BNSLの概要
CRAN Rパッケージ BNSLの概要
Joe Suzuki
 
Forest Learning from Data
Forest Learning from DataForest Learning from Data
Forest Learning from Data
Joe Suzuki
 
A Bayesian Approach to Data Compression
A Bayesian Approach to Data CompressionA Bayesian Approach to Data Compression
A Bayesian Approach to Data Compression
Joe Suzuki
 
研究紹介(学生向け)
研究紹介(学生向け)研究紹介(学生向け)
研究紹介(学生向け)
Joe Suzuki
 
Efficietly Learning Bayesian Network Structures based on the B&B Strategy: A ...
Efficietly Learning Bayesian Network Structuresbased on the B&B Strategy: A ...Efficietly Learning Bayesian Network Structuresbased on the B&B Strategy: A ...
Efficietly Learning Bayesian Network Structures based on the B&B Strategy: A ...
Joe Suzuki
 
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
Joe Suzuki
 
2016 7-13
2016 7-132016 7-13
2016 7-13
Joe Suzuki
 
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
Joe Suzuki
 
連続変量を含む相互情報量の推定
連続変量を含む相互情報量の推定連続変量を含む相互情報量の推定
連続変量を含む相互情報量の推定
Joe Suzuki
 
Jeffreys' and BDeu Priors for Model Selection
Jeffreys' and BDeu Priors for Model SelectionJeffreys' and BDeu Priors for Model Selection
Jeffreys' and BDeu Priors for Model Selection
Joe Suzuki
 
離散と連続の入り混じった相互情報量を推定して、 SNP と遺伝子発現量の因果関係をさぐる
離散と連続の入り混じった相互情報量を推定して、SNP と遺伝子発現量の因果関係をさぐる離散と連続の入り混じった相互情報量を推定して、SNP と遺伝子発現量の因果関係をさぐる
離散と連続の入り混じった相互情報量を推定して、 SNP と遺伝子発現量の因果関係をさぐる
Joe Suzuki
 
MaCaulay2 Miuraパッケージの開発と今後
MaCaulay2 Miuraパッケージの開発と今後MaCaulay2 Miuraパッケージの開発と今後
MaCaulay2 Miuraパッケージの開発と今後
Joe Suzuki
 

More from Joe Suzuki (20)

RとPythonを比較する
RとPythonを比較するRとPythonを比較する
RとPythonを比較する
 
R集会@統数研
R集会@統数研R集会@統数研
R集会@統数研
 
E-learning Development of Statistics and in Duex: Practical Approaches and Th...
E-learning Development of Statistics and in Duex: Practical Approaches and Th...E-learning Development of Statistics and in Duex: Practical Approaches and Th...
E-learning Development of Statistics and in Duex: Practical Approaches and Th...
 
分枝限定法でモデル選択の計算量を低減する
分枝限定法でモデル選択の計算量を低減する分枝限定法でモデル選択の計算量を低減する
分枝限定法でモデル選択の計算量を低減する
 
連続変量を含む条件付相互情報量の推定
連続変量を含む条件付相互情報量の推定連続変量を含む条件付相互情報量の推定
連続変量を含む条件付相互情報量の推定
 
E-learning Design and Development for Data Science in Osaka University
E-learning Design and Development for Data Science in Osaka UniversityE-learning Design and Development for Data Science in Osaka University
E-learning Design and Development for Data Science in Osaka University
 
UAI 2017
UAI 2017UAI 2017
UAI 2017
 
AMBN2017 サテライトワークショップ
AMBN2017 サテライトワークショップAMBN2017 サテライトワークショップ
AMBN2017 サテライトワークショップ
 
CRAN Rパッケージ BNSLの概要
CRAN Rパッケージ BNSLの概要CRAN Rパッケージ BNSLの概要
CRAN Rパッケージ BNSLの概要
 
Forest Learning from Data
Forest Learning from DataForest Learning from Data
Forest Learning from Data
 
A Bayesian Approach to Data Compression
A Bayesian Approach to Data CompressionA Bayesian Approach to Data Compression
A Bayesian Approach to Data Compression
 
研究紹介(学生向け)
研究紹介(学生向け)研究紹介(学生向け)
研究紹介(学生向け)
 
Efficietly Learning Bayesian Network Structures based on the B&B Strategy: A ...
Efficietly Learning Bayesian Network Structuresbased on the B&B Strategy: A ...Efficietly Learning Bayesian Network Structuresbased on the B&B Strategy: A ...
Efficietly Learning Bayesian Network Structures based on the B&B Strategy: A ...
 
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
Forest Learning based on the Chow-Liu Algorithm and its Application to Genom...
 
2016 7-13
2016 7-132016 7-13
2016 7-13
 
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
Structure Learning of Bayesian Networks with p Nodes from n Samples when n&lt...
 
連続変量を含む相互情報量の推定
連続変量を含む相互情報量の推定連続変量を含む相互情報量の推定
連続変量を含む相互情報量の推定
 
Jeffreys' and BDeu Priors for Model Selection
Jeffreys' and BDeu Priors for Model SelectionJeffreys' and BDeu Priors for Model Selection
Jeffreys' and BDeu Priors for Model Selection
 
離散と連続の入り混じった相互情報量を推定して、 SNP と遺伝子発現量の因果関係をさぐる
離散と連続の入り混じった相互情報量を推定して、SNP と遺伝子発現量の因果関係をさぐる離散と連続の入り混じった相互情報量を推定して、SNP と遺伝子発現量の因果関係をさぐる
離散と連続の入り混じった相互情報量を推定して、 SNP と遺伝子発現量の因果関係をさぐる
 
MaCaulay2 Miuraパッケージの開発と今後
MaCaulay2 Miuraパッケージの開発と今後MaCaulay2 Miuraパッケージの開発と今後
MaCaulay2 Miuraパッケージの開発と今後
 

Recently uploaded

Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
NathanBaughman3
 
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Ana Luísa Pinho
 
Seminar of U.V. Spectroscopy by SAMIR PANDA
 Seminar of U.V. Spectroscopy by SAMIR PANDA Seminar of U.V. Spectroscopy by SAMIR PANDA
Seminar of U.V. Spectroscopy by SAMIR PANDA
SAMIR PANDA
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram Staining
Areesha Ahmad
 
platelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptxplatelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptx
muralinath2
 
Richard's entangled aventures in wonderland
Richard's entangled aventures in wonderlandRichard's entangled aventures in wonderland
Richard's entangled aventures in wonderland
Richard Gill
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
Scintica Instrumentation
 
Richard's aventures in two entangled wonderlands
Richard's aventures in two entangled wonderlandsRichard's aventures in two entangled wonderlands
Richard's aventures in two entangled wonderlands
Richard Gill
 
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptxBody fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
muralinath2
 
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
University of Maribor
 
NuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final versionNuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final version
pablovgd
 
in vitro propagation of plants lecture note.pptx
in vitro propagation of plants lecture note.pptxin vitro propagation of plants lecture note.pptx
in vitro propagation of plants lecture note.pptx
yusufzako14
 
Mammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also FunctionsMammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also Functions
YOGESH DOGRA
 
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATIONPRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
ChetanK57
 
Citrus Greening Disease and its Management
Citrus Greening Disease and its ManagementCitrus Greening Disease and its Management
Citrus Greening Disease and its Management
subedisuryaofficial
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
RenuJangid3
 
Cancer cell metabolism: special Reference to Lactate Pathway
Cancer cell metabolism: special Reference to Lactate PathwayCancer cell metabolism: special Reference to Lactate Pathway
Cancer cell metabolism: special Reference to Lactate Pathway
AADYARAJPANDEY1
 
Structures and textures of metamorphic rocks
Structures and textures of metamorphic rocksStructures and textures of metamorphic rocks
Structures and textures of metamorphic rocks
kumarmathi863
 
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdfSCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SELF-EXPLANATORY
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
silvermistyshot
 

Recently uploaded (20)

Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
 
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
 
Seminar of U.V. Spectroscopy by SAMIR PANDA
 Seminar of U.V. Spectroscopy by SAMIR PANDA Seminar of U.V. Spectroscopy by SAMIR PANDA
Seminar of U.V. Spectroscopy by SAMIR PANDA
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram Staining
 
platelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptxplatelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptx
 
Richard's entangled aventures in wonderland
Richard's entangled aventures in wonderlandRichard's entangled aventures in wonderland
Richard's entangled aventures in wonderland
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
 
Richard's aventures in two entangled wonderlands
Richard's aventures in two entangled wonderlandsRichard's aventures in two entangled wonderlands
Richard's aventures in two entangled wonderlands
 
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptxBody fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
Body fluids_tonicity_dehydration_hypovolemia_hypervolemia.pptx
 
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
 
NuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final versionNuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final version
 
in vitro propagation of plants lecture note.pptx
in vitro propagation of plants lecture note.pptxin vitro propagation of plants lecture note.pptx
in vitro propagation of plants lecture note.pptx
 
Mammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also FunctionsMammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also Functions
 
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATIONPRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
 
Citrus Greening Disease and its Management
Citrus Greening Disease and its ManagementCitrus Greening Disease and its Management
Citrus Greening Disease and its Management
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
 
Cancer cell metabolism: special Reference to Lactate Pathway
Cancer cell metabolism: special Reference to Lactate PathwayCancer cell metabolism: special Reference to Lactate Pathway
Cancer cell metabolism: special Reference to Lactate Pathway
 
Structures and textures of metamorphic rocks
Structures and textures of metamorphic rocksStructures and textures of metamorphic rocks
Structures and textures of metamorphic rocks
 
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdfSCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
 

A Conjecture on Strongly Consistent Learning

  • 1. . . A Conjecture on Strongly Consistent Learning Joe Suzuki Osaka University July 7, 2009 Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 1 / 21
  • 2. Outline Outline 1 Stochastic Learning with Model Selection 2 Learning CP 3 Learning ARMA 4 Conjecture 5 Summary Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 2 / 21
  • 3. Stochastic Learning with Model Selection Probability Space (Ω, F, µ) Ω: the entire set F: the set of events over Ω µ : F → [0, 1]: a probability measure (µ(Ω) = 1) Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 3 / 21
  • 4. Stochastic Learning with Model Selection Stochastic Learning X : Ω → R: random variable µX : the measure w.r.t. X x1, · · · , xn ∈ X(Ω): n samples   Induction/Inference . µX −→ x1, · · · , xn (Random Number Generation) x1, · · · , xn −→ µX (Stochastic Learning) Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 4 / 21
  • 5. Stochastic Learning with Model Selection Two Problems with Model Selection Conditional Probability for Y given X . . µY |X Identify the equivalence relation in X from samples. x ∼ x′ ⇐⇒ µ(Y = y|X = x) = µ(Y = y|X = x′ ) for y ∈ Y (Ω) ARMA for {Xn}∞ n=−∞ . Xn + ∑k j=1 λj Xn−j ∼ N(0, σ2) with {λ}k j=1 (0 ≤ k < ∞) Identify the true order k from samples. This paper compares CP and ARMA to find how close they are. Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 5 / 21
  • 6. Learning CP CP A ∈ F G ⊆ F CP of A given G (Radon-Nykodim) . . ∃G-measurable g : Ω → R s.t µ(A ∩ G) = ∫ G g(ω)µ(dω) for G ∈ G. CP µY |X A := (Y = y), y ∈ Y (Ω) G: generated by subsets of X. Assumption |Y (Ω)| < ∞ Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 6 / 21
  • 7. Learning CP Applications for CP Stochastic Decision Trees, Stochastic Horn Clauseses Y |X, etc. . . Find G from samples. G = {G1, G2, G3, G4, G5} "! # "! # "! # "! # "! # "! # "! # "! # ¡ ¡¡ e ee ¡ ¡¡ e ee ¨¨¨¨¨¨ rrrrrr Weather Temp G3 Wind G1 G2 G4 G5 Clear Cloud Rain ≤ 75 > 75 Yes No Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 7 / 21
  • 8. Learning CP Applications for CP (cont’d) Finite Bayesian Networks Xi |Xj , j ∈ π(i) . . Find π(i) ⊆ {1, · · · , i − 1} from samples. X3 X4 X5 X2 X1 T E E d d ds      © d d d‚ Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 8 / 21
  • 9. Learning CP Learning CP zi := (xi , yi ) ∈ Z(Ω) := X(Ω) × Y (Ω) From zn := (z1, · · · , zn) ∈ Zn(Ω), find a minimal G. G∗: true ˆGn: estimated Assumption . . |G∗| ∞ Two Types of Errors G∗ ⊂ ˆGn (Over Estimation) G∗ ̸⊆ ˆGn (Under Estimation) Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 9 / 21
  • 10. Learning CP Example: Quinlan’s Q4.5 (a) G∗ ¡¡ ee ¡¡ ee ¨¨¨¨ rrrr Weather Temp 3 Wind 1 2 4 5 Clear Cloud Rain ≤ 75 75 Yes No (b) Overestimation ¡¡ ee ¡¡ ee ¨¨¨¨ rrrr Weather Temp Wind Wind 1 2 5 6 3 4 Clear Cloud Rain ≤ 75 75 Yes No Yes No ¡¡ ee (c) Underestimation ¡¡ ee ¨¨¨¨ rrrr Weather Temp 3 4 1 2 Clear Cloud Rain ≤ 75 75 (d) Underestimation ¨¨¨¨ rrrr Weather 1 Wind Wind 4 5 ¡¡ ee ¡¡ ee 2 3 Clear Cloud Rain Yes No ¡¡ ee Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 10 / 21
  • 11. Learning CP Information Criteria for CP Given zn ∈ Z(Ω), find G minimizing I(G, zn ) := H(G, zn ) + k(G) 2 dn H(G, zn): Empirical Entropy (Fitness of zn to G) k(G): the # of Parameters (Simplicity of G) dn ≥ 0: dn n → 0 dn = log n BIC/MDL dn = 2 AIC Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 11 / 21
  • 12. Learning CP Consistency Consistency ( ˆGn −→ G∗ (n → ∞)) . . Weak Consistency Probability Convergence (O(1) dn o(n)) Strong Consistency Almost Surely Convergence (MDL/BIC etc.) AIC (dn = 2) is not consistent because {dn} is too small ! Problem What is the minimum {dn} for Strong Consistency ? Answer (Suzuki, 2006) dn = 2 log log n (the Law of Iterated Logarithms) Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 12 / 21
  • 13. Learning CP Error Probability for CP G∗ ⊂ G (Over Estimation) . . µ{ω ∈ Ω|I(G, Zn (ω)) I(G∗ , Zn (ω))} = µ{ω ∈ Ω|χ2 K(G)−K(G∗)(ω) (K(G) − K(G∗ ))dn} χ2 K(G)−K(G∗) ∼ χ2 of freedom K(G) − K(G∗) G∗ ̸⊆ G (Under Estimation) µ{ω ∈ Ω|I(G, Zn (ω)) I(G∗ , Zn (ω)) diminishes exponentially with n Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 13 / 21
  • 14. Learning CP Applications for ARMA Time Series Xi |Xi−k, · · · , Xi−1 . . Xi + k∑ j=1 λj Xi−j ∼ N(0, σ2 ) Find k from samples. Gaussian Bayesian Networks Xi |Xj , j ∈ π(i) Xi + ∑ j∈π(i) λj,i Xj ∼ N(0, σ2 i ) Find π(i) ⊆ {1, · · · , i − 1} from samples. Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 14 / 21
  • 15. Learning ARMA Learning ARMA k ≥ 0 {λj }k j=1: λi ∈ R σ2 ∈ R0 {Xi }∞ i=−∞: Xi + k∑ j=1 λj Xi−j = ϵi ∼ N(0, σ2 ) Given xn := (x1, · · · , xn) ∈ X1(Ω) × · · · × Xn(Ω), estimate k: known {λj }k j=1, σ2 k: Unknown k as well as {λj }k j=1, σ2 Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 15 / 21
  • 16. Learning ARMA Yule-Walker If k is known, solve {ˆλj,k}k j=1 and ˆσ2 k w.r.t. ¯x := 1 n n∑ i=1 xi cj := 1 n n−j∑ i=1 (xi − ¯x)(xi+j − ¯x) , j = 0, · · · , k        −1 c1 c2 · · · ck 0 c0 c1 · · · ck−1 0 c1 c0 · · · ck−2 ... ... ... ... ... 0 ck−1 ck−2 · · · c0               ˆσ2 k ˆλ1,k ˆλ2,k ... ˆλk,k        =        −c0 −c1 −c2 ... −ck        Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 16 / 21
  • 17. Learning ARMA Information Criteria for ARMA If k is unknown, given xn ∈ Xn(Ω), find k minimiziing I(k, xn ) := 1 2 log ˆσ2 k + k 2 dn ˆσ2 k: obtain using Yule-Walker dn ≥ 0: dn n → 0 dn = log n BIC/MDL dn = 2 AIC dn = 2 log log n Hannan-Quinn (1979) =⇒the minimum {dn} satisfying Strong Consistency Suzuki (2006) was inspired by Hannan-Quinn (1979) ! Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 17 / 21
  • 18. Learning ARMA Error Probability for ARMA k∗: true Order k∗ k (Under Estimation) . µ{ω ∈ Ω|I(k, Xn (ω)) I(k∗, Xn (ω))} diminishes exponentially with n Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 18 / 21
  • 19. Conjecture Error Probability for ARMA (cont’d) Conjecture: k∗ k (Over Estimation) . µ{ω ∈ Ω|I(k, Xn (ω)) I(k∗, Xn (ω))} = µ{ω ∈ Ω|χ2 k−k∗ (ω) (k − k∗)dn} χ2 k−k∗ ∼ χ2 of freedom k − k∗ Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 19 / 21
  • 20. Conjecture In fact, ... For k k∗ and large n (use ˆσ2 k = (1 − ˆλ2 k,k)σ2 k−1), 1 2 log ˆσ2 k + k 2 dn k∗ 2 log σ2 k∗ + k∗ 2 dn ⇐⇒ k∑ j=k∗+1 log ˆσ2 j ˆσ2 j−1 (k − k∗)dn ⇐⇒ k∑ j=k∗+1 log(1 − ˆλ2 k,k) dn ⇐⇒ k∑ j=k∗+1 ˆλ2 k,k (k − k∗)dn It is very likely that ∑k j=k∗+1 ˆλ2 k,k ∼ χ2 k−k∗ although ˆλk,k ̸∼ N(0, 1). Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 20 / 21
  • 21. Summary Summary CP and ARMA are similar . . 1 Results in ARMA can be applied to CP 2 Results in Gausssian BN can be applied to finite BN.   The conjecture is likely enough ? . . Answer: Perhaps. Several evidences. {Zi }n i=1: independent Zi ∼ N(0, 1) rankA = n − k, A2 = A =⇒ t [Z1, · · · , Zn]A[Z1, · · · , Zn] ∼ χ2 n−k Joe Suzuki (Osaka University) A Conjecture on Strongly Consistent Learning July 7, 2009 21 / 21