SlideShare a Scribd company logo
1 of 29
Download to read offline
On Information Geometry and its Applications
                                     .

                                                                                       .
                                       Masaki Asano (M1)
                                       Osaka City University
                                       My advisor is Prof. Ohnita

                                                 July 24, 2012




Masaki Asano (Osaka City University)    On Information Geometry and its Applications       July 24, 2012   1 / 16
Contents of this talk




1   Statistical model, Fisher metric and α-connection


2   Statistical manifold


3   Applications




Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   2 / 16
Statistical model, Fisher metric and α-connection


     Statistical model

  (Ω, F , P)           :     a probability space
      Ξ                :     an open domain of Rn (a parameter space)

Definition 1.1
S is a statistical model or parametric model on Ω
 def
⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such
that

            S = p(x; ξ)                      p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn ,
                                         Ω

where P(A) =                 A
                                 p(x; ξ)dx,             (A ∈ F ).

We assume S is a smooth manifold with local coordinate system Ξ.
                                                     .

Masaki Asano (Osaka City University)       On Information Geometry and its Applications   July 24, 2012   3 / 16
Statistical model, Fisher metric and α-connection


     Statistical model

  (Ω, F , P)           :     a probability space
      Ξ                :     an open domain of Rn (a parameter space)

Definition 1.1
S is a statistical model or parametric model on Ω
 def
⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such
that

            S = p(x; ξ)                      p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn ,
                                         Ω

where P(A) =                 A
                                 p(x; ξ)dx,             (A ∈ F ).

We assume S is a smooth manifold with local coordinate system Ξ.
                                                     .

Masaki Asano (Osaka City University)       On Information Geometry and its Applications   July 24, 2012   3 / 16
Statistical model, Fisher metric and α-connection


     Statistical model

  (Ω, F , P)           :     a probability space
      Ξ                :     an open domain of Rn (a parameter space)

Definition 1.1
S is a statistical model or parametric model on Ω
 def
⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such
that

            S = p(x; ξ)                      p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn ,
                                         Ω

where P(A) =                 A
                                 p(x; ξ)dx,             (A ∈ F ).

We assume S is a smooth manifold with local coordinate system Ξ.
                                                     .

Masaki Asano (Osaka City University)       On Information Geometry and its Applications   July 24, 2012   3 / 16
Statistical model, Fisher metric and α-connection


     Fisher metric
For simplicity,

  Eξ [ f ] =            f (x)p(x; ξ)dx,                   (the expectation of f (x) w.r.t. p(x; ξ))
                    Ω
         lξ = l(x; ξ) = log p(x; ξ)                                          (the information of p(x; ξ))
               ∂
         ∂i = i
              ∂ξ


Definition 1.2 (Fisher information matrix)
g = (gi j ) is the Fisher information matrix of S .
def
⇐⇒

                                              gi j (ξ) := Eξ ∂i lξ ∂ j lξ

Masaki Asano (Osaka City University)       On Information Geometry and its Applications       July 24, 2012   4 / 16
Statistical model, Fisher metric and α-connection


     Fisher metric
For simplicity,

  Eξ [ f ] =            f (x)p(x; ξ)dx,                   (the expectation of f (x) w.r.t. p(x; ξ))
                    Ω
         lξ = l(x; ξ) = log p(x; ξ)                                          (the information of p(x; ξ))
               ∂
         ∂i = i
              ∂ξ


Definition 1.2 (Fisher information matrix)
g = (gi j ) is the Fisher information matrix of S .
def
⇐⇒

                                              gi j (ξ) := Eξ ∂i lξ ∂ j lξ

Masaki Asano (Osaka City University)       On Information Geometry and its Applications       July 24, 2012   4 / 16
Statistical model, Fisher metric and α-connection




Proposition 1.3
The following conditions are equivalent.
  • g is positive definite.
    • {∂1 pξ , · · · , ∂n pξ } are linearly independent.
                                                                                              .
    • {∂1 lξ , · · · , ∂n lξ } are linearly independent.

We assume that one of the above conditions is satisfied and gi j (ξ)
is finite for all i, j, ξ .
=⇒ We can define a Riemannian metric on S .
                                                                                          .
=⇒ The metric g is called Fisher metric.



Masaki Asano (Osaka City University)       On Information Geometry and its Applications           July 24, 2012   5 / 16
Statistical model, Fisher metric and α-connection




Proposition 1.3
The following conditions are equivalent.
  • g is positive definite.
    • {∂1 pξ , · · · , ∂n pξ } are linearly independent.
                                                                                              .
    • {∂1 lξ , · · · , ∂n lξ } are linearly independent.

We assume that one of the above conditions is satisfied and gi j (ξ)
is finite for all i, j, ξ .
=⇒ We can define a Riemannian metric on S .
                                                                                          .
=⇒ The metric g is called Fisher metric.



Masaki Asano (Osaka City University)       On Information Geometry and its Applications           July 24, 2012   5 / 16
Statistical model, Fisher metric and α-connection




Proposition 1.3
The following conditions are equivalent.
  • g is positive definite.
    • {∂1 pξ , · · · , ∂n pξ } are linearly independent.
                                                                                              .
    • {∂1 lξ , · · · , ∂n lξ } are linearly independent.

We assume that one of the above conditions is satisfied and gi j (ξ)
is finite for all i, j, ξ .
=⇒ We can define a Riemannian metric on S .
                                                                                          .
=⇒ The metric g is called Fisher metric.



Masaki Asano (Osaka City University)       On Information Geometry and its Applications           July 24, 2012   5 / 16
Statistical model, Fisher metric and α-connection




Proposition 1.3
The following conditions are equivalent.
  • g is positive definite.
    • {∂1 pξ , · · · , ∂n pξ } are linearly independent.
                                                                                              .
    • {∂1 lξ , · · · , ∂n lξ } are linearly independent.

We assume that one of the above conditions is satisfied and gi j (ξ)
is finite for all i, j, ξ .
=⇒ We can define a Riemannian metric on S .
                                                                                          .
=⇒ The metric g is called Fisher metric.



Masaki Asano (Osaka City University)       On Information Geometry and its Applications           July 24, 2012   5 / 16
Statistical model, Fisher metric and α-connection


      α-connection


Definition 1.4
For α ∈ R, we define the α-connection                                          (α)
                                                                                    by the following formula,

                             (α)                                         1−α
                     g       ∂i ∂ j , ∂k       = E ∂i ∂ j l ξ +              ∂i lξ ∂ j lξ ∂k lξ .
                                                                          2                     .
By the definition of α-connection,
          (0)
(1)             is the Levi-Civita connection of the Fisher metric g.
(2)       (α)
                is torsion-free (∀ α),
                    (α)                     (α)    (α)
       i.e. T             (X, Y) :=          X Y − Y X              − [X, Y] ≡ 0.              .
           (α)                           (α)
(3) (      X g)(Y, Z)           =(       Y g)(X, Z).




Masaki Asano (Osaka City University)        On Information Geometry and its Applications            July 24, 2012   6 / 16
Statistical model, Fisher metric and α-connection


      α-connection


Definition 1.4
For α ∈ R, we define the α-connection                                          (α)
                                                                                    by the following formula,

                             (α)                                         1−α
                     g       ∂i ∂ j , ∂k       = E ∂i ∂ j l ξ +              ∂i lξ ∂ j lξ ∂k lξ .
                                                                          2                     .
By the definition of α-connection,
          (0)
(1)             is the Levi-Civita connection of the Fisher metric g.
(2)       (α)
                is torsion-free (∀ α),
                    (α)                     (α)    (α)
       i.e. T             (X, Y) :=          X Y − Y X              − [X, Y] ≡ 0.              .
           (α)                           (α)
(3) (      X g)(Y, Z)           =(       Y g)(X, Z).




Masaki Asano (Osaka City University)        On Information Geometry and its Applications            July 24, 2012   6 / 16
Statistical model, Fisher metric and α-connection




Example 1.5 (Normal distribution)
ξ = (µ, σ) ∈ Ξ = R × R+ (Upper half-plane)
µ : mean (−∞ < µ < ∞), σ : standard deviation (0 < σ < ∞),

                                                                       1                  (x − µ)2
                  S = p(x; ξ) p(x; µ, σ) = √                                   exp −
                                                                      2πσ                   2σ2
                                                                                                     .
                          1                                1 0                         1
. The Fisher metric : g = 2                                    . The curvature of S : − .
                         σ                                 0 2                         2

Hyperbolic plane.
(x, y) ∈ H = {(x, y) ∈ R2 | y > 0}
                              1 1 0
The Poincare metric : g = 2         . The curvature of H : −1.
                              y 0 1
                                                                                                .

Masaki Asano (Osaka City University)       On Information Geometry and its Applications                  July 24, 2012   7 / 16
Statistical model, Fisher metric and α-connection




Example 1.5 (Normal distribution)
ξ = (µ, σ) ∈ Ξ = R × R+ (Upper half-plane)
µ : mean (−∞ < µ < ∞), σ : standard deviation (0 < σ < ∞),

                                                                       1                  (x − µ)2
                  S = p(x; ξ) p(x; µ, σ) = √                                   exp −
                                                                      2πσ                   2σ2
                                                                                                     .
                          1                                1 0                         1
. The Fisher metric : g = 2                                    . The curvature of S : − .
                         σ                                 0 2                         2

Hyperbolic plane.
(x, y) ∈ H = {(x, y) ∈ R2 | y > 0}
                              1 1 0
The Poincare metric : g = 2         . The curvature of H : −1.
                              y 0 1
                                                                                                .

Masaki Asano (Osaka City University)       On Information Geometry and its Applications                  July 24, 2012   7 / 16
Statistical manifold


     Statistical manifold



  (M, g)         :     a Riemannian manifold
                 :     a torsion-free affine connection on M
                       i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0

Definition 2.1
(M, , g) is a statistical manifold
 def
⇐⇒ g is totally symmetric (0, 3)-tensor field.

We can find that connection                             satisfies the properties (1)-(3) of
α-connection (α) .

                                                                                      .


Masaki Asano (Osaka City University)   On Information Geometry and its Applications       July 24, 2012   8 / 16
Statistical manifold


     Statistical manifold



  (M, g)         :     a Riemannian manifold
                 :     a torsion-free affine connection on M
                       i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0

Definition 2.1
(M, , g) is a statistical manifold
 def
⇐⇒ g is totally symmetric (0, 3)-tensor field.

We can find that connection                             satisfies the properties (1)-(3) of
α-connection (α) .

                                                                                      .


Masaki Asano (Osaka City University)   On Information Geometry and its Applications       July 24, 2012   8 / 16
Statistical manifold


     Statistical manifold



  (M, g)         :     a Riemannian manifold
                 :     a torsion-free affine connection on M
                       i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0

Definition 2.1
(M, , g) is a statistical manifold
 def
⇐⇒ g is totally symmetric (0, 3)-tensor field.

We can find that connection                             satisfies the properties (1)-(3) of
α-connection (α) .

                                                                                      .


Masaki Asano (Osaka City University)   On Information Geometry and its Applications       July 24, 2012   8 / 16
Applications


     Application 1




                                        Statistical Models

         Theorem             Theorem
    (Hon Van Le(2005))⇑⇓(Hon Van Le(2005))
      ˆ   ˆ   ˆ           ˆ   ˆ   ˆ

                                       Statistical Manifolds



Masaki Asano (Osaka City University)     On Information Geometry and its Applications   July 24, 2012   9 / 16
Applications


     Application 1




                                        Statistical Models

         Theorem             Theorem
    (Hon Van Le(2005))⇑⇓(Hon Van Le(2005))
      ˆ   ˆ   ˆ           ˆ   ˆ   ˆ

                                       Statistical Manifolds



Masaki Asano (Osaka City University)     On Information Geometry and its Applications   July 24, 2012   10 / 16
Applications


     Application 2


Information geometry is related to ...
    • statistics,
    • information theory,
    • (almost) complex geometry,
    • symplectic geometry,
    • contact geometry,
    • Wasserstein geometry...etc.
However, I am not sure.
These are that I want to study in future.



Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   11 / 16
Applications


     Application 2


Information geometry is related to ...
    • statistics,
    • information theory,
    • (almost) complex geometry,
    • symplectic geometry,
    • contact geometry,
    • Wasserstein geometry...etc.
However, I am not sure.
These are that I want to study in future.



Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   11 / 16
Applications


     Application 2


Information geometry is related to ...
    • statistics,
    • information theory,
    • (almost) complex geometry,
    • symplectic geometry,
    • contact geometry,
    • Wasserstein geometry...etc.
However, I am not sure.
These are that I want to study in future.



Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   11 / 16
Applications




      Thank you for your attention!!




Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   12 / 16
Applications


     References



    • S. Amari and H. Nagaoka, Methods of Information Geometry,
      Trans. of Math. Monograph, AMS, 2000.
    • M. Gromov, Partial differential relations. Springer-Verlag,
      Berlin, 1986.
    • H-V. Le, Statistical manifolds are statistical models. J. Geom.
            ˆ
      84 (2005), no. 1-2, 83-93.
    • J. Nash, C 1 -isometric imbeddings. Ann. of Math. 60, (1954).
      383-396.




Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   13 / 16
Applications


     For Question




Proof of Proposition 1.3.
For any n-dimensional vector c = t (c1 , c2 , . . . , cn ) (t denotes
transpose),
                                                                                             2
                            t
                            cgc =             ci c j gi j (ξ) = Eξ                ci ∂i lξ
                                       i, j




                                                                                                 .



Masaki Asano (Osaka City University)   On Information Geometry and its Applications              July 24, 2012   14 / 16
Applications


     For Question


Proof of Theorem (roughly).
To prove this theorem,

any statistical manifold can be immersed into the statistical manifold
 which the space of positive probability distributions is generalized
                                  to.

We use the following two important immersion theorems.
 • The Nash Immersion Theorem
    • The Gromov Immersion Theorem
                                                                                      .


Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   15 / 16
Applications


     For Question


Proof of Theorem (roughly).
To prove this theorem,

any statistical manifold can be immersed into the statistical manifold
 which the space of positive probability distributions is generalized
                                  to.

We use the following two important immersion theorems.
 • The Nash Immersion Theorem
    • The Gromov Immersion Theorem
                                                                                      .


Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   15 / 16
Applications


     For Question



Theorem 3.1 (THE NASH IMMERSION THEOREM (1954))
Any smooth Riemannian manifold (M m , g) can be isometrically
                    N
immersed into (RN , i=1 dxi2 ) for some N depending on M m
                                          .
Theorem 3.2 (THE GROMOV IMMERSION THEOREM (1986))
Suppose that M m is given with a smooth symmetric 3-form T . Then
there exists an immersion f : M m → RN1 (m) with     .
                 m+1   m+2            ∗     N1 (m)
N1 (m) = 3(m + 2 + 3 ) such that f ( i=1 dxi ) = T 3




Masaki Asano (Osaka City University)   On Information Geometry and its Applications   July 24, 2012   16 / 16

More Related Content

What's hot

Introductory maths analysis chapter 07 official
Introductory maths analysis   chapter 07 officialIntroductory maths analysis   chapter 07 official
Introductory maths analysis chapter 07 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 17 official
Introductory maths analysis   chapter 17 officialIntroductory maths analysis   chapter 17 official
Introductory maths analysis chapter 17 officialEvert Sandye Taasiringan
 
Lecture 9 eigenvalues - 5-1 & 5-2
Lecture 9   eigenvalues -  5-1 & 5-2Lecture 9   eigenvalues -  5-1 & 5-2
Lecture 9 eigenvalues - 5-1 & 5-2njit-ronbrown
 
Introductory maths analysis chapter 15 official
Introductory maths analysis   chapter 15 officialIntroductory maths analysis   chapter 15 official
Introductory maths analysis chapter 15 officialEvert Sandye Taasiringan
 
Lecture 8 nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6
Lecture 8   nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6Lecture 8   nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6
Lecture 8 nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6njit-ronbrown
 
On Fully Indecomposable Quaternion Doubly Stochastic Matrices
On Fully Indecomposable Quaternion Doubly Stochastic MatricesOn Fully Indecomposable Quaternion Doubly Stochastic Matrices
On Fully Indecomposable Quaternion Doubly Stochastic Matricesijtsrd
 
Introductory maths analysis chapter 09 official
Introductory maths analysis   chapter 09 officialIntroductory maths analysis   chapter 09 official
Introductory maths analysis chapter 09 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 08 official
Introductory maths analysis   chapter 08 officialIntroductory maths analysis   chapter 08 official
Introductory maths analysis chapter 08 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 16 official
Introductory maths analysis   chapter 16 officialIntroductory maths analysis   chapter 16 official
Introductory maths analysis chapter 16 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 03 official
Introductory maths analysis   chapter 03 officialIntroductory maths analysis   chapter 03 official
Introductory maths analysis chapter 03 officialEvert Sandye Taasiringan
 
Common fixed point for two weakly compatible pairs ...
Common fixed point for two weakly compatible pairs                           ...Common fixed point for two weakly compatible pairs                           ...
Common fixed point for two weakly compatible pairs ...Alexander Decker
 
Introductory maths analysis chapter 05 official
Introductory maths analysis   chapter 05 officialIntroductory maths analysis   chapter 05 official
Introductory maths analysis chapter 05 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 06 official
Introductory maths analysis   chapter 06 officialIntroductory maths analysis   chapter 06 official
Introductory maths analysis chapter 06 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 02 official
Introductory maths analysis   chapter 02 officialIntroductory maths analysis   chapter 02 official
Introductory maths analysis chapter 02 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 10 official
Introductory maths analysis   chapter 10 officialIntroductory maths analysis   chapter 10 official
Introductory maths analysis chapter 10 officialEvert Sandye Taasiringan
 
Introductory maths analysis chapter 13 official
Introductory maths analysis   chapter 13 officialIntroductory maths analysis   chapter 13 official
Introductory maths analysis chapter 13 officialEvert Sandye Taasiringan
 

What's hot (16)

Introductory maths analysis chapter 07 official
Introductory maths analysis   chapter 07 officialIntroductory maths analysis   chapter 07 official
Introductory maths analysis chapter 07 official
 
Introductory maths analysis chapter 17 official
Introductory maths analysis   chapter 17 officialIntroductory maths analysis   chapter 17 official
Introductory maths analysis chapter 17 official
 
Lecture 9 eigenvalues - 5-1 & 5-2
Lecture 9   eigenvalues -  5-1 & 5-2Lecture 9   eigenvalues -  5-1 & 5-2
Lecture 9 eigenvalues - 5-1 & 5-2
 
Introductory maths analysis chapter 15 official
Introductory maths analysis   chapter 15 officialIntroductory maths analysis   chapter 15 official
Introductory maths analysis chapter 15 official
 
Lecture 8 nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6
Lecture 8   nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6Lecture 8   nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6
Lecture 8 nul col bases dim & rank - section 4-2, 4-3, 4-5 & 4-6
 
On Fully Indecomposable Quaternion Doubly Stochastic Matrices
On Fully Indecomposable Quaternion Doubly Stochastic MatricesOn Fully Indecomposable Quaternion Doubly Stochastic Matrices
On Fully Indecomposable Quaternion Doubly Stochastic Matrices
 
Introductory maths analysis chapter 09 official
Introductory maths analysis   chapter 09 officialIntroductory maths analysis   chapter 09 official
Introductory maths analysis chapter 09 official
 
Introductory maths analysis chapter 08 official
Introductory maths analysis   chapter 08 officialIntroductory maths analysis   chapter 08 official
Introductory maths analysis chapter 08 official
 
Introductory maths analysis chapter 16 official
Introductory maths analysis   chapter 16 officialIntroductory maths analysis   chapter 16 official
Introductory maths analysis chapter 16 official
 
Introductory maths analysis chapter 03 official
Introductory maths analysis   chapter 03 officialIntroductory maths analysis   chapter 03 official
Introductory maths analysis chapter 03 official
 
Common fixed point for two weakly compatible pairs ...
Common fixed point for two weakly compatible pairs                           ...Common fixed point for two weakly compatible pairs                           ...
Common fixed point for two weakly compatible pairs ...
 
Introductory maths analysis chapter 05 official
Introductory maths analysis   chapter 05 officialIntroductory maths analysis   chapter 05 official
Introductory maths analysis chapter 05 official
 
Introductory maths analysis chapter 06 official
Introductory maths analysis   chapter 06 officialIntroductory maths analysis   chapter 06 official
Introductory maths analysis chapter 06 official
 
Introductory maths analysis chapter 02 official
Introductory maths analysis   chapter 02 officialIntroductory maths analysis   chapter 02 official
Introductory maths analysis chapter 02 official
 
Introductory maths analysis chapter 10 official
Introductory maths analysis   chapter 10 officialIntroductory maths analysis   chapter 10 official
Introductory maths analysis chapter 10 official
 
Introductory maths analysis chapter 13 official
Introductory maths analysis   chapter 13 officialIntroductory maths analysis   chapter 13 official
Introductory maths analysis chapter 13 official
 

Similar to 韓国での研究集会で用いた発表スライド

COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
bau-22-1.pdf
bau-22-1.pdfbau-22-1.pdf
bau-22-1.pdfShengBau
 
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...mathsjournal
 
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...Joe Suzuki
 
Stochastic Differentiation
Stochastic DifferentiationStochastic Differentiation
Stochastic DifferentiationSSA KPI
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingFrank Nielsen
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Christian Robert
 
Structured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelStructured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelLaboratoire Statistique et génome
 
The Universal Measure for General Sources and its Application to MDL/Bayesian...
The Universal Measure for General Sources and its Application to MDL/Bayesian...The Universal Measure for General Sources and its Application to MDL/Bayesian...
The Universal Measure for General Sources and its Application to MDL/Bayesian...Joe Suzuki
 
Prof. Rob Leigh (University of Illinois)
Prof. Rob Leigh (University of Illinois)Prof. Rob Leigh (University of Illinois)
Prof. Rob Leigh (University of Illinois)Rene Kotze
 
Introduction to Stochastic calculus
Introduction to Stochastic calculusIntroduction to Stochastic calculus
Introduction to Stochastic calculusAshwin Rao
 

Similar to 韓国での研究集会で用いた発表スライド (20)

Proba stats-r1-2017
Proba stats-r1-2017Proba stats-r1-2017
Proba stats-r1-2017
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
bau-22-1.pdf
bau-22-1.pdfbau-22-1.pdf
bau-22-1.pdf
 
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...
Common Fixed Point Theorems in Compatible Mappings of Type (P*) of Generalize...
 
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...
A Generalization of Nonparametric Estimation and On-Line Prediction for Stati...
 
Side 2019 #9
Side 2019 #9Side 2019 #9
Side 2019 #9
 
Side 2019, part 1
Side 2019, part 1Side 2019, part 1
Side 2019, part 1
 
Stochastic Differentiation
Stochastic DifferentiationStochastic Differentiation
Stochastic Differentiation
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processing
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
 
Structured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelStructured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical model
 
Lad
LadLad
Lad
 
The Universal Measure for General Sources and its Application to MDL/Bayesian...
The Universal Measure for General Sources and its Application to MDL/Bayesian...The Universal Measure for General Sources and its Application to MDL/Bayesian...
The Universal Measure for General Sources and its Application to MDL/Bayesian...
 
Prof. Rob Leigh (University of Illinois)
Prof. Rob Leigh (University of Illinois)Prof. Rob Leigh (University of Illinois)
Prof. Rob Leigh (University of Illinois)
 
Presentation
PresentationPresentation
Presentation
 
Introduction to Stochastic calculus
Introduction to Stochastic calculusIntroduction to Stochastic calculus
Introduction to Stochastic calculus
 
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
 

韓国での研究集会で用いた発表スライド

  • 1. On Information Geometry and its Applications . . Masaki Asano (M1) Osaka City University My advisor is Prof. Ohnita July 24, 2012 Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 1 / 16
  • 2. Contents of this talk 1 Statistical model, Fisher metric and α-connection 2 Statistical manifold 3 Applications Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 2 / 16
  • 3. Statistical model, Fisher metric and α-connection Statistical model (Ω, F , P) : a probability space Ξ : an open domain of Rn (a parameter space) Definition 1.1 S is a statistical model or parametric model on Ω def ⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such that S = p(x; ξ) p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn , Ω where P(A) = A p(x; ξ)dx, (A ∈ F ). We assume S is a smooth manifold with local coordinate system Ξ. . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 3 / 16
  • 4. Statistical model, Fisher metric and α-connection Statistical model (Ω, F , P) : a probability space Ξ : an open domain of Rn (a parameter space) Definition 1.1 S is a statistical model or parametric model on Ω def ⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such that S = p(x; ξ) p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn , Ω where P(A) = A p(x; ξ)dx, (A ∈ F ). We assume S is a smooth manifold with local coordinate system Ξ. . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 3 / 16
  • 5. Statistical model, Fisher metric and α-connection Statistical model (Ω, F , P) : a probability space Ξ : an open domain of Rn (a parameter space) Definition 1.1 S is a statistical model or parametric model on Ω def ⇐⇒ S is a set of probability densities with parameter ξ ∈ Ξ such that S = p(x; ξ) p(x; ξ)dx = 1, p(x; ξ) > 0, ξ ∈ Ξ ⊂ Rn , Ω where P(A) = A p(x; ξ)dx, (A ∈ F ). We assume S is a smooth manifold with local coordinate system Ξ. . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 3 / 16
  • 6. Statistical model, Fisher metric and α-connection Fisher metric For simplicity, Eξ [ f ] = f (x)p(x; ξ)dx, (the expectation of f (x) w.r.t. p(x; ξ)) Ω lξ = l(x; ξ) = log p(x; ξ) (the information of p(x; ξ)) ∂ ∂i = i ∂ξ Definition 1.2 (Fisher information matrix) g = (gi j ) is the Fisher information matrix of S . def ⇐⇒ gi j (ξ) := Eξ ∂i lξ ∂ j lξ Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 4 / 16
  • 7. Statistical model, Fisher metric and α-connection Fisher metric For simplicity, Eξ [ f ] = f (x)p(x; ξ)dx, (the expectation of f (x) w.r.t. p(x; ξ)) Ω lξ = l(x; ξ) = log p(x; ξ) (the information of p(x; ξ)) ∂ ∂i = i ∂ξ Definition 1.2 (Fisher information matrix) g = (gi j ) is the Fisher information matrix of S . def ⇐⇒ gi j (ξ) := Eξ ∂i lξ ∂ j lξ Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 4 / 16
  • 8. Statistical model, Fisher metric and α-connection Proposition 1.3 The following conditions are equivalent. • g is positive definite. • {∂1 pξ , · · · , ∂n pξ } are linearly independent. . • {∂1 lξ , · · · , ∂n lξ } are linearly independent. We assume that one of the above conditions is satisfied and gi j (ξ) is finite for all i, j, ξ . =⇒ We can define a Riemannian metric on S . . =⇒ The metric g is called Fisher metric. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 5 / 16
  • 9. Statistical model, Fisher metric and α-connection Proposition 1.3 The following conditions are equivalent. • g is positive definite. • {∂1 pξ , · · · , ∂n pξ } are linearly independent. . • {∂1 lξ , · · · , ∂n lξ } are linearly independent. We assume that one of the above conditions is satisfied and gi j (ξ) is finite for all i, j, ξ . =⇒ We can define a Riemannian metric on S . . =⇒ The metric g is called Fisher metric. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 5 / 16
  • 10. Statistical model, Fisher metric and α-connection Proposition 1.3 The following conditions are equivalent. • g is positive definite. • {∂1 pξ , · · · , ∂n pξ } are linearly independent. . • {∂1 lξ , · · · , ∂n lξ } are linearly independent. We assume that one of the above conditions is satisfied and gi j (ξ) is finite for all i, j, ξ . =⇒ We can define a Riemannian metric on S . . =⇒ The metric g is called Fisher metric. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 5 / 16
  • 11. Statistical model, Fisher metric and α-connection Proposition 1.3 The following conditions are equivalent. • g is positive definite. • {∂1 pξ , · · · , ∂n pξ } are linearly independent. . • {∂1 lξ , · · · , ∂n lξ } are linearly independent. We assume that one of the above conditions is satisfied and gi j (ξ) is finite for all i, j, ξ . =⇒ We can define a Riemannian metric on S . . =⇒ The metric g is called Fisher metric. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 5 / 16
  • 12. Statistical model, Fisher metric and α-connection α-connection Definition 1.4 For α ∈ R, we define the α-connection (α) by the following formula, (α) 1−α g ∂i ∂ j , ∂k = E ∂i ∂ j l ξ + ∂i lξ ∂ j lξ ∂k lξ . 2 . By the definition of α-connection, (0) (1) is the Levi-Civita connection of the Fisher metric g. (2) (α) is torsion-free (∀ α), (α) (α) (α) i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0. . (α) (α) (3) ( X g)(Y, Z) =( Y g)(X, Z). Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 6 / 16
  • 13. Statistical model, Fisher metric and α-connection α-connection Definition 1.4 For α ∈ R, we define the α-connection (α) by the following formula, (α) 1−α g ∂i ∂ j , ∂k = E ∂i ∂ j l ξ + ∂i lξ ∂ j lξ ∂k lξ . 2 . By the definition of α-connection, (0) (1) is the Levi-Civita connection of the Fisher metric g. (2) (α) is torsion-free (∀ α), (α) (α) (α) i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0. . (α) (α) (3) ( X g)(Y, Z) =( Y g)(X, Z). Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 6 / 16
  • 14. Statistical model, Fisher metric and α-connection Example 1.5 (Normal distribution) ξ = (µ, σ) ∈ Ξ = R × R+ (Upper half-plane) µ : mean (−∞ < µ < ∞), σ : standard deviation (0 < σ < ∞), 1 (x − µ)2 S = p(x; ξ) p(x; µ, σ) = √ exp − 2πσ 2σ2 . 1 1 0 1 . The Fisher metric : g = 2 . The curvature of S : − . σ 0 2 2 Hyperbolic plane. (x, y) ∈ H = {(x, y) ∈ R2 | y > 0} 1 1 0 The Poincare metric : g = 2 . The curvature of H : −1. y 0 1 . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 7 / 16
  • 15. Statistical model, Fisher metric and α-connection Example 1.5 (Normal distribution) ξ = (µ, σ) ∈ Ξ = R × R+ (Upper half-plane) µ : mean (−∞ < µ < ∞), σ : standard deviation (0 < σ < ∞), 1 (x − µ)2 S = p(x; ξ) p(x; µ, σ) = √ exp − 2πσ 2σ2 . 1 1 0 1 . The Fisher metric : g = 2 . The curvature of S : − . σ 0 2 2 Hyperbolic plane. (x, y) ∈ H = {(x, y) ∈ R2 | y > 0} 1 1 0 The Poincare metric : g = 2 . The curvature of H : −1. y 0 1 . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 7 / 16
  • 16. Statistical manifold Statistical manifold (M, g) : a Riemannian manifold : a torsion-free affine connection on M i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0 Definition 2.1 (M, , g) is a statistical manifold def ⇐⇒ g is totally symmetric (0, 3)-tensor field. We can find that connection satisfies the properties (1)-(3) of α-connection (α) . . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 8 / 16
  • 17. Statistical manifold Statistical manifold (M, g) : a Riemannian manifold : a torsion-free affine connection on M i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0 Definition 2.1 (M, , g) is a statistical manifold def ⇐⇒ g is totally symmetric (0, 3)-tensor field. We can find that connection satisfies the properties (1)-(3) of α-connection (α) . . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 8 / 16
  • 18. Statistical manifold Statistical manifold (M, g) : a Riemannian manifold : a torsion-free affine connection on M i.e. T (X, Y) := X Y − Y X − [X, Y] ≡ 0 Definition 2.1 (M, , g) is a statistical manifold def ⇐⇒ g is totally symmetric (0, 3)-tensor field. We can find that connection satisfies the properties (1)-(3) of α-connection (α) . . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 8 / 16
  • 19. Applications Application 1 Statistical Models Theorem Theorem (Hon Van Le(2005))⇑⇓(Hon Van Le(2005)) ˆ ˆ ˆ ˆ ˆ ˆ Statistical Manifolds Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 9 / 16
  • 20. Applications Application 1 Statistical Models Theorem Theorem (Hon Van Le(2005))⇑⇓(Hon Van Le(2005)) ˆ ˆ ˆ ˆ ˆ ˆ Statistical Manifolds Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 10 / 16
  • 21. Applications Application 2 Information geometry is related to ... • statistics, • information theory, • (almost) complex geometry, • symplectic geometry, • contact geometry, • Wasserstein geometry...etc. However, I am not sure. These are that I want to study in future. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 11 / 16
  • 22. Applications Application 2 Information geometry is related to ... • statistics, • information theory, • (almost) complex geometry, • symplectic geometry, • contact geometry, • Wasserstein geometry...etc. However, I am not sure. These are that I want to study in future. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 11 / 16
  • 23. Applications Application 2 Information geometry is related to ... • statistics, • information theory, • (almost) complex geometry, • symplectic geometry, • contact geometry, • Wasserstein geometry...etc. However, I am not sure. These are that I want to study in future. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 11 / 16
  • 24. Applications Thank you for your attention!! Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 12 / 16
  • 25. Applications References • S. Amari and H. Nagaoka, Methods of Information Geometry, Trans. of Math. Monograph, AMS, 2000. • M. Gromov, Partial differential relations. Springer-Verlag, Berlin, 1986. • H-V. Le, Statistical manifolds are statistical models. J. Geom. ˆ 84 (2005), no. 1-2, 83-93. • J. Nash, C 1 -isometric imbeddings. Ann. of Math. 60, (1954). 383-396. Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 13 / 16
  • 26. Applications For Question Proof of Proposition 1.3. For any n-dimensional vector c = t (c1 , c2 , . . . , cn ) (t denotes transpose), 2 t cgc = ci c j gi j (ξ) = Eξ ci ∂i lξ i, j . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 14 / 16
  • 27. Applications For Question Proof of Theorem (roughly). To prove this theorem, any statistical manifold can be immersed into the statistical manifold which the space of positive probability distributions is generalized to. We use the following two important immersion theorems. • The Nash Immersion Theorem • The Gromov Immersion Theorem . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 15 / 16
  • 28. Applications For Question Proof of Theorem (roughly). To prove this theorem, any statistical manifold can be immersed into the statistical manifold which the space of positive probability distributions is generalized to. We use the following two important immersion theorems. • The Nash Immersion Theorem • The Gromov Immersion Theorem . Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 15 / 16
  • 29. Applications For Question Theorem 3.1 (THE NASH IMMERSION THEOREM (1954)) Any smooth Riemannian manifold (M m , g) can be isometrically N immersed into (RN , i=1 dxi2 ) for some N depending on M m . Theorem 3.2 (THE GROMOV IMMERSION THEOREM (1986)) Suppose that M m is given with a smooth symmetric 3-form T . Then there exists an immersion f : M m → RN1 (m) with . m+1 m+2 ∗ N1 (m) N1 (m) = 3(m + 2 + 3 ) such that f ( i=1 dxi ) = T 3 Masaki Asano (Osaka City University) On Information Geometry and its Applications July 24, 2012 16 / 16