SlideShare a Scribd company logo
1 of 165
Download to read offline
CSR-2011




   Kolmogorov complexity as a language
                   Alexander Shen
           LIF CNRS, Marseille; on leave from
                 ИППИ РАН, Москва




                                      .   .     .   .   .   .
Kolmogorov complexity




                        .   .   .   .   .   .
Kolmogorov complexity
     A powerful tool




                        .   .   .   .   .   .
Kolmogorov complexity
     A powerful tool
     Just a way to reformulate arguments




                                           .   .   .   .   .   .
Kolmogorov complexity
     A powerful tool
     Just a way to reformulate arguments
     three languages: combinatorial/algorithmic/probabilistic




                                           .    .   .    .      .   .
Kolmogorov complexity
     A powerful tool
     Just a way to reformulate arguments
     three languages: combinatorial/algorithmic/probabilistic




                                           .    .   .    .      .   .
Reminder and notation




                        .   .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x




                                           .   .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}




                                           .   .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}
      depends on the interpreter D




                                           .   .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}
      depends on the interpreter D
      optimal D makes it minimal up to O(1) additive term




                                           .    .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}
      depends on the interpreter D
      optimal D makes it minimal up to O(1) additive term
      Variations:




                                           .    .   .   .   .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}
      depends on the interpreter D
      optimal D makes it minimal up to O(1) additive term
      Variations:

                      p = string     p = prefix of a sequence
       x = string       plain                  prefix
                      K(x), C(x)           KP(x), K(x)
        x = prefix     decision             monotone
     of a sequence   KR(x), KD(x)         KM(x), Km(x)




                                           .    .   .   .       .   .
Reminder and notation
      K(x) = minimal length of a program that produces x
      KD (x) = min{|p| : D(p) = x}
      depends on the interpreter D
      optimal D makes it minimal up to O(1) additive term
      Variations:

                      p = string     p = prefix of a sequence
       x = string       plain                  prefix
                      K(x), C(x)           KP(x), K(x)
        x = prefix     decision             monotone
     of a sequence   KR(x), KD(x)         KM(x), Km(x)

  Conditional complexity C(x|y):
  minimal length of a program p : y → x.


                                           .    .   .   .       .   .
Reminder and notation
       K(x) = minimal length of a program that produces x
       KD (x) = min{|p| : D(p) = x}
       depends on the interpreter D
       optimal D makes it minimal up to O(1) additive term
       Variations:

                          p = string        p = prefix of a sequence
        x = string          plain                     prefix
                          K(x), C(x)              KP(x), K(x)
        x = prefix         decision                monotone
     of a sequence       KR(x), KD(x)            KM(x), Km(x)

  Conditional complexity C(x|y):
  minimal length of a program p : y → x.
  There is also a priori probability (in two versions: discrete, on strings;
  continuous, on prefixes)
                                                    .    .    .    .    .      .
Foundations of probability theory




                                    .   .   .   .   .   .
Foundations of probability theory

      Random object or random process?




                                         .   .   .   .   .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?




                                           .    .   .   .   .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                               [xkcd cartoon]




                                           .    .   .   .       .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                               [xkcd cartoon]



      randomness = incompressibility (maximal complexity)




                                           .    .   .   .       .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                                    [xkcd cartoon]



      randomness = incompressibility (maximal complexity)
      ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1)




                                                .    .    .       .   .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                                    [xkcd cartoon]



      randomness = incompressibility (maximal complexity)
      ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1)
      Classical probability theory: random sequence satisfies the
      Strong Law of Large Numbers with probability 1




                                                .    .    .       .   .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                                    [xkcd cartoon]



      randomness = incompressibility (maximal complexity)
      ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1)
      Classical probability theory: random sequence satisfies the
      Strong Law of Large Numbers with probability 1
      Algorithmic version: every (algorithmically) random
      sequence satisfies SLLN



                                                .    .    .       .   .   .
Foundations of probability theory

      Random object or random process?
      “well shuffled deck of cards”: any meaning?

                                                    [xkcd cartoon]



      randomness = incompressibility (maximal complexity)
      ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1)
      Classical probability theory: random sequence satisfies the
      Strong Law of Large Numbers with probability 1
      Algorithmic version: every (algorithmically) random
      sequence satisfies SLLN
      algorithmic ⇒ classical: Martin-Löf random sequences
      form a set of measure 1.
                                                .    .    .       .   .   .
Sampling random strings (S.Aaronson)




                                .      .   .   .   .   .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops




                                            .   .    .   .   .      .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops
      “The device produces a random string”: what does it
      mean?




                                            .   .    .   .   .      .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops
      “The device produces a random string”: what does it
      mean?
      classical: the output distribution is close to the uniform one




                                              .    .   .    .   .      .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops
      “The device produces a random string”: what does it
      mean?
      classical: the output distribution is close to the uniform one
      effective: with high probability the output string is
      incompressible




                                               .    .    .    .   .    .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops
      “The device produces a random string”: what does it
      mean?
      classical: the output distribution is close to the uniform one
      effective: with high probability the output string is
      incompressible
      not equivalent if no assumptions about the device




                                               .    .    .    .   .    .
Sampling random strings (S.Aaronson)



      A device that (being switched on) produces N-bit string and
      stops
      “The device produces a random string”: what does it
      mean?
      classical: the output distribution is close to the uniform one
      effective: with high probability the output string is
      incompressible
      not equivalent if no assumptions about the device
      but are related under some assumptions




                                               .    .    .    .   .    .
Example: matrices without uniform minors




                                 .   .     .   .   .   .
Example: matrices without uniform minors

      k × k minor of n × n Boolean matrix: select k rows and k
      columns




                                            .   .   .    .   .   .
Example: matrices without uniform minors

      k × k minor of n × n Boolean matrix: select k rows and k
      columns




      minor is uniform if it is all-0 or all-1.
      claim: there is a n × n bit matrix without k × k uniform
      minors for k = 3 log n.


                                                  .   .   .   .   .   .
Counting argument and complexity reformulation




                                 .   .   .   .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]




                                              .   .    .   .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)




                                              .   .    .   .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest




                                              .   .    .   .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2
      n2k   × 2 × 2n           =




                                              .   .    .   .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2                         2 −9 log2             2
      n2k   × 2 × 2n           = 2log n×2×3 log n+1+(n           n)   < 2n .




                                                         .   .         .       .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2                         2 −9 log2             2
      n2k   × 2 × 2n           = 2log n×2×3 log n+1+(n           n)   < 2n .

      log n bits to specify a column or row: 2k log n bits in total




                                                         .   .         .       .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2                         2 −9 log2             2
      n2k   × 2 × 2n           = 2log n×2×3 log n+1+(n           n)   < 2n .

      log n bits to specify a column or row: 2k log n bits in total
      one additional bit to specify the type of minor (0/1)




                                                         .   .         .       .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2                         2 −9 log2             2
      n2k   × 2 × 2n           = 2log n×2×3 log n+1+(n           n)   < 2n .

      log n bits to specify a column or row: 2k log n bits in total
      one additional bit to specify the type of minor (0/1)
      n2 − k2 bits to specify the rest of the matrix




                                                         .   .         .       .   .   .
Counting argument and complexity reformulation


      ≤ nk × nk positions of the minor [k = 3 log n]
      2 types of uniform minors (0/1)
        2 −k2
      2n        possibilities for the rest
                       2 −k2                         2 −9 log2             2
      n2k   × 2 × 2n           = 2log n×2×3 log n+1+(n           n)   < 2n .

      log n bits to specify a column or row: 2k log n bits in total
      one additional bit to specify the type of minor (0/1)
      n2 − k2 bits to specify the rest of the matrix
      2k log n + 1 + (n2 − k2 ) = 6 log2 n + 1 + (n2 − 9 log2 n) < n2 .




                                                         .   .         .       .   .   .
One-tape Turing machines




                           .   .   .   .   .   .
One-tape Turing machines
     copying n-bit string on 1-tape TM requires Ω(n2 ) time




                                            .   .   .    .    .   .
One-tape Turing machines
     copying n-bit string on 1-tape TM requires Ω(n2 ) time




     complexity version: if initially the tape was empty on the
     right of the border, then after n steps the complexity of a
     zone that is d cells far from the border is O(n/d).




     K(u(t)) ≤ O(n/d)



                                             .   .    .    .   .   .
One-tape Turing machines
     copying n-bit string on 1-tape TM requires Ω(n2 ) time




     complexity version: if initially the tape was empty on the
     right of the border, then after n steps the complexity of a
     zone that is d cells far from the border is O(n/d).




     K(u(t)) ≤ O(n/d)
     proof: border guards in each cell of the border security zone write down the
     contents of the head of TM; each of the records is enough to reconstruct u(t) so
     the length of it should be Ω(K(u(t)); the sum of lengths does not exceed time
                                                         .     .      .     .       .   .
Everywhere complex sequences




                               .   .   .   .   .   .
Everywhere complex sequences

     Random sequence has n-bit prefix of complexity n




                                         .   .   .      .   .   .
Everywhere complex sequences

     Random sequence has n-bit prefix of complexity n
     but some factors (substrings) have small complexity




                                          .   .    .    .   .   .
Everywhere complex sequences

     Random sequence has n-bit prefix of complexity n
     but some factors (substrings) have small complexity
     Levin: there exist everywhere complex sequences: every
     n-bit substring has complexity 0.99n − O(1)




                                          .   .    .    .   .   .
Everywhere complex sequences

     Random sequence has n-bit prefix of complexity n
     but some factors (substrings) have small complexity
     Levin: there exist everywhere complex sequences: every
     n-bit substring has complexity 0.99n − O(1)




     Combinatorial equivalent: Let F be a set of strings that has
     at most 20.99n strings of length n. Then there is a sequence
     ω s.t. all sufficiently long substrings of ω are not in F.




                                            .   .   .    .   .      .
Everywhere complex sequences

     Random sequence has n-bit prefix of complexity n
     but some factors (substrings) have small complexity
     Levin: there exist everywhere complex sequences: every
     n-bit substring has complexity 0.99n − O(1)




     Combinatorial equivalent: Let F be a set of strings that has
     at most 20.99n strings of length n. Then there is a sequence
     ω s.t. all sufficiently long substrings of ω are not in F.
     combinatorial and complexity proofs not just translations of
     each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)


                                            .   .    .   .   .      .
Gilbert-Varshamov complexity bound




                                .    .   .   .   .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




                                               .    .   .    .    .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




      lower bound (Gilbert–Varshamov)




                                               .    .   .    .    .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




      lower bound (Gilbert–Varshamov)
      then < d/2 changed bits are harmless




                                               .    .   .    .    .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




      lower bound (Gilbert–Varshamov)
      then < d/2 changed bits are harmless
      but bit insertion or deletions could be




                                                .   .   .    .    .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




      lower bound (Gilbert–Varshamov)
      then < d/2 changed bits are harmless
      but bit insertion or deletions could be
      general requirement: C(xi |xj ) ≥ d




                                                .   .   .    .    .   .
Gilbert-Varshamov complexity bound

      coding theory: how many n-bit strings x1 , . . . , xk one can
      find if Hamming distance between every two is at least d




      lower bound (Gilbert–Varshamov)
      then < d/2 changed bits are harmless
      but bit insertion or deletions could be
      general requirement: C(xi |xj ) ≥ d
      generalization of GV bound: d-separated family of size
      Ω(2n−d )

                                                .   .   .    .    .   .
Inequalities for complexities and combinatorial
interpretation




                                    .   .   .   .   .   .
Inequalities for complexities and combinatorial
interpretation

      C(x, y) ≤ C(x) + C(y|x) + O(log)




                                         .   .   .   .   .   .
Inequalities for complexities and combinatorial
interpretation

      C(x, y) ≤ C(x) + C(y|x) + O(log)




                                         .   .   .   .   .   .
Inequalities for complexities and combinatorial
interpretation

      C(x, y) ≤ C(x) + C(y|x) + O(log)




      C(x, y) ≥ C(x) + C(y|x) + O(log)




                                         .   .   .   .   .   .
Inequalities for complexities and combinatorial
interpretation

      C(x, y) ≤ C(x) + C(y|x) + O(log)




      C(x, y) ≥ C(x) + C(y|x) + O(log)
      C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log)




                                            .   .    .   .   .     .
Inequalities for complexities and combinatorial
interpretation

      C(x, y) ≤ C(x) + C(y|x) + O(log)




      C(x, y) ≥ C(x) + C(y|x) + O(log)
      C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log)
      every set A of size < 2k+l can be split into two parts
      A = A1 ∪ A2 such that w(A1 ) ≤ 2k and h(A2 ) ≤ 2l


                                              .   .    .   .   .   .
One more inequality




                      .   .   .   .   .   .
One more inequality



      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z)




                                              .   .   .   .   .   .
One more inequality



      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z)




      V2 ≤ S1 × S2 × S3




                                              .   .   .   .   .   .
One more inequality



      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z)




      V2 ≤ S1 × S2 × S3
      Also for Shannon entropies; special case of Shearer
      lemma




                                              .   .   .   .   .   .
Common information and graph minors




                               .   .   .   .   .   .
Common information and graph minors

     mutual information: I(a : b) = C(a) + C(b) − C(a, b)




                                           .   .    .   .   .   .
Common information and graph minors

     mutual information: I(a : b) = C(a) + C(b) − C(a, b)




     common information:




                                           .   .    .   .   .   .
Common information and graph minors

     mutual information: I(a : b) = C(a) + C(b) − C(a, b)




     common information:




     combinatorial: graph minors



     can the graph be covered by 2δ minors of size 2α−δ × 2β−δ ?

                                           .   .    .   .   .      .
Almost uniform sets




                      .   .   .   .   .   .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




                                           .   .   .       .   .   .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




      Theorem: every set of N elements can be represented as
      union of polylog(N) sets whose nonuniformity is polylog(N).




                                            .   .    .     .   .    .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




      Theorem: every set of N elements can be represented as
      union of polylog(N) sets whose nonuniformity is polylog(N).
      multidimensional version




                                            .   .    .     .   .    .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




      Theorem: every set of N elements can be represented as
      union of polylog(N) sets whose nonuniformity is polylog(N).
      multidimensional version
      how to construct parts using Kolmogorov complexity: take
      strings with given complexity bounds




                                            .   .    .     .   .    .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




      Theorem: every set of N elements can be represented as
      union of polylog(N) sets whose nonuniformity is polylog(N).
      multidimensional version
      how to construct parts using Kolmogorov complexity: take
      strings with given complexity bounds
      so simple that it is not clear what is the combinatorial
      translation



                                              .    .   .    .    .   .
Almost uniform sets
      nonuniformity= (maximal section)/(average section)




      Theorem: every set of N elements can be represented as
      union of polylog(N) sets whose nonuniformity is polylog(N).
      multidimensional version
      how to construct parts using Kolmogorov complexity: take
      strings with given complexity bounds
      so simple that it is not clear what is the combinatorial
      translation
      but combinatorial argument exists (and gives even a
      stronger result)
                                              .    .   .    .    .   .
Shannon coding theorem




                         .   .   .   .   .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk




                                                .    .    .    .       .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk
     ξ N : N independent trials of ξ




                                                .    .    .    .       .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk
     ξ N : N independent trials of ξ
     Shannon’s informal question: how many bits are needed to
     encode a “typical” value of ξ N ?




                                                .    .    .    .       .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk
     ξ N : N independent trials of ξ
     Shannon’s informal question: how many bits are needed to
     encode a “typical” value of ξ N ?
     Shannon’s answer: NH(ξ), where

               H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ).




                                                .    .    .    .       .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk
     ξ N : N independent trials of ξ
     Shannon’s informal question: how many bits are needed to
     encode a “typical” value of ξ N ?
     Shannon’s answer: NH(ξ), where

               H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ).

     formal statement is a bit complicated




                                                .    .    .    .       .   .
Shannon coding theorem


     ξ is a random variable; k values, probabilities p1 , . . . , pk
     ξ N : N independent trials of ξ
     Shannon’s informal question: how many bits are needed to
     encode a “typical” value of ξ N ?
     Shannon’s answer: NH(ξ), where

               H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ).

     formal statement is a bit complicated
     Complexity version: with high probablity the value of ξ N
     has complexity close to NH(ξ).




                                                .    .    .    .       .   .
Complexity, entropy and group size




                                     .   .   .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)




                                              .   .    .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)
      The same for entropy:
      2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ )




                                                 .   .   .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)
      The same for entropy:
      2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ )
      …and even for the sizes of subgroups U, V, W of some
      finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤
      log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).




                                                 .   .   .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)
      The same for entropy:
      2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ )
      …and even for the sizes of subgroups U, V, W of some
      finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤
      log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).
      in all three cases inequalities are the same
      (Romashchenko, Chan, Yeung)




                                                 .   .   .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)
      The same for entropy:
      2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ )
      …and even for the sizes of subgroups U, V, W of some
      finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤
      log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).
      in all three cases inequalities are the same
      (Romashchenko, Chan, Yeung)
      some of them are quite strange:

        I(a : b) ≤
      ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a)




                                                 .   .   .   .   .   .
Complexity, entropy and group size
      2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)
      The same for entropy:
      2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ )
      …and even for the sizes of subgroups U, V, W of some
      finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤
      log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).
      in all three cases inequalities are the same
      (Romashchenko, Chan, Yeung)
      some of them are quite strange:

        I(a : b) ≤
      ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a)

      Related to Romashchenko’s theorem: if three last terms
      are zeros, one can extract common information from a, b, e.
                                                 .   .   .   .   .   .
Muchnik and Slepian–Wolf




                           .   .   .   .   .   .
Muchnik and Slepian–Wolf

     a, b: two strings




                           .   .   .   .   .   .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b




                                          .     .   .   .   .   .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher




                                             .   .    .   .      .   .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher
     there exist p : a → b that is simple relative to b, e.g., “map
     everything to b”




                                              .   .    .    .    .    .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher
     there exist p : a → b that is simple relative to b, e.g., “map
     everything to b”
     Muchnik theorem: it is possible to combine these two
     conditions: there exists p : a → b such that C(p) ≈ C(b|a)
     and C(p|b) ≈ 0




                                              .   .    .    .    .    .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher
     there exist p : a → b that is simple relative to b, e.g., “map
     everything to b”
     Muchnik theorem: it is possible to combine these two
     conditions: there exists p : a → b such that C(p) ≈ C(b|a)
     and C(p|b) ≈ 0
     information theory analog: Wolf–Slepian




                                              .   .    .    .    .    .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher
     there exist p : a → b that is simple relative to b, e.g., “map
     everything to b”
     Muchnik theorem: it is possible to combine these two
     conditions: there exists p : a → b such that C(p) ≈ C(b|a)
     and C(p|b) ≈ 0
     information theory analog: Wolf–Slepian
     similar technique was developed by Fortnow and Laplante
     (randomness extractors)




                                              .   .    .    .    .    .
Muchnik and Slepian–Wolf

     a, b: two strings
     we look for a program p that maps a to b
     by definition C(p) is at least C(b|a) but could be higher
     there exist p : a → b that is simple relative to b, e.g., “map
     everything to b”
     Muchnik theorem: it is possible to combine these two
     conditions: there exists p : a → b such that C(p) ≈ C(b|a)
     and C(p|b) ≈ 0
     information theory analog: Wolf–Slepian
     similar technique was developed by Fortnow and Laplante
     (randomness extractors)
     (Romashchenko, Musatov): how to use explicit extractors
     and derandomization to get space-bounded versions

                                              .   .    .    .    .    .
Computability theory: simple sets




                                    .   .   .   .   .   .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement




                                           .    .   .   .   .    .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.




                                           .    .   .   .   .     .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.
      Most strings are not simple ⇒ infinite complement




                                           .    .   .     .   .   .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.
      Most strings are not simple ⇒ infinite complement
      Let x1 , x2 , . . . be a computable sequence of different
      non-simple strings




                                                .    .   .    .   .   .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.
      Most strings are not simple ⇒ infinite complement
      Let x1 , x2 , . . . be a computable sequence of different
      non-simple strings
      May assume wlog that |xi | > i and therefore C(xi ) > i/2




                                                .    .   .    .   .   .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.
      Most strings are not simple ⇒ infinite complement
      Let x1 , x2 , . . . be a computable sequence of different
      non-simple strings
      May assume wlog that |xi | > i and therefore C(xi ) > i/2
      but to specify xi we need O(log i) bits only




                                                .    .   .    .   .   .
Computability theory: simple sets

      Simple set: enumerable set with infinite complement, but
      no algorithm can generate infinitely many elements from
      the complement
      Construction using Kolmogorov complexity: a simple string
      x has C(x) ≤ |x|/2.
      Most strings are not simple ⇒ infinite complement
      Let x1 , x2 , . . . be a computable sequence of different
      non-simple strings
      May assume wlog that |xi | > i and therefore C(xi ) > i/2
      but to specify xi we need O(log i) bits only
      “Minimal integer that cannot be described in ten English
      words” (Berry)


                                                .    .   .    .   .   .
Lower semicomputable random reals




                               .    .   .   .   .   .
Lower semicomputable random reals
     ∑
         ai computable converging series with rational terms




                                            .   .   .    .     .   .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?




                                           .   .   .    .     .   .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)




                                           .   .   .    .     .   .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals




                                           .   .   .    .     .   .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α




                                           .    .   .    .    .     .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α
     There are maximal elements




                                           .    .   .    .    .     .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α
     There are maximal elements = random lower
     semicomputable reals




                                           .    .   .    .    .     .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α
     There are maximal elements = random lower
     semicomputable reals
     = slowly converging series




                                           .    .   .    .    .     .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α
     There are maximal elements = random lower
     semicomputable reals
     = slowly converging series
     modulus of convergence: ε → N(ε) = how many terms are
     needed for ε-precision




                                           .    .   .    .    .     .
Lower semicomputable random reals
     ∑
        ai computable converging series with rational terms
            ∑
     is α =   ai computable (ε → ε-approximation)?
     not necessarily (Specker example)
     lower semicomputable reals
     Solovay classification: α ≼ β if ε-approximation to β can be
     effectively converted to O(ε)-approximation to α
     There are maximal elements = random lower
     semicomputable reals
     = slowly converging series
     modulus of convergence: ε → N(ε) = how many terms are
     needed for ε-precision
     maximal elements: N(2−n ) > BP(n − O(1)), where BP(k)
     is the maximal integer whose prefix complexity is k or less.

                                            .   .    .   .    .     .
Lovasz local lemma: constructive proof




                                  .   .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors
      ⇒ CNF is satisfiable




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors
      ⇒ CNF is satisfiable
      Non-constructive proof: lower bound for probability, Lovasz
      local lemma




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors
      ⇒ CNF is satisfiable
      Non-constructive proof: lower bound for probability, Lovasz
      local lemma
      Naïve algorithm: just resample false clause (while they
      exist)




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors
      ⇒ CNF is satisfiable
      Non-constructive proof: lower bound for probability, Lovasz
      local lemma
      Naïve algorithm: just resample false clause (while they
      exist)
      Recent breakthrough (Moser): this algorithm with high
      probability terminates quickly




                                                   .     .   .   .   .   .
Lovasz local lemma: constructive proof
      CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .
      neighbors: clauses having common variables
      several clauses with k literals in each
      each clause has o(2k ) neighbors
      ⇒ CNF is satisfiable
      Non-constructive proof: lower bound for probability, Lovasz
      local lemma
      Naïve algorithm: just resample false clause (while they
      exist)
      Recent breakthrough (Moser): this algorithm with high
      probability terminates quickly
      Explanation: if not, the sequence of resampled clauses
      would encode the random bits used in resampling making
      them compressible
                                                   .     .   .   .   .   .
Berry, Gödel, Chaitin, Raz




                             .   .   .   .   .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n




                                            .    .   .    .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true




                                              .   .    .   .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?




                                              .   .    .   .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?
      No, otherwise we could effectively generate string of
      complexity > n by enumerating all proofs




                                              .   .    .   .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?
      No, otherwise we could effectively generate string of
      complexity > n by enumerating all proofs
      and get Berry’s paradox: the first provable statement
      C(x) > n for given n gives some x of complexity > n that
      can be described by O(log n) bits




                                              .   .    .   .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?
      No, otherwise we could effectively generate string of
      complexity > n by enumerating all proofs
      and get Berry’s paradox: the first provable statement
      C(x) > n for given n gives some x of complexity > n that
      can be described by O(log n) bits
      (Gödel theorem in Chaitin form): There are only finitely
      many n such that C(x) > n is provable for some x.




                                              .   .    .   .     .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?
      No, otherwise we could effectively generate string of
      complexity > n by enumerating all proofs
      and get Berry’s paradox: the first provable statement
      C(x) > n for given n gives some x of complexity > n that
      can be described by O(log n) bits
      (Gödel theorem in Chaitin form): There are only finitely
      many n such that C(x) > n is provable for some x. (Note
      that ∃x C(x) > n is always provable!)




                                              .   .    .   .    .   .
Berry, Gödel, Chaitin, Raz
      There are only finitely many strings of complexity < n
      for all strings (except finitely many ones) x the statement
      K(x) > n is true
      Can all true statements of this form be provable?
      No, otherwise we could effectively generate string of
      complexity > n by enumerating all proofs
      and get Berry’s paradox: the first provable statement
      C(x) > n for given n gives some x of complexity > n that
      can be described by O(log n) bits
      (Gödel theorem in Chaitin form): There are only finitely
      many n such that C(x) > n is provable for some x. (Note
      that ∃x C(x) > n is always provable!)
      (Gödel second theorem, Kritchman–Raz proof): the
      “unexpected test paradox” (a test will be given next week
      but it won’t be known before the day of the test)
                                              .   .    .   .    .   .
13th Hilbert’s problem




                         .   .   .   .   .   .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients




                                             .    .   .    .    .    .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients
      Is it possible to represent this function as a composition of
      functions of at most 2 variables?




                                              .   .    .    .   .     .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients
      Is it possible to represent this function as a composition of
      functions of at most 2 variables?
      Yes, with weird functions (Cantor)




                                              .   .    .    .   .     .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients
      Is it possible to represent this function as a composition of
      functions of at most 2 variables?
      Yes, with weird functions (Cantor)
      Yes, even with continuous functions (Kolmogorov, Arnold)




                                              .   .    .    .   .     .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients
      Is it possible to represent this function as a composition of
      functions of at most 2 variables?
      Yes, with weird functions (Cantor)
      Yes, even with continuous functions (Kolmogorov, Arnold)
      Circuit version: explicit function Bn × Bn × Bn → Bn
      (polynomial circuit size?) that cannot be represented as
      composition of O(1) functions Bn × Bn → Bn . Not known.




                                              .   .    .    .   .     .
13th Hilbert’s problem
      Function of ≥ 3 variables: e.g., solution of a polynomial of
      degree 7 as function of its coefficients
      Is it possible to represent this function as a composition of
      functions of at most 2 variables?
      Yes, with weird functions (Cantor)
      Yes, even with continuous functions (Kolmogorov, Arnold)
      Circuit version: explicit function Bn × Bn × Bn → Bn
      (polynomial circuit size?) that cannot be represented as
      composition of O(1) functions Bn × Bn → Bn . Not known.
      Kolmogorov complexity version: we have three strings
      a, b, c on a blackboard. It is allowed to write (add) a new
      string if it simple relative to two of the strings on the board.
      Which strings we can obtain in O(1) steps? Only strings of
      small complexity relative to a, b, c, but not all of them (for
      random a, b, c)
                                               .    .    .    .    .     .
Secret sharing




                 .   .   .   .   .   .
Secret sharing

      secret s and three people; any two are able to reconstruct
      the secret working together; but each one in isolation has
      no information about it




                                            .   .    .   .    .    .
Secret sharing

      secret s and three people; any two are able to reconstruct
      the secret working together; but each one in isolation has
      no information about it
      assume s ∈ F (a field); take random a and tell the people
      a, a + s and a + 2s (assuming 2 ̸= 0 in F)




                                            .   .    .   .    .    .
Secret sharing

      secret s and three people; any two are able to reconstruct
      the secret working together; but each one in isolation has
      no information about it
      assume s ∈ F (a field); take random a and tell the people
      a, a + s and a + 2s (assuming 2 ̸= 0 in F)
      other secret sharing schemes; how long should be secrets
      (not understood)




                                            .   .    .   .    .    .
Secret sharing

      secret s and three people; any two are able to reconstruct
      the secret working together; but each one in isolation has
      no information about it
      assume s ∈ F (a field); take random a and tell the people
      a, a + s and a + 2s (assuming 2 ̸= 0 in F)
      other secret sharing schemes; how long should be secrets
      (not understood)
      Kolmogorov complexity settings: for a given s find a, b, c
      such that

      C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0




                                              .    .    .   .    .   .
Secret sharing

      secret s and three people; any two are able to reconstruct
      the secret working together; but each one in isolation has
      no information about it
      assume s ∈ F (a field); take random a and tell the people
      a, a + s and a + 2s (assuming 2 ̸= 0 in F)
      other secret sharing schemes; how long should be secrets
      (not understood)
      Kolmogorov complexity settings: for a given s find a, b, c
      such that

      C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0

      Some relation between Kolmogorov and traditional settings
      (Romashchenko, Kaced)

                                              .    .    .   .    .   .
Quasi-cryptography




                     .   .   .   .   .   .
Quasi-cryptography


     Alice has some information a




                                    .   .   .   .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b




                                        .   .   .   .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b
     by sending some message f




                                        .   .   .   .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b
     by sending some message f
     in such a way that Eve gets minimal information about b




                                          .   .    .   .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b
     by sending some message f
     in such a way that Eve gets minimal information about b
     Formally: for given a, b find f such that C(b|a, f) ≈ 0 and
     C(b|f) → max.




                                             .    .   .    .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b
     by sending some message f
     in such a way that Eve gets minimal information about b
     Formally: for given a, b find f such that C(b|a, f) ≈ 0 and
     C(b|f) → max.
     Theorem (Muchnik): it is always possible to have
     C(b|f) ≈ min(C(b), C(a))




                                             .    .   .    .   .   .
Quasi-cryptography


     Alice has some information a
     Bob wants to let her know some b
     by sending some message f
     in such a way that Eve gets minimal information about b
     Formally: for given a, b find f such that C(b|a, f) ≈ 0 and
     C(b|f) → max.
     Theorem (Muchnik): it is always possible to have
     C(b|f) ≈ min(C(b), C(a))
     Full version: Eve knows some c and we want to send
     message of a minimal possible length C(b|a)




                                             .    .   .    .   .   .
Andrej Muchnik (1958-2007)




                             .   .   .   .   .   .

More Related Content

What's hot

Information Gain
Information GainInformation Gain
Information Gainguest32311f
 
Minimax Rates for Homology Inference
Minimax Rates for Homology InferenceMinimax Rates for Homology Inference
Minimax Rates for Homology InferenceDon Sheehy
 
2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy
2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy
2013-1 Machine Learning Lecture 02 - Andrew Moore: EntropyDongseo University
 
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...Universitat Politècnica de Catalunya
 
Framework for the analysis and design of encryption strategies based on d...
Framework for the analysis and design of encryption strategies     based on d...Framework for the analysis and design of encryption strategies     based on d...
Framework for the analysis and design of encryption strategies based on d...darg0001
 
Winter school-pq2016v2
Winter school-pq2016v2Winter school-pq2016v2
Winter school-pq2016v2Ludovic Perret
 
Cryptanalysis Project Report
Cryptanalysis Project ReportCryptanalysis Project Report
Cryptanalysis Project Reportshahparin
 
Learning Financial Market Data with Recurrent Autoencoders and TensorFlow
Learning Financial Market Data with Recurrent Autoencoders and TensorFlowLearning Financial Market Data with Recurrent Autoencoders and TensorFlow
Learning Financial Market Data with Recurrent Autoencoders and TensorFlowAltoros
 
Social Network Analysis
Social Network AnalysisSocial Network Analysis
Social Network Analysisrik0
 
Lesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesLesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesMatthew Leingang
 
Lesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesLesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesMatthew Leingang
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Universitat Politècnica de Catalunya
 
Bayesian Subset Simulation
Bayesian Subset SimulationBayesian Subset Simulation
Bayesian Subset SimulationJulien Bect
 
Finite automata intro
Finite automata introFinite automata intro
Finite automata introlavishka_anuj
 
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Universitat Politècnica de Catalunya
 
Algorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlgorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlex Prut
 
Unit 3 greedy method
Unit 3  greedy methodUnit 3  greedy method
Unit 3 greedy methodMaryJacob24
 

What's hot (19)

Information Gain
Information GainInformation Gain
Information Gain
 
Neural Networks - How do they work?
Neural Networks - How do they work?Neural Networks - How do they work?
Neural Networks - How do they work?
 
Minimax Rates for Homology Inference
Minimax Rates for Homology InferenceMinimax Rates for Homology Inference
Minimax Rates for Homology Inference
 
2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy
2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy
2013-1 Machine Learning Lecture 02 - Andrew Moore: Entropy
 
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...
Deep Generative Models I (DLAI D9L2 2017 UPC Deep Learning for Artificial Int...
 
Framework for the analysis and design of encryption strategies based on d...
Framework for the analysis and design of encryption strategies     based on d...Framework for the analysis and design of encryption strategies     based on d...
Framework for the analysis and design of encryption strategies based on d...
 
Arvindsujeeth scaladays12
Arvindsujeeth scaladays12Arvindsujeeth scaladays12
Arvindsujeeth scaladays12
 
Winter school-pq2016v2
Winter school-pq2016v2Winter school-pq2016v2
Winter school-pq2016v2
 
Cryptanalysis Project Report
Cryptanalysis Project ReportCryptanalysis Project Report
Cryptanalysis Project Report
 
Learning Financial Market Data with Recurrent Autoencoders and TensorFlow
Learning Financial Market Data with Recurrent Autoencoders and TensorFlowLearning Financial Market Data with Recurrent Autoencoders and TensorFlow
Learning Financial Market Data with Recurrent Autoencoders and TensorFlow
 
Social Network Analysis
Social Network AnalysisSocial Network Analysis
Social Network Analysis
 
Lesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesLesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum Vaues
 
Lesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum VauesLesson 18: Maximum and Minimum Vaues
Lesson 18: Maximum and Minimum Vaues
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
 
Bayesian Subset Simulation
Bayesian Subset SimulationBayesian Subset Simulation
Bayesian Subset Simulation
 
Finite automata intro
Finite automata introFinite automata intro
Finite automata intro
 
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
 
Algorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlgorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography Theory
 
Unit 3 greedy method
Unit 3  greedy methodUnit 3  greedy method
Unit 3 greedy method
 

Viewers also liked

Csr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCsr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCSR2011
 
Exemples Gràfics de Funcions
Exemples Gràfics de FuncionsExemples Gràfics de Funcions
Exemples Gràfics de Funcions090452
 
LTR: Open Source Public Workstations
LTR: Open Source Public Workstations LTR: Open Source Public Workstations
LTR: Open Source Public Workstations koegeljm
 
Vademecum innovation
Vademecum innovationVademecum innovation
Vademecum innovationfcolombat
 
Cyia Next Levels Rev 11 07
Cyia Next Levels Rev 11 07Cyia Next Levels Rev 11 07
Cyia Next Levels Rev 11 07leecef
 
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печать
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печатьпрезентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печать
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печатьВиктор Сулимов
 
SIUE Cougar Athletics Branding Presentation 5 20-10
SIUE Cougar Athletics Branding Presentation 5 20-10SIUE Cougar Athletics Branding Presentation 5 20-10
SIUE Cougar Athletics Branding Presentation 5 20-10Elizabeth Keserauskis
 
Take Back Your Education (Ignite Portland 2013 Talk)
Take Back Your Education (Ignite Portland 2013 Talk)Take Back Your Education (Ignite Portland 2013 Talk)
Take Back Your Education (Ignite Portland 2013 Talk)Catherine Stevens
 

Viewers also liked (10)

Csr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCsr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatov
 
Exemples Gràfics de Funcions
Exemples Gràfics de FuncionsExemples Gràfics de Funcions
Exemples Gràfics de Funcions
 
LTR: Open Source Public Workstations
LTR: Open Source Public Workstations LTR: Open Source Public Workstations
LTR: Open Source Public Workstations
 
Vademecum innovation
Vademecum innovationVademecum innovation
Vademecum innovation
 
Call Back
Call BackCall Back
Call Back
 
Cyia Next Levels Rev 11 07
Cyia Next Levels Rev 11 07Cyia Next Levels Rev 11 07
Cyia Next Levels Rev 11 07
 
Doc. 113
Doc. 113Doc. 113
Doc. 113
 
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печать
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печатьпрезентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печать
презентация по итогам форума ключевых бизнес партнеровюв. херсон 2013 в печать
 
SIUE Cougar Athletics Branding Presentation 5 20-10
SIUE Cougar Athletics Branding Presentation 5 20-10SIUE Cougar Athletics Branding Presentation 5 20-10
SIUE Cougar Athletics Branding Presentation 5 20-10
 
Take Back Your Education (Ignite Portland 2013 Talk)
Take Back Your Education (Ignite Portland 2013 Talk)Take Back Your Education (Ignite Portland 2013 Talk)
Take Back Your Education (Ignite Portland 2013 Talk)
 

Similar to Csr2011 june15 09_30_shen

Kolmogorov complexity and topological arguments
Kolmogorov complexity and topological argumentsKolmogorov complexity and topological arguments
Kolmogorov complexity and topological argumentsshenmontpellier
 
02-VariableLengthCodes_pres.pdf
02-VariableLengthCodes_pres.pdf02-VariableLengthCodes_pres.pdf
02-VariableLengthCodes_pres.pdfJunZhao68
 
Lesson 19: Maximum and Minimum Values
Lesson 19: Maximum and Minimum ValuesLesson 19: Maximum and Minimum Values
Lesson 19: Maximum and Minimum ValuesMatthew Leingang
 
Probability and Statistics
Probability and StatisticsProbability and Statistics
Probability and StatisticsMalik Sb
 
Dependent Types and Dynamics of Natural Language
Dependent Types and Dynamics of Natural LanguageDependent Types and Dynamics of Natural Language
Dependent Types and Dynamics of Natural LanguageDaisuke BEKKI
 
Cuckoo Search Algorithm: An Introduction
Cuckoo Search Algorithm: An IntroductionCuckoo Search Algorithm: An Introduction
Cuckoo Search Algorithm: An IntroductionXin-She Yang
 
Breaking the Softmax Bottleneck: a high-rank RNN Language Model
Breaking the Softmax Bottleneck: a high-rank RNN Language ModelBreaking the Softmax Bottleneck: a high-rank RNN Language Model
Breaking the Softmax Bottleneck: a high-rank RNN Language ModelSsu-Rui Lee
 
Cheat Sheets for Hard Problems
Cheat Sheets for Hard ProblemsCheat Sheets for Hard Problems
Cheat Sheets for Hard ProblemsNeeldhara Misra
 
Random Oracle Model & Hashing - Cryptography & Network Security
Random Oracle Model & Hashing - Cryptography & Network SecurityRandom Oracle Model & Hashing - Cryptography & Network Security
Random Oracle Model & Hashing - Cryptography & Network SecurityMahbubur Rahman
 
從 VAE 走向深度學習新理論
從 VAE 走向深度學習新理論從 VAE 走向深度學習新理論
從 VAE 走向深度學習新理論岳華 杜
 
CiE 2010 talk
CiE 2010 talkCiE 2010 talk
CiE 2010 talkilyaraz
 
5 stream ciphers
5 stream ciphers5 stream ciphers
5 stream ciphersHarish Sahu
 
Chapter-4 combined.pptx
Chapter-4 combined.pptxChapter-4 combined.pptx
Chapter-4 combined.pptxHamzaHaji6
 
ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2zukun
 
An FPT Algorithm for Maximum Edge Coloring
An FPT Algorithm for Maximum Edge ColoringAn FPT Algorithm for Maximum Edge Coloring
An FPT Algorithm for Maximum Edge ColoringNeeldhara Misra
 
pptx - Psuedo Random Generator for Halfspaces
pptx - Psuedo Random Generator for Halfspacespptx - Psuedo Random Generator for Halfspaces
pptx - Psuedo Random Generator for Halfspacesbutest
 

Similar to Csr2011 june15 09_30_shen (20)

Kolmogorov complexity and topological arguments
Kolmogorov complexity and topological argumentsKolmogorov complexity and topological arguments
Kolmogorov complexity and topological arguments
 
Flat
FlatFlat
Flat
 
Losseless
LosselessLosseless
Losseless
 
02-VariableLengthCodes_pres.pdf
02-VariableLengthCodes_pres.pdf02-VariableLengthCodes_pres.pdf
02-VariableLengthCodes_pres.pdf
 
Unit 3
Unit 3Unit 3
Unit 3
 
Unit 3
Unit 3Unit 3
Unit 3
 
Lesson 19: Maximum and Minimum Values
Lesson 19: Maximum and Minimum ValuesLesson 19: Maximum and Minimum Values
Lesson 19: Maximum and Minimum Values
 
Probability and Statistics
Probability and StatisticsProbability and Statistics
Probability and Statistics
 
Dependent Types and Dynamics of Natural Language
Dependent Types and Dynamics of Natural LanguageDependent Types and Dynamics of Natural Language
Dependent Types and Dynamics of Natural Language
 
Cuckoo Search Algorithm: An Introduction
Cuckoo Search Algorithm: An IntroductionCuckoo Search Algorithm: An Introduction
Cuckoo Search Algorithm: An Introduction
 
Breaking the Softmax Bottleneck: a high-rank RNN Language Model
Breaking the Softmax Bottleneck: a high-rank RNN Language ModelBreaking the Softmax Bottleneck: a high-rank RNN Language Model
Breaking the Softmax Bottleneck: a high-rank RNN Language Model
 
Cheat Sheets for Hard Problems
Cheat Sheets for Hard ProblemsCheat Sheets for Hard Problems
Cheat Sheets for Hard Problems
 
Random Oracle Model & Hashing - Cryptography & Network Security
Random Oracle Model & Hashing - Cryptography & Network SecurityRandom Oracle Model & Hashing - Cryptography & Network Security
Random Oracle Model & Hashing - Cryptography & Network Security
 
從 VAE 走向深度學習新理論
從 VAE 走向深度學習新理論從 VAE 走向深度學習新理論
從 VAE 走向深度學習新理論
 
CiE 2010 talk
CiE 2010 talkCiE 2010 talk
CiE 2010 talk
 
5 stream ciphers
5 stream ciphers5 stream ciphers
5 stream ciphers
 
Chapter-4 combined.pptx
Chapter-4 combined.pptxChapter-4 combined.pptx
Chapter-4 combined.pptx
 
ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2
 
An FPT Algorithm for Maximum Edge Coloring
An FPT Algorithm for Maximum Edge ColoringAn FPT Algorithm for Maximum Edge Coloring
An FPT Algorithm for Maximum Edge Coloring
 
pptx - Psuedo Random Generator for Halfspaces
pptx - Psuedo Random Generator for Halfspacespptx - Psuedo Random Generator for Halfspaces
pptx - Psuedo Random Generator for Halfspaces
 

More from CSR2011

Csr2011 june14 09_30_grigoriev
Csr2011 june14 09_30_grigorievCsr2011 june14 09_30_grigoriev
Csr2011 june14 09_30_grigorievCSR2011
 
Csr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCsr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCSR2011
 
Csr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCsr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCSR2011
 
Csr2011 june18 14_00_sudan
Csr2011 june18 14_00_sudanCsr2011 june18 14_00_sudan
Csr2011 june18 14_00_sudanCSR2011
 
Csr2011 june18 15_45_avron
Csr2011 june18 15_45_avronCsr2011 june18 15_45_avron
Csr2011 june18 15_45_avronCSR2011
 
Csr2011 june18 09_30_shpilka
Csr2011 june18 09_30_shpilkaCsr2011 june18 09_30_shpilka
Csr2011 june18 09_30_shpilkaCSR2011
 
Csr2011 june18 12_00_nguyen
Csr2011 june18 12_00_nguyenCsr2011 june18 12_00_nguyen
Csr2011 june18 12_00_nguyenCSR2011
 
Csr2011 june18 11_00_tiskin
Csr2011 june18 11_00_tiskinCsr2011 june18 11_00_tiskin
Csr2011 june18 11_00_tiskinCSR2011
 
Csr2011 june18 11_30_remila
Csr2011 june18 11_30_remilaCsr2011 june18 11_30_remila
Csr2011 june18 11_30_remilaCSR2011
 
Csr2011 june17 17_00_likhomanov
Csr2011 june17 17_00_likhomanovCsr2011 june17 17_00_likhomanov
Csr2011 june17 17_00_likhomanovCSR2011
 
Csr2011 june17 16_30_blin
Csr2011 june17 16_30_blinCsr2011 june17 16_30_blin
Csr2011 june17 16_30_blinCSR2011
 
Csr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCsr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCSR2011
 
Csr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCsr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCSR2011
 
Csr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCsr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCSR2011
 
Csr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCsr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCSR2011
 
Csr2011 june17 11_00_lonati
Csr2011 june17 11_00_lonatiCsr2011 june17 11_00_lonati
Csr2011 june17 11_00_lonatiCSR2011
 
Csr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCsr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCSR2011
 
Csr2011 june17 15_15_kaminski
Csr2011 june17 15_15_kaminskiCsr2011 june17 15_15_kaminski
Csr2011 june17 15_15_kaminskiCSR2011
 
Csr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCsr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCSR2011
 
Csr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCsr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCSR2011
 

More from CSR2011 (20)

Csr2011 june14 09_30_grigoriev
Csr2011 june14 09_30_grigorievCsr2011 june14 09_30_grigoriev
Csr2011 june14 09_30_grigoriev
 
Csr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCsr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoff
 
Csr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoffCsr2011 june18 15_15_bomhoff
Csr2011 june18 15_15_bomhoff
 
Csr2011 june18 14_00_sudan
Csr2011 june18 14_00_sudanCsr2011 june18 14_00_sudan
Csr2011 june18 14_00_sudan
 
Csr2011 june18 15_45_avron
Csr2011 june18 15_45_avronCsr2011 june18 15_45_avron
Csr2011 june18 15_45_avron
 
Csr2011 june18 09_30_shpilka
Csr2011 june18 09_30_shpilkaCsr2011 june18 09_30_shpilka
Csr2011 june18 09_30_shpilka
 
Csr2011 june18 12_00_nguyen
Csr2011 june18 12_00_nguyenCsr2011 june18 12_00_nguyen
Csr2011 june18 12_00_nguyen
 
Csr2011 june18 11_00_tiskin
Csr2011 june18 11_00_tiskinCsr2011 june18 11_00_tiskin
Csr2011 june18 11_00_tiskin
 
Csr2011 june18 11_30_remila
Csr2011 june18 11_30_remilaCsr2011 june18 11_30_remila
Csr2011 june18 11_30_remila
 
Csr2011 june17 17_00_likhomanov
Csr2011 june17 17_00_likhomanovCsr2011 june17 17_00_likhomanov
Csr2011 june17 17_00_likhomanov
 
Csr2011 june17 16_30_blin
Csr2011 june17 16_30_blinCsr2011 june17 16_30_blin
Csr2011 june17 16_30_blin
 
Csr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCsr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhanin
 
Csr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhaninCsr2011 june17 09_30_yekhanin
Csr2011 june17 09_30_yekhanin
 
Csr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCsr2011 june17 12_00_morin
Csr2011 june17 12_00_morin
 
Csr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCsr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyi
 
Csr2011 june17 11_00_lonati
Csr2011 june17 11_00_lonatiCsr2011 june17 11_00_lonati
Csr2011 june17 11_00_lonati
 
Csr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatovCsr2011 june17 14_00_bulatov
Csr2011 june17 14_00_bulatov
 
Csr2011 june17 15_15_kaminski
Csr2011 june17 15_15_kaminskiCsr2011 june17 15_15_kaminski
Csr2011 june17 15_15_kaminski
 
Csr2011 june17 12_00_morin
Csr2011 june17 12_00_morinCsr2011 june17 12_00_morin
Csr2011 june17 12_00_morin
 
Csr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyiCsr2011 june17 11_30_vyalyi
Csr2011 june17 11_30_vyalyi
 

Csr2011 june15 09_30_shen

  • 1. CSR-2011 Kolmogorov complexity as a language Alexander Shen LIF CNRS, Marseille; on leave from ИППИ РАН, Москва . . . . . .
  • 2. Kolmogorov complexity . . . . . .
  • 3. Kolmogorov complexity A powerful tool . . . . . .
  • 4. Kolmogorov complexity A powerful tool Just a way to reformulate arguments . . . . . .
  • 5. Kolmogorov complexity A powerful tool Just a way to reformulate arguments three languages: combinatorial/algorithmic/probabilistic . . . . . .
  • 6. Kolmogorov complexity A powerful tool Just a way to reformulate arguments three languages: combinatorial/algorithmic/probabilistic . . . . . .
  • 7. Reminder and notation . . . . . .
  • 8. Reminder and notation K(x) = minimal length of a program that produces x . . . . . .
  • 9. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} . . . . . .
  • 10. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D . . . . . .
  • 11. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D optimal D makes it minimal up to O(1) additive term . . . . . .
  • 12. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D optimal D makes it minimal up to O(1) additive term Variations: . . . . . .
  • 13. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D optimal D makes it minimal up to O(1) additive term Variations: p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone of a sequence KR(x), KD(x) KM(x), Km(x) . . . . . .
  • 14. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D optimal D makes it minimal up to O(1) additive term Variations: p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone of a sequence KR(x), KD(x) KM(x), Km(x) Conditional complexity C(x|y): minimal length of a program p : y → x. . . . . . .
  • 15. Reminder and notation K(x) = minimal length of a program that produces x KD (x) = min{|p| : D(p) = x} depends on the interpreter D optimal D makes it minimal up to O(1) additive term Variations: p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone of a sequence KR(x), KD(x) KM(x), Km(x) Conditional complexity C(x|y): minimal length of a program p : y → x. There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes) . . . . . .
  • 16. Foundations of probability theory . . . . . .
  • 17. Foundations of probability theory Random object or random process? . . . . . .
  • 18. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? . . . . . .
  • 19. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] . . . . . .
  • 20. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) . . . . . .
  • 21. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1) . . . . . .
  • 22. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1) Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 . . . . . .
  • 23. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1) Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN . . . . . .
  • 24. Foundations of probability theory Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) ω = ω1 ω2 . . . is random iff KM(ω1 . . . ωn ) ≥ n − O(1) Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic ⇒ classical: Martin-Löf random sequences form a set of measure 1. . . . . . .
  • 25. Sampling random strings (S.Aaronson) . . . . . .
  • 26. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops . . . . . .
  • 27. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? . . . . . .
  • 28. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one . . . . . .
  • 29. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible . . . . . .
  • 30. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible not equivalent if no assumptions about the device . . . . . .
  • 31. Sampling random strings (S.Aaronson) A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible not equivalent if no assumptions about the device but are related under some assumptions . . . . . .
  • 32. Example: matrices without uniform minors . . . . . .
  • 33. Example: matrices without uniform minors k × k minor of n × n Boolean matrix: select k rows and k columns . . . . . .
  • 34. Example: matrices without uniform minors k × k minor of n × n Boolean matrix: select k rows and k columns minor is uniform if it is all-0 or all-1. claim: there is a n × n bit matrix without k × k uniform minors for k = 3 log n. . . . . . .
  • 35. Counting argument and complexity reformulation . . . . . .
  • 36. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] . . . . . .
  • 37. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) . . . . . .
  • 38. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest . . . . . .
  • 39. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 n2k × 2 × 2n = . . . . . .
  • 40. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 2 −9 log2 2 n2k × 2 × 2n = 2log n×2×3 log n+1+(n n) < 2n . . . . . . .
  • 41. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 2 −9 log2 2 n2k × 2 × 2n = 2log n×2×3 log n+1+(n n) < 2n . log n bits to specify a column or row: 2k log n bits in total . . . . . .
  • 42. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 2 −9 log2 2 n2k × 2 × 2n = 2log n×2×3 log n+1+(n n) < 2n . log n bits to specify a column or row: 2k log n bits in total one additional bit to specify the type of minor (0/1) . . . . . .
  • 43. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 2 −9 log2 2 n2k × 2 × 2n = 2log n×2×3 log n+1+(n n) < 2n . log n bits to specify a column or row: 2k log n bits in total one additional bit to specify the type of minor (0/1) n2 − k2 bits to specify the rest of the matrix . . . . . .
  • 44. Counting argument and complexity reformulation ≤ nk × nk positions of the minor [k = 3 log n] 2 types of uniform minors (0/1) 2 −k2 2n possibilities for the rest 2 −k2 2 −9 log2 2 n2k × 2 × 2n = 2log n×2×3 log n+1+(n n) < 2n . log n bits to specify a column or row: 2k log n bits in total one additional bit to specify the type of minor (0/1) n2 − k2 bits to specify the rest of the matrix 2k log n + 1 + (n2 − k2 ) = 6 log2 n + 1 + (n2 − 9 log2 n) < n2 . . . . . . .
  • 46. One-tape Turing machines copying n-bit string on 1-tape TM requires Ω(n2 ) time . . . . . .
  • 47. One-tape Turing machines copying n-bit string on 1-tape TM requires Ω(n2 ) time complexity version: if initially the tape was empty on the right of the border, then after n steps the complexity of a zone that is d cells far from the border is O(n/d). K(u(t)) ≤ O(n/d) . . . . . .
  • 48. One-tape Turing machines copying n-bit string on 1-tape TM requires Ω(n2 ) time complexity version: if initially the tape was empty on the right of the border, then after n steps the complexity of a zone that is d cells far from the border is O(n/d). K(u(t)) ≤ O(n/d) proof: border guards in each cell of the border security zone write down the contents of the head of TM; each of the records is enough to reconstruct u(t) so the length of it should be Ω(K(u(t)); the sum of lengths does not exceed time . . . . . .
  • 50. Everywhere complex sequences Random sequence has n-bit prefix of complexity n . . . . . .
  • 51. Everywhere complex sequences Random sequence has n-bit prefix of complexity n but some factors (substrings) have small complexity . . . . . .
  • 52. Everywhere complex sequences Random sequence has n-bit prefix of complexity n but some factors (substrings) have small complexity Levin: there exist everywhere complex sequences: every n-bit substring has complexity 0.99n − O(1) . . . . . .
  • 53. Everywhere complex sequences Random sequence has n-bit prefix of complexity n but some factors (substrings) have small complexity Levin: there exist everywhere complex sequences: every n-bit substring has complexity 0.99n − O(1) Combinatorial equivalent: Let F be a set of strings that has at most 20.99n strings of length n. Then there is a sequence ω s.t. all sufficiently long substrings of ω are not in F. . . . . . .
  • 54. Everywhere complex sequences Random sequence has n-bit prefix of complexity n but some factors (substrings) have small complexity Levin: there exist everywhere complex sequences: every n-bit substring has complexity 0.99n − O(1) Combinatorial equivalent: Let F be a set of strings that has at most 20.99n strings of length n. Then there is a sequence ω s.t. all sufficiently long substrings of ω are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik) . . . . . .
  • 56. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d . . . . . .
  • 57. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) . . . . . .
  • 58. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then < d/2 changed bits are harmless . . . . . .
  • 59. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then < d/2 changed bits are harmless but bit insertion or deletions could be . . . . . .
  • 60. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then < d/2 changed bits are harmless but bit insertion or deletions could be general requirement: C(xi |xj ) ≥ d . . . . . .
  • 61. Gilbert-Varshamov complexity bound coding theory: how many n-bit strings x1 , . . . , xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then < d/2 changed bits are harmless but bit insertion or deletions could be general requirement: C(xi |xj ) ≥ d generalization of GV bound: d-separated family of size Ω(2n−d ) . . . . . .
  • 62. Inequalities for complexities and combinatorial interpretation . . . . . .
  • 63. Inequalities for complexities and combinatorial interpretation C(x, y) ≤ C(x) + C(y|x) + O(log) . . . . . .
  • 64. Inequalities for complexities and combinatorial interpretation C(x, y) ≤ C(x) + C(y|x) + O(log) . . . . . .
  • 65. Inequalities for complexities and combinatorial interpretation C(x, y) ≤ C(x) + C(y|x) + O(log) C(x, y) ≥ C(x) + C(y|x) + O(log) . . . . . .
  • 66. Inequalities for complexities and combinatorial interpretation C(x, y) ≤ C(x) + C(y|x) + O(log) C(x, y) ≥ C(x) + C(y|x) + O(log) C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log) . . . . . .
  • 67. Inequalities for complexities and combinatorial interpretation C(x, y) ≤ C(x) + C(y|x) + O(log) C(x, y) ≥ C(x) + C(y|x) + O(log) C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log) every set A of size < 2k+l can be split into two parts A = A1 ∪ A2 such that w(A1 ) ≤ 2k and h(A2 ) ≤ 2l . . . . . .
  • 68. One more inequality . . . . . .
  • 69. One more inequality 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) . . . . . .
  • 70. One more inequality 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) V2 ≤ S1 × S2 × S3 . . . . . .
  • 71. One more inequality 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) V2 ≤ S1 × S2 × S3 Also for Shannon entropies; special case of Shearer lemma . . . . . .
  • 72. Common information and graph minors . . . . . .
  • 73. Common information and graph minors mutual information: I(a : b) = C(a) + C(b) − C(a, b) . . . . . .
  • 74. Common information and graph minors mutual information: I(a : b) = C(a) + C(b) − C(a, b) common information: . . . . . .
  • 75. Common information and graph minors mutual information: I(a : b) = C(a) + C(b) − C(a, b) common information: combinatorial: graph minors can the graph be covered by 2δ minors of size 2α−δ × 2β−δ ? . . . . . .
  • 76. Almost uniform sets . . . . . .
  • 77. Almost uniform sets nonuniformity= (maximal section)/(average section) . . . . . .
  • 78. Almost uniform sets nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog(N) sets whose nonuniformity is polylog(N). . . . . . .
  • 79. Almost uniform sets nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog(N) sets whose nonuniformity is polylog(N). multidimensional version . . . . . .
  • 80. Almost uniform sets nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog(N) sets whose nonuniformity is polylog(N). multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds . . . . . .
  • 81. Almost uniform sets nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog(N) sets whose nonuniformity is polylog(N). multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation . . . . . .
  • 82. Almost uniform sets nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog(N) sets whose nonuniformity is polylog(N). multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result) . . . . . .
  • 83. Shannon coding theorem . . . . . .
  • 84. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk . . . . . .
  • 85. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk ξ N : N independent trials of ξ . . . . . .
  • 86. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk ξ N : N independent trials of ξ Shannon’s informal question: how many bits are needed to encode a “typical” value of ξ N ? . . . . . .
  • 87. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk ξ N : N independent trials of ξ Shannon’s informal question: how many bits are needed to encode a “typical” value of ξ N ? Shannon’s answer: NH(ξ), where H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ). . . . . . .
  • 88. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk ξ N : N independent trials of ξ Shannon’s informal question: how many bits are needed to encode a “typical” value of ξ N ? Shannon’s answer: NH(ξ), where H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ). formal statement is a bit complicated . . . . . .
  • 89. Shannon coding theorem ξ is a random variable; k values, probabilities p1 , . . . , pk ξ N : N independent trials of ξ Shannon’s informal question: how many bits are needed to encode a “typical” value of ξ N ? Shannon’s answer: NH(ξ), where H(ξ) = p1 log(1/p1 ) + . . . + pn log(1/pn ). formal statement is a bit complicated Complexity version: with high probablity the value of ξ N has complexity close to NH(ξ). . . . . . .
  • 90. Complexity, entropy and group size . . . . . .
  • 91. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) . . . . . .
  • 92. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) The same for entropy: 2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ ) . . . . . .
  • 93. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) The same for entropy: 2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ ) …and even for the sizes of subgroups U, V, W of some finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|). . . . . . .
  • 94. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) The same for entropy: 2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ ) …and even for the sizes of subgroups U, V, W of some finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|). in all three cases inequalities are the same (Romashchenko, Chan, Yeung) . . . . . .
  • 95. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) The same for entropy: 2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ ) …and even for the sizes of subgroups U, V, W of some finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|). in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I(a : b) ≤ ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a) . . . . . .
  • 96. Complexity, entropy and group size 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) The same for entropy: 2H(ξ, η, τ ) ≤ H(ξ, η) + H(ξ, τ ) + H(η, τ ) …and even for the sizes of subgroups U, V, W of some finite group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|). in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I(a : b) ≤ ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a) Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a, b, e. . . . . . .
  • 98. Muchnik and Slepian–Wolf a, b: two strings . . . . . .
  • 99. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b . . . . . .
  • 100. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher . . . . . .
  • 101. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher there exist p : a → b that is simple relative to b, e.g., “map everything to b” . . . . . .
  • 102. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher there exist p : a → b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p : a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0 . . . . . .
  • 103. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher there exist p : a → b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p : a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0 information theory analog: Wolf–Slepian . . . . . .
  • 104. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher there exist p : a → b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p : a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0 information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) . . . . . .
  • 105. Muchnik and Slepian–Wolf a, b: two strings we look for a program p that maps a to b by definition C(p) is at least C(b|a) but could be higher there exist p : a → b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p : a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0 information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions . . . . . .
  • 106. Computability theory: simple sets . . . . . .
  • 107. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement . . . . . .
  • 108. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. . . . . . .
  • 109. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. Most strings are not simple ⇒ infinite complement . . . . . .
  • 110. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. Most strings are not simple ⇒ infinite complement Let x1 , x2 , . . . be a computable sequence of different non-simple strings . . . . . .
  • 111. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. Most strings are not simple ⇒ infinite complement Let x1 , x2 , . . . be a computable sequence of different non-simple strings May assume wlog that |xi | > i and therefore C(xi ) > i/2 . . . . . .
  • 112. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. Most strings are not simple ⇒ infinite complement Let x1 , x2 , . . . be a computable sequence of different non-simple strings May assume wlog that |xi | > i and therefore C(xi ) > i/2 but to specify xi we need O(log i) bits only . . . . . .
  • 113. Computability theory: simple sets Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C(x) ≤ |x|/2. Most strings are not simple ⇒ infinite complement Let x1 , x2 , . . . be a computable sequence of different non-simple strings May assume wlog that |xi | > i and therefore C(xi ) > i/2 but to specify xi we need O(log i) bits only “Minimal integer that cannot be described in ten English words” (Berry) . . . . . .
  • 114. Lower semicomputable random reals . . . . . .
  • 115. Lower semicomputable random reals ∑ ai computable converging series with rational terms . . . . . .
  • 116. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? . . . . . .
  • 117. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) . . . . . .
  • 118. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals . . . . . .
  • 119. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α . . . . . .
  • 120. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α There are maximal elements . . . . . .
  • 121. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α There are maximal elements = random lower semicomputable reals . . . . . .
  • 122. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α There are maximal elements = random lower semicomputable reals = slowly converging series . . . . . .
  • 123. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: ε → N(ε) = how many terms are needed for ε-precision . . . . . .
  • 124. Lower semicomputable random reals ∑ ai computable converging series with rational terms ∑ is α = ai computable (ε → ε-approximation)? not necessarily (Specker example) lower semicomputable reals Solovay classification: α ≼ β if ε-approximation to β can be effectively converted to O(ε)-approximation to α There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: ε → N(ε) = how many terms are needed for ε-precision maximal elements: N(2−n ) > BP(n − O(1)), where BP(k) is the maximal integer whose prefix complexity is k or less. . . . . . .
  • 125. Lovasz local lemma: constructive proof . . . . . .
  • 126. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . . . . . . .
  • 127. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables . . . . . .
  • 128. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each . . . . . .
  • 129. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors . . . . . .
  • 130. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors ⇒ CNF is satisfiable . . . . . .
  • 131. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors ⇒ CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma . . . . . .
  • 132. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors ⇒ CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) . . . . . .
  • 133. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors ⇒ CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly . . . . . .
  • 134. Lovasz local lemma: constructive proof CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . neighbors: clauses having common variables several clauses with k literals in each each clause has o(2k ) neighbors ⇒ CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible . . . . . .
  • 135. Berry, Gödel, Chaitin, Raz . . . . . .
  • 136. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n . . . . . .
  • 137. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true . . . . . .
  • 138. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? . . . . . .
  • 139. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity > n by enumerating all proofs . . . . . .
  • 140. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity > n by enumerating all proofs and get Berry’s paradox: the first provable statement C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits . . . . . .
  • 141. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity > n by enumerating all proofs and get Berry’s paradox: the first provable statement C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits (Gödel theorem in Chaitin form): There are only finitely many n such that C(x) > n is provable for some x. . . . . . .
  • 142. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity > n by enumerating all proofs and get Berry’s paradox: the first provable statement C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits (Gödel theorem in Chaitin form): There are only finitely many n such that C(x) > n is provable for some x. (Note that ∃x C(x) > n is always provable!) . . . . . .
  • 143. Berry, Gödel, Chaitin, Raz There are only finitely many strings of complexity < n for all strings (except finitely many ones) x the statement K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity > n by enumerating all proofs and get Berry’s paradox: the first provable statement C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits (Gödel theorem in Chaitin form): There are only finitely many n such that C(x) > n is provable for some x. (Note that ∃x C(x) > n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test) . . . . . .
  • 144. 13th Hilbert’s problem . . . . . .
  • 145. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients . . . . . .
  • 146. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most 2 variables? . . . . . .
  • 147. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most 2 variables? Yes, with weird functions (Cantor) . . . . . .
  • 148. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most 2 variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) . . . . . .
  • 149. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most 2 variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function Bn × Bn × Bn → Bn (polynomial circuit size?) that cannot be represented as composition of O(1) functions Bn × Bn → Bn . Not known. . . . . . .
  • 150. 13th Hilbert’s problem Function of ≥ 3 variables: e.g., solution of a polynomial of degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most 2 variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function Bn × Bn × Bn → Bn (polynomial circuit size?) that cannot be represented as composition of O(1) functions Bn × Bn → Bn . Not known. Kolmogorov complexity version: we have three strings a, b, c on a blackboard. It is allowed to write (add) a new string if it simple relative to two of the strings on the board. Which strings we can obtain in O(1) steps? Only strings of small complexity relative to a, b, c, but not all of them (for random a, b, c) . . . . . .
  • 151. Secret sharing . . . . . .
  • 152. Secret sharing secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it . . . . . .
  • 153. Secret sharing secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it assume s ∈ F (a field); take random a and tell the people a, a + s and a + 2s (assuming 2 ̸= 0 in F) . . . . . .
  • 154. Secret sharing secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it assume s ∈ F (a field); take random a and tell the people a, a + s and a + 2s (assuming 2 ̸= 0 in F) other secret sharing schemes; how long should be secrets (not understood) . . . . . .
  • 155. Secret sharing secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it assume s ∈ F (a field); take random a and tell the people a, a + s and a + 2s (assuming 2 ̸= 0 in F) other secret sharing schemes; how long should be secrets (not understood) Kolmogorov complexity settings: for a given s find a, b, c such that C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0 . . . . . .
  • 156. Secret sharing secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it assume s ∈ F (a field); take random a and tell the people a, a + s and a + 2s (assuming 2 ̸= 0 in F) other secret sharing schemes; how long should be secrets (not understood) Kolmogorov complexity settings: for a given s find a, b, c such that C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0 Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced) . . . . . .
  • 157. Quasi-cryptography . . . . . .
  • 158. Quasi-cryptography Alice has some information a . . . . . .
  • 159. Quasi-cryptography Alice has some information a Bob wants to let her know some b . . . . . .
  • 160. Quasi-cryptography Alice has some information a Bob wants to let her know some b by sending some message f . . . . . .
  • 161. Quasi-cryptography Alice has some information a Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b . . . . . .
  • 162. Quasi-cryptography Alice has some information a Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C(b|a, f) ≈ 0 and C(b|f) → max. . . . . . .
  • 163. Quasi-cryptography Alice has some information a Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C(b|a, f) ≈ 0 and C(b|f) → max. Theorem (Muchnik): it is always possible to have C(b|f) ≈ min(C(b), C(a)) . . . . . .
  • 164. Quasi-cryptography Alice has some information a Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C(b|a, f) ≈ 0 and C(b|f) → max. Theorem (Muchnik): it is always possible to have C(b|f) ≈ min(C(b), C(a)) Full version: Eve knows some c and we want to send message of a minimal possible length C(b|a) . . . . . .