Universit`a degli Studi di Roma Tor Vergata
MACROAREA DI SCIENZE MATEMATICHE FISICHE E NATURALI
Corso di Laurea Magistrale in Fisica
Tesi di laurea magistrale
BosonSampling validation with
integrated photonics
Candidato:
Luca Innocenti
Matricola 0206284
Relatore Interno:
Dott. Alessandro Cianchi
Relatore Esterno:
Prof. Fabio Sciarrino
Anno Accademico 2014-2015
2
List of Acronyms
AA Aaronson and Arkhipov . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
TM Turing machine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
ECT Extended Church-Turing thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
SPDC Spontaneous parametric downconversion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
HOM Hong-Ou-Mandel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
BS beam splitter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
ZTL Zero-Transmission Law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
MOL Mode Occupation List . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
MAL Mode Assignment List . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
BM binary matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
qFFT Fast Fourier Transform. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .90
3
4
Contents
Introduction 7
1 Foundations of classical and quantum information 11
1.1 Turing machine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.2 Church-Turing thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3 Complexity theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
1.4 Quantum information and quantum computation . . . . . . . . . . . . 20
2 Quantum and non-linear optics 23
2.1 Description of many-body states . . . . . . . . . . . . . . . . . . . . . . 23
2.2 Counting many-body states . . . . . . . . . . . . . . . . . . . . . . . . 30
2.3 Tools for quantum optics experiments . . . . . . . . . . . . . . . . . . . 33
2.3.1 Single-mode phase shift . . . . . . . . . . . . . . . . . . . . . . 33
2.3.2 Beam splitter . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
2.3.3 Wave plates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
2.3.4 Single-photon sources . . . . . . . . . . . . . . . . . . . . . . . 37
2.3.5 Single-photon detectors . . . . . . . . . . . . . . . . . . . . . . 39
2.4 Hong-Ou-Mandel effect . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3 BosonSampling 43
3.1 Importance of BosonSampling . . . . . . . . . . . . . . . . . . . . . . 43
3.2 The BosonSampling computational problem . . . . . . . . . . . . . . . 45
3.3 Scaling experimental boson sampling implementations . . . . . . . . . 51
3.4 Scattershot boson sampling . . . . . . . . . . . . . . . . . . . . . . . . . 52
4 Boson sampling validation 55
4.1 Validating boson sampling implementations . . . . . . . . . . . . . . . 55
4.2 Fourier suppression law . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
4.3 Fourier suppression law and scattershot boson sampling . . . . . . . . 65
4.4 Sylvester suppression law . . . . . . . . . . . . . . . . . . . . . . . . . . 70
5
5 Theoretical results 75
5.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
5.2 Main results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
5.3 Bayesian validation scheme . . . . . . . . . . . . . . . . . . . . . . . . . 82
5.4 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
6 Experimental boson sampling validation with integrated photonics 89
6.1 Overview of the experiment . . . . . . . . . . . . . . . . . . . . . . . . 90
6.2 Femtosecond laser writing . . . . . . . . . . . . . . . . . . . . . . . . . 90
6.3 Realization of the quantum Fast Fourier Transform . . . . . . . . . . . 92
6.4 Photon generation and manipulation . . . . . . . . . . . . . . . . . . . 94
6.5 One- and two-photon measurements in integrated Fourier interferometers 94
6.6 Observation of the suppression law . . . . . . . . . . . . . . . . . . . . 100
6.7 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
Conclusions 105
6
Introduction
Since the early 1980s, it has been argued [1] that simulating quantum systems is a
very challenging task. One source of difficulty is the number of parameters needed to
characterize a generic quantum system, which grows exponentially with the size of
the system. This means that even only storing the state of a large quantum system is
not feasible with classical computer memories. Furthermore, the number of operations
needed to simulate the temporal evolution of such a system also scales exponentially
with the size. Thus, the only way to avoid this exponential overhead in the evolution
is the use of approximation methods (such as Monte Carlo methods). However, for
many problems of interest, no good approximation scheme are available. It is then still
an open problem whether they can be efficiently simulated with a classical approach.
Hence, it is widely accepted that classical systems cannot in general efficiently simulate
quantum systems. While it is not yet possible to prove it, neither mathematically nor
experimentally, there are strong evidences to believe that this is the case.
This distinction between classical and quantum world has many implications. One
of the most notable examples concerns the possibility that computers exploiting the
weirdnesses of quantum mechanics may be able to carry out computations impossible
with only classical resources. With the current available technologies, the experimental
observation of this quantum advantage (sometimes referred to as quantum supremacy [2,
3]) has proven itself to be rather difficult to achieve. In particular, to observe a post-
classical computation with a universal quantum computer one first needs to solve the
problem of fault-tolerant quantum computation [4], which is known to be possible
in principle [5, 6, 7], but might require decoherence rates that are several orders of
magnitude below what achievable today. In the case of linear optics, a number of no-go
theorems led to the widespread belief that linear interferometry alone could not provide
a path to universal quantum computation. For this reason the result of Aaronson and
Arkhipov (AA), that passive linear optical interferometers with many-photon inputs
cannot be efficiently simulated by a classical computer [8], represented a significant
advance. The related computational problem, that is, sampling from the output proba-
bility distribution of such an apparatus, was named by AA the BosonSampling problem.
A quantum device able to efficiently solve it is referred to as a boson sampler.
More in detail, the BosonSampling computational problem consists in sampling
7
from the output probability distribution resulting from the time-evolution of n indis-
tinguishable photons into a random m × m unitary transformation. The hardness of
BosonSampling arises from the fact that the scattering amplitude between an input and
an output state configuration is proportional to the permanent of a suitable n×n matrix,
where the permanent is a particular function, defined similarly to the determinant,
which in the general case is known to be hard to compute classically. This immediately
suggests an experimental scheme to build a boson sampler using only linear optical
elements: just inject n indistinguishable photons into an appropriate linear optical
interferometer, and use photon-counting detectors to detect the resulting output states.
AA showed that, already with 20 < n < 30 and m n, this would provide direct
evidence that a quantum computer can solve a problem faster than what is possible with
any classical device. While this regime is far from our current technological capabilities,
several implementations of 2- and 3-photon devices have soon been reported [9, 10, 11,
12], and other more complex implementations followed [13, 14, 15, 16, 17, 18].
However, the originally proposed scheme to implement BosonSampling, that is,
to generate the input n-photon state through Spontaneous parametric downconver-
sion (SPDC), suffers from scalability problems. Indeed, it is unfeasible to generate high
numbers of indistinguishable input photons with this method, due to the generation
probability decreasing exponentially with n. For this reason, an alternative scheme,
named scattershot boson sampling [19], has been devised [20], and subsequently im-
plemented [18]. Contrarily to a classical boson sampler, a scattershot boson sampler
uses m SPDC sources, one per input mode of the interferometer, to generate random
(but known) n-photon input states, with n m. Each SPDC source generates a pair
of photons, one of which is injected into the interferometer, while the other is used
to herald the SPDC generation event. The use of a scattershot boson sampling scheme
results in an exponential increase of the probability of generating n indistinguishable
photons, for m and n large enough.
While the key part of BosonSampling resides in its simulation complexity, this very
hardness also poses a problem of certification of such a device. Indeed, it is believed [8]
that, when n is large enough, a classical computer cannot even verify that the device is
solving BosonSampling correctly. However, it is still possible to obtain circumstantial
evidence of the correct functioning of a device, and efficiently distinguish the output of
a boson sampler from that resulting from alternative probability distributions, like the
output produced by classical particles evolving through the same interferometer.
A number of validation schemes were subsequently devised to validate the output
resulting from true many-boson interference [14, 16, 21, 22, 23]. In particular, the tests
currently more suitable to identify true many-body interference [22] are those based
on Zero-Transmission Laws (ZTLs) [24]. A ZTL, also often referred to as suppression
law, is a rule which, for certain particular unitary evolutions, is able to predict that
the probability of certain input-output configurations is exactly zero, without having to
8
compute any permanent.
However, for the current validation schemes based on ZTLs, it is mandatory that
the input states possess particular symmetries. This requirement may thus be an issue
when the ZTLs are applied to validate a scattershot boson sampler. Indeed, the input
state in scattershot boson sampling is not fixed, but changes randomly at each n-photon
generation event. This mandates for a new ZTL-based validation scheme to be devised,
able to efficiently validate a scattershot boson sampling experiment, but still keeping
the capability of distinguishing alternative probability distributions.
In this thesis we report on both theoretical and experimental advances in the context
of validating classical and scattershot boson sampling experiments:
• From the theoretical point of view, we devise a new validation scheme, more
suitable to validate scattershot boson sampling experiments. This scheme, based
on a ZTL valid for a particular class of matrices, the so-called Sylvester matrices,
generalizes the ZTL reported in [23], presenting significantly higher predictive
capabilities.
• From the experimental point of view, we report on the experimental implementa-
tion [25] of the validation scheme proposed in [22] based on the ZTL for Fourier
matrices [24]. To this end, a scalable methodology to implement the Fourier
transformation on integrated photonics was adopted. This approach exploits the
3-D capabilities of femtosecond laser writing technique, together with a recently
proposed [26] quantum generalization of the Fast Fourier transform algorithm
[27], which allows a significant improvement in the number of elementary optical
elements required to implement the desired Fourier transformation.
The thesis is structured as follows: Chapter 1 opens with a brief survey of classical
and quantum computer science. In chapter 2, after a brief exposition of the theoretical
formalism for many-body quantum states, the fundamental tools used in quantum op-
tics experiments are presented. In chapter 3 the BosonSampling problem is introduced.
The problem of scaling boson sampling experiments is discussed, together with the
recently proposed alternative scheme named scattershot boson sampling. In chapter 4
the subject of boson sampling validation is introduced, and an outline of the proposed
solutions is provided. In particular, the focus is on the validation schemes based on
zero-transmission laws for Fourier and Sylvester matrices, and the possibility of apply-
ing them to scattershot boson sampling experiments. In chapter 5 we present a new
zero-transmission law for Sylvester matrices. Exploiting this zero-transmission law, we
present a scheme to validate scattershot boson sampling experiments. The thesis closes
with chapter 6, where we present the experimental implementation of a validation
scheme for Fourier matrices. The experimental and technological aspects of the experi-
ment are discussed, from the femtosecond laser-written technology employed to build
the integrated interferometers, to a novel method to efficiently implement the Fourier
9
transform on an integrated photonics chip. A full analysis of the chip reconstruction
and the observed suppression effects follows. The chapter closes with a discussion of
the usefulness of the presented work, and the possible future improvements.
The work presented in this thesis was carried out at the Quantum Information Lab,
University of Rome La Sapienza.
10
Chapter 1
Foundations of classical and
quantum information
Quantum information theory is a relatively recent field, and most of the main concepts
have been developed only in the last few decades. It is therefore natural that quantum
information relies heavily on ideas developed in the context of classical information
theory. These provide on one side tools that can be adapted to the quantum realm to
tackle quantum information problems, and on the other a benchmark to measure the
advantages provided by quantum resources with respect to their classical counterparts.
To meaningfully talk about the efficiency of an algorithm in performing a given
task, both in the classical and quantum context, it is necessary to have a mathematically
precise notion of what an “algorithm” is, and a mean to quantify its efficiency in solving
a given computational problem. The formalization of the concept of an algorithm
requires, in turn, the introduction of a computational model. One of the most important
computational models is the so-called Turing machine (TM) model of computation,
discussed in section 1.1. While a variety of other computational models have been
introduced and studied over the years, we will see that, thanks to the Church-Turing
thesis, it is enough to just consider the TM model in defining what algorithms can and
cannot do. Indeed, until a few decades ago, a much stronger conjecture was believed to
be true: the so-called Extended Church-Turing thesis states that any physically plausible
model of computation can be efficiently simulated by a TM. This last conjecture is
however currently under debate, as we now have reasons to believe that computational
devices able to solve certain problems exponentially faster than any classical computer
can indeed be devised. The Church-Turing thesis, as well as its extended version, is
discussed in section 1.2. To meaningfully assess whether an algorithm is more efficient
than another, it is necessary to introduce the idea of complexity classes, studied in the
field of computational complexity theory. These are classes of algorithms sharing some
common scaling properties, and are discussed in section 1.3.
11
1.1 Turing machine
While the concept of algorithm as a sequence of operations aimed to obtain a given result
may seem intuitively obvious, a mathematically precise formulation of the concept
was only given in 1930s, thanks to the work of Alonzo Church, Alan Turing, and other
pioneers of the computer era. This work resulted in the development of what may
arguably be considered as the most important computational model of computer science:
the Turing machine (TM) model.
A TM captures the notion of an algorithm performing a computational task, and is
composed of four main elements:
1. a program, which similarly to an ordinary computer program is a sequence of
program lines, each one describing an elementary operation for the TM,
2. a finite state control, which co-ordinates the other operations of the machine,
similarly to a modern computer CPU,
3. a tape, which acts like a computer memory,
4. a read-write tape-head, pointing to the position on the tape which is currently
readable or writable.
The finite state control for a TM consists of a finite set of internal states, q1, . . . , qm. The
number of states m can be varied, however it turns out that for m sufficiently large this
change does not affect the power of the machine in any essential way, so without loss
of generality we may suppose that m is some fixed constant. The finite state control
can be thought of as a sort of microprocessor, co-ordinating the TM’s operation. It
provides temporary storage off-tape, and a central place where all processing for the
machine may be done. In addition to the states q1, . . . , qm, there are also two special
internal states, labelled qs and qh. We call these the starting state and the halting state,
respectively. The idea is that at the beginning of the computation, the TM is in the
starting state qs. The execution of the computation causes the TM’s internal state to
change. If the computation ever finishes, the TM ends up in the state qh to indicate that
the machine has completed its operation.
The TM tape is a one-dimensional object, which stretches off to infinity in one
direction. The tape consists of an infinite sequence of tape squares numbered starting
from 0. Each tape square contains one symbol drawn from an alphabet, Γ, composed of
a finite number of distinct symbols. The read-write tape-head identifies a single square
on the TM tape as the square that is currently being accessed by the machine.
Summarizing, a TM starts its operation with the finite state control in the state
qs, and with the read-write head at the leftmost tape square. The computation then
proceeds step by step according to a predefined program. The computation is halted
12
when the current state is qh, and the output of the computation is the current contents
of the tape.
A program for a TM is a finite ordered list of program lines of the form q, x, q , x , s ,
where q, q are internal states of the machine, x, x are symbols of the alphabet Γ, and s
is equal to −1, +1, or 0.
At any given step of the computation, if q and x are the current internal state and
the symbol under the read-write tape-head respectively, the TM looks through the list
of program lines in order, searching for a line of the form q, x, ·, ·, · . If it does not find
such a line, the internal state is changed to qh and the machine halts. If such a line is
found, than that program line is executed: the internal state is changed to q , the current
symbol on the tape is overwritten by the symbol x , and the tape-head moves left, right,
or stands still, depending on whether s is −1, +1, or 0, respectively.
Example 1 (Computation of the constant function f(x) = 1). Consider a TM with
three internal states, q1, q2, and q3, in addition to the starting state qs and the halting
state qh. The used alphabet will contain four symbols: Γ = { , b, 0, 1}, with the zeros
and ones used to denote the input number, the to mark the beginning of the input
number, and the blanks (b) used on all the other cells of the tape. Finally, we program
the TM with the following program lines:
1 : qs, , q1, , +1 ,
2 : q1, 0, q1, b, +1 ,
3 : q1, 1, q1, b, +1 ,
4 : q1, b, q2, b, −1 ,
5 : q2, b, q2, b, −1 ,
6 : q2, , q3, , +1 ,
7 : q3, b, qh, 1, 0 .
If we start the TM with the tape containing a number expressed in binary form,
surrounded by blank (b) tapes, with the symbol marking the beginning of the input
number, and the tape-head starting on the cell containing , like the following:
b b b 1 1 1 0 1 b b b b
qs
start
. . .. . .
we will get as output the following state:
13
b b b 1 b b b b b b b b
qh
end
. . .. . .
To see this, we have to analyse how the program given to the TM acts on the
initial tape: starting in the state qs on a tape cell with the symbol , the first line of
the program mandates the tape-head to move right and switch to the state q1. In
the q1 state, following the lines 2 and 3, the tape-head will move right over-writing
all the ones and zeros it finds, until it reaches a blank cell. When a blank cell is
reached, according to line 4, the tape-head changes its state to q2 and starts moving
left, continuing moving left, following line 5, until it reaches again the cell. When
the cell is reached, the state is changed to q3 and the tape-head is moved once on the
right. At this point, because of the line 7 of the program, the cell under the tape-head
- that is, the cell on the right of the one with - is over-written to 1, and the tape-head
state changed to qh, halting the execution.
The above analysis shows that this program computes the constant function
f(x) = 1. That is, regardless of what number is given in input onto the tape, the TM
halts with the number 1 represented onto the tape.
In general, a TM can be thought of as computing functions from the non-negative
integers to the non-negative integers, with the input to the function represented by the
initial state of the tape, and the output of the function by the final state of the tape.
The above presented TM is used to formalize the concept of a deterministic algorithm.
To also consider non-deterministic algorithms, this model must however be extended. For
this purpose, the TM model is generalized to that of a probabilistic TM. In a probabilistic
TM, the state transitions are choosen according to some probability distribution, instead
of being completely predetermined.
A further generalization of TMs provides a theoretical basis for quantum algorithms.
These are a special kind of algorithms which, exploiting the properties of quantum
mechanics, can potentially outperform any classical algorithm in certain tasks.
1.2 Church-Turing thesis
An interesting question is what class of functions is it possible to compute using a Turing
machine. Despite its apparent simplicity, the TM model can be used to simulate all
the operations performed on a modern computer. Indeed, according to a thesis put
forward independently by Church and Turing, the TM model completely captures the
notion of computing a function using an algorithm. This is known as the Church-Turing
thesis [28]:
14
Church-Turing thesis: The class of functions computable by a Turing ma-
chine corresponds exactly to the class of functions which we would naturally
regard as being computable by an algorithm.
The Church-Turing thesis asserts an equivalence between the rigorous mathematical
concept of “function computable by a Turing machine”, and the intuitive concept of
what it means for a function to be computable by an algorithm. In this sense it is
nothing more than a definition of what we mean when we talk of the “computability” of
a function. This thesis is relevant because it makes the study of real-world algorithms
amenable to rigorous mathematical analysis.
We remark that it is not obvious that every function which we would intuitively
regard as computable by an algorithm can be computed using a TM. Indeed, it is conceiv-
able that in the future we will discover in Nature a process which computes a function
not computable by a TM. Up to now, however, no such process has been observed.
Indeed, as will be discussed in more detail in later sections, quantum computers also
obey the Church-Turing thesis. That is, quantum computers can compute the same
class of functions computable by a TM.
A much stronger statement than the Church-Turing thesis is the so-called Extended
Church-Turing thesis (ECT) [8] (also sometimes referred to as Strong Church-Turing
thesis [28]):
Extended Church-Turing thesis: All computational problems that are
efficiently solvable by realistic physical devices, are efficiently solvable by a
Turing machine.
The ECT was however already found to be insufficient to capture all realistic com-
putational models in the 1970s, when Solovay and Strassen [29] devised an efficient,
probabilistic primality test. As the Solovay-Strassen algorithm relied essentially on
randomness, it provided the first evidence that probabilistic Turing machines are capa-
ble to solve certain problems more efficiently than deterministic ones. This led to the
following ad-hoc modification to the ECT:
Extended Probabilistic Church-Turing thesis: All computational prob-
lems that are efficiently solvable by realistic physical devices, are efficiently
solvable by a probabilistic Turing machine.
As this is the form the ECT is currently usually stated as, this is the version we will refer
to when talking in the following of “ECT”.
However, even in this modified form, the ECT still does seem to be in contrast with
the currently accepted physical laws. The first evidence in this direction was given by
Shor [30], which proved that two very important problems - the problem of finding
the prime factors of an integer, and the so-called discrete logarithm problem - could
be solved efficiently on a quantum computer. Since no efficient classical algorithm -
neither deterministic nor probabilistic - is currently known to be able to efficiently solve
15
these problems, Shor’s algorithm strongly suggests that quantum mechanics allows
to solve certain problems exponentially faster than any classical computer, and this
directly contradicts the ECT.
1.3 Complexity theory
Computational complexity theory analyzes the time and space resources required to
solve computational problems [28]. Generally speaking, the typical problem faced
in computational complexity theory is proving some lower bounds on the resources
required by the best possible algorithm for solving a problem, even if that algorithm is
not explicitly known.
One difficulty in formulating a theory of computational complexity is that different
computational models may lead to different resource requirements for the same problem.
For instance, multiple-tape TMs can solve many problems significantly faster than single-
tape TMs. On the other hand, the strong Church-Turing thesis states that any model of
computation can be simulated on a probabilistic TM with at most a polynomial increase
in the number of elementary operations required. This means that if we make the
coarse distinction between problems which can be solved using resources which are
bounded by a polynomial in n, and those whose resource requirements grow faster than
any polynomial in n, then this distinction will be well-defined and independent of the
considered computational model. This is the chief distinction made in computational
complexity.
With abuse of the term exponential, the algorithms with resource requirements
growing faster than any polynomial in n are said to require an amount of resources
scaling exponentially in the problem size. This includes function like nlog n
, which grow
faster than any polynomial but lower than a true exponential, and are nonetheless said
to be scaling exponentially, in this context. A problem is regarded as easy, tractable, or
feasible, if an algorithm for solving the problem using polynomial resources exists, and
as hard, intractable, or infeasible, if the best possible algorithm requires exponential
resources.
Many computational problems are formulated as decision problems, that is problems
with a yes or no answer. For example, the question is a given number m a prime number
or not? is a decision problem. Although most decision problems can easily be stated in
simple, familiar language, discussions of the general properties of decision problems are
greatly helped by the terminology of formal languages. In this terminology, a language
L over the alphabet Σ is a subset of the set Σ∗
of all finite strings of symbols from Σ.
For example, if Σ = {0, 1}, then the set of binary representations of even numbers
L = {0, 10, 100, 110, . . . } is a language over Σ. A language L is said to be decided by a
TM if for every possible input x ∈ Σ∗
, the TM is able to decide whether x belongs to
L or not. In other words, the language L is decided if the TM will eventually halt in a
16
state encoding a “yes” answer if x ∈ L, and eventually halt to a state encoding a “no”
answer otherwise.
Decision problems are naturally encoded as problems about languages. For instance,
the primality decision problem can be encoded using the binary alphabet Σ = {0, 1},
interpreting strings from Σ∗
as non-negative integers, and defining the language L to
consist of all binary strings such that the corresponding number is prime. The primality
decision problem is then translated to the problem of finding a TM which decides the
language L. More generally, to each decision problem is associated a language L over
an alphabet Σ∗
, and the problem is translated to that of finding a TM which decides L.
To study the relations between computational problems, it is useful to classify them
into complexity classes, each one grouping all problems (that is, in the case of decision
problems, all languages) sharing some common properties. Most of computational
complexity theory is aimed at defining various complexity classes, and at understanding
of the relationships between different complexity classes.
A brief description of the most important complexity classes for decision problems
is provided in the following:
• P: We say that a given problem is in TIME(f(n)) if there is a deterministic TM
which decides whether a candidate x is in the corresponding language in time
O(f(n)), with n the length of x. A problem is said to be solvable in polynomial
time if it is in TIME(nk
) for some k. The collection of all languages which are in
TIME(nk
), for some k, is denoted P, which is an example of a complexity class
Some examples of problems in P are linear programming, the calculation of the
greatest common divisors of two numbers, and the problem of determining if a
number is prime or not.
Not surprisingly, there are lots of problems for which no polynomial-time algo-
rithm is known. Proving that a given decision problem is not in P, however, is very
difficult. A couple of examples of such problems are 1) given a non-deterministic
Turing machine M and an integer n written in binary, does M accept the empty
string in at most n steps? and 2) given a pair of regular expressions, do they represent
different sets?. Many other problems are believed to not be in P. Among these
are notable ones such as Factoring, which is the problem of finding the prime
factors decomposition of an integer. This problem is believed to hard problem for
classical computers, though no proof, nor compelling evidences for it, are known
to date. Factoring is particularly important, since its hardness lies at the heart
of wisely used algorithms in cryptography such as the RSA cryptosystem [28].
• NP: An interesting property of the prime factorization problem is that, even if
finding the prime factorization of an integer n is very hard, it is easy to check if a
proposed set of primes is indeed the correct factorization of n: just multiply the
numbers and check if they equal n. The class of decision problems sharing this
17
property is called NP. More generally NP, standing for “nondeterministic poly-
nomial time”, is the class of all decision problems for which there are efficiently
verifiable proofs. A NP problem can often be intuitively stated in the form are
there any solutions that satisfy certain constraints?
While it is clear that P is a subset of NP, the converse is currently not known.
Indeed, whether P equals NP is arguably the most famous open problem in
computer science, often abbreviated as the P = NP problem. Many computer
scientists believe [31, 32, 33] that P = NP. However, despite decades of work,
nobody has been able to prove this, and the possibility that P = NP cannot be
excluded. Some implications of either of these possibilities are shown in fig. 1.1.
A related complexity class is NP-hard, which groups all decision problems that
are, informally, at least as hard as the hardest problems in NP. More precisely, a
problem L is NP-hard if every problem in NP can be reduced to L in polynomial
time. As a consequence, a polynomial algorithm solving an NP-hard would
also automatically provide a polynomial algorithm for all problems in NP. While
this is considered highly unlikely, as many NP problems are believed to not be
solvable in polynomial time, it has never been proved that this is not the case.
Finally, the intersection between NP and NP-hard is the class of the so-called
NP-complete problems.
• BPP: If we extend our definition of a TM allowing it to have access to a source of
randomness, let’s say the ability to flip a fair coin, other complexity classes can
be defined. Such a probabilistic Turing machine may only accept or reject inputs
with a certain probability, but if the probability of an incorrect accept or reject is
low enough, they are as useful as their deterministic counterparts. One of the
most important such classes is BPP, which stands for Bounded-error Probabilistic
Polynomial time. BPP is the class of decision problems solvable by a probabilistic
TM in polynomial time with a probability of error less than 1/3.
The choice of 1/3 as error bound is mostly arbitrary, as any error bound strictly
less than 1/2 can be reduced to practically zero with only a small increase in
the resource requirements. For this reason, problems in BPP are regarded as
as efficiently solvable as P problems. In fact, for practical purposes, BPP is
considered, even more than P, as the class of problems which are efficiently
solvable on a classical computer.
18
Complexity
P ≠ NP P = NP
NP-Hard
NP-Complete
P
NP
NP-Hard
P = NP =
NP-Complete
Figure 1.1: Relations between the fundamental complexity classes.
All the above considered computational classes only took into account classical
Turing machines. The advent of quantum mechanics and the conception of quantum
computers, however, led to the question of what classes of problems can a quantum
computer solve? To try to answer this question, one must study another kind of
complexity classes, entering the realm of quantum complexity theory. In this context,
arguably the most fundamental complexity class if BQP, standing for Bounded-error
Quantum Polynomial time. This is the quantum generalization of BPP, and is defined
as the set of decision problems solvable by a quantum computer in polynomial time,
with an error probability of at most 1/3 for all instances. Probably the most notable
problem which has been shown to be in BQP is Factoring. Indeed, Shor’s algorithm
[30] was one of the first devised quantum algorithms able to efficiently solve a problem
that the best-known classical counterparts can solve only in exponential time.
While only decision problems have been mentioned to this point, these are not the
only kind of computational problems. Function problems are a generalized version of
decision problems, where the output of the algorithm is not bounded to be a simple
YES/NO answer. More formally, a function problem P is defined as a relation R over
the cartesian product over strings of an alphabet Σ, that is R ⊂ Σ∗
× Σ∗
. An algorithm
is said to solve P if for every pair (x, y) ∈ R, it produces y when given x as input.
A class of function problems that will be of interest in the following are the so-called
counting problems, which are problems that can be stated as how many X satisfy a certain
19
property P? An example of such a complexity class is #P (pronounced “sharp P”), which
is the set of counting problems associated with the decision problems in NP. Intuitively,
to each NP problem which can be stated in the form “Are there any solutions having
the property P?” is associated a #P problem which can be stated in the form “How many
solutions are there which satisfy the property P?”. As can be intuitively deduced from
this definition, #P problems are generally believed to be even more difficult than NP
problems.
A pair of computational classes related to #P are #P-hard and #P-complete. These
are defined in a completely analogous way to NP-hard and NP-complete, containing
the class of counting problems at least as hard as any #P problem.
A notable instance of a #P-hard problem is the calculation of the permanent of a
complex-valued matrix. The permanent is a function of matrices defined similarly to
the determinant (see the discussions in the following sections, like definition 2), but
which, contrarily to the latter, is known to not be computable in polynomial time, for
general matrices [34]. Indeed, the problem of computing the permanent of a matrix is
known to be a #P-hard problem, and even #P-complete in special circumstances [34].
1.4 Quantum information and quantum computation
Quantum information theory [28, 35, 36, 37] is the study of the information processing
tasks that can be accomplished using quantum mechanical systems. One of the main
goals of quantum information theory is to investigate how information is stored in the
state of a quantum system, how does it differ from that stored in a classical system, and
how can this difference be exploited to build quantum devices with capabilities superior
to that of their classical counterparts. To this end several concepts and ideas are drawn
from other disciplines, such as quantum mechanics, computer science, information
theory, and cryptography, and merged with the goal of generalizing the concepts of
information and computing to the quantum realm.
In the last few decades, information and computation theory have undergone a
spurt of new growth, expanding to treat the intact transmission and processing of
quantum states, and the interaction of such quantum information with traditional forms
of information. We now know that a fully quantum theory of information offers, among
other benefits, a brand of cryptography whose security rests on fundamental physics,
and a reasonable hope of constructing quantum computers that could dramatically speed-
up the solution of certain mathematical problems. Moreover, at a more fundamental
level, it has become clear that an information theory based on quantum principles
extends and completes classical information theory, much like complex numbers extend
and complete the reals. One of the conceptual building blocks of quantum information
and quantum computation is that of a qubit. This is the quantum generalization of
the classical concept of bit, and the fundamental processing unit of most quantum
20
devices. While a bit can be in one of two states, traditionally referred to as 0 and 1, a
qubit is allowed to be in a superposition of these basis states. Properly handling such
qubits, quantum computers are able to process information in ways impossible with
any classical computer.
The first to envisage the notion of a quantum computer was Feynman [1], as a
possible solution to the problem of the exponentially increasing amount of resources
required to simulate complex quantum systems with classical computers. More than a
decade later, Lloyd [38] showed that a quantum computer can indeed act as a universal
quantum simulator, where the word universal refers to the fact that the same machine
is capable of tackling vastly different problems by simply changing the program it runs.
There are a lot of candidate implementations for quantum computation. Among
these, in no particular order, are implementations using superconductors, trapped
ions, quantum dots, nuclear magnetic resonance, diamond nitrogen vacancies, silicon,
linear optics, and many other proposed technologies. Here we will only focus on linear
optical implementations of quantum computing, to highlight the difficulties inherent
to implement universal quantum computers, as opposite to the relatively much easier
demands of boson sampling devices, which will be described in the following sections.
Linear optics quantum computation (LOQC) with single photons has the advan-
tage that photons have very long decoherence times, which means that the quantum
information stored in photons tends to stay there, and that linear optical elements are
arguably the simplest building blocks to realize quantum information processing. The
downside is that photons do not naturally interact with each other, and in order to
apply two-qubit quantum gates, which are necessary to implement universal quantum
computation, such interactions are essential. Because of this, effective interactions
among photons have to be introduced somehow.
The two main methods to implement such interactions among photons are 1) using
Kerr nonlinearities, and 2) the use of projective measurements with photodetectors.
Unfortunately, present-day nonlinear Kerr media exhibit very poor efficiency [28] and
very weak nonlinearities, while projective measurements have the disadvantage of
producing probabilistic quantum gates: more often than not these gates fail, destroying
the quantum information.
In the case of projective-measurements-induced nonlinearities there is however
a way to avoid the issue of nondeterministic gates, still mantaning feasible resource
requirements: the Knill, Laflamme, and Milburn [7] (KLM) scheme. Introduced in 2001,
the KLM protocol allows scalable linear optics quantum computing by using quantum
gate teleportation to increase the probability of success of nondeterministic gates [7, 39,
40]. The downside of the KLM scheme is that, for its implementation, it is still necessary
to overcome a series of experimental challenges, such as the synchronization of pulses,
mode-matching, quickly controllable delay lines, tunable beam splitters and phase
shifters, single-photon sources, accurate, fast, single-photon detectors, and extremely
21
fast feedback control of these detectors. While most of these features are not terribly
unrealistic to implement, the experimental state of the art is simply not at the point at
which more complex gate operations such as two-qubit operations can be implemented.
On the other hand, a quantum computer is not necessary to implement quantum
simulation. Dropping the requirement of being able to simulate any kind of system,
special purpose devices can be built to tackle specific problems better than the clas-
sical counterparts, in the simpler conceivable case by just emulating, in an analog
manner, the behaviour of a complex quantum system on a simpler quantum device.
Being generally these special purpose devices easier to implement than full-fledged
quantum computers, it is expected that practical quantum simulation will become a
reality well before quantum computers. However, despite the undeniable practical
usefullness of implementing quantum simulation on a classically intractable quantum
system, this would hardly give a definite answer to the question: are there tasks which
quantum computers can solve exponentially faster than any classical computer? Indeed,
a quantum system that is hard to classically simulate is also typically hard to define as
a computational problem. This makes extremely difficult to definitively prove whether
a classical algorithm, able to efficiently carry out such a simulation, exists.
It is for this reasons that the proposal of a boson computer by Aaronson and Arkhipov
[8] gained much interest in the quantum optics community. This kind of special purpose
linear optics quantum computer requires only to send n indistinguishable photons
through a random unitary evolution, and detect the output photons with standard
photodetectors. No teleportation or feedback mechanisms are required, which makes
the experimental implementation of such a device much easier than that of a quantum
computer following the KLM scheme. Furthermore, the related computational problem
is simple enough to be analytically tractable with the tools of computational complexity,
allowing to obtain very strong theoretical evidence of its hardness.
22
Chapter 2
Quantum and non-linear optics
In section 2.1 the formalism required to describe many-body quantum states is intro-
duced. In particular, the focus is on many-boson states, and their evolution through
a unitary transformation. In section 2.2 a derivation of the total number of physical
states according to the bosonic, fermionic, and classical statistics is presented. In sec-
tion 2.3 the focus shifts to a more experimental point of view, and the main tools used
in quantum optics experiments are presented. The chapter closes with an outline of the
Hong-Ou-Mandel effect in section 2.4, which is a striking evidence of how quantum
interference results in highly non-classical consequences.
2.1 Description of many-body states
To describe a quantum mechanical system with a fixed number of particles, it is
enough to use a ket state |Ψ(t) , corresponding to a wave function Ψ(r1, . . . , rn, t) ≡
r1, . . . , rn|Ψ(t) . This is interpreted through Born’s rule as the probability amplitude
of the i-th particle being found at the position ri at the time t, that is, as a function such
that
|Ψ(r1, . . . , rn, t)|2
= probability of finding the i-th particle at the position ri, at the time t.
(2.1)
Such a wave function has to satisfy certain symmetrization rules, depending on the
statistical nature of the particles described, and its time-evolution is characterized by
the Schrödinger equation,
i
∂
∂t
|Ψ(t) = H |Ψ(t) , (2.2)
where H is the Hamiltonian of the system. The Hamiltonian is an operator, correspond-
ing to the total energy of the system in most of the cases, which characterizes the
dynamics of the system. Equation (2.2), together with an initial condition |Ψ(t0) , is
23
sufficient to obtain the wave function at each time t: its solution is, at least formally,
given by |Ψ(t) = e−i(t−t0)H/
|Ψ(t0) .
The above described formalism is also called first-quantization, to distinguish it
from another way of dealing with quantum systems, named second-quantization. The
latter differs from the former by a shift in focus: instead of considering the number
of particles as a fixed property of the system and using the wave function to describe
their states, the system is characterized by the number of particles contained in each
possible mode, which are however no longer necessarily fixed. To this end, a creation
operator is defined, for each mode, as an operator which acts on a quantum state and
produces another quantum state differing from the former for a single quantum added
to that mode. The hermitian conjugate of a creation operator is called an annihilation
operator (also destruction operator), and instead of adding a quantum to a given mode,
it does the opposite, producing a new state with one less particle in the mode.
The exact rules followed by creation and annihilation operators depend on the
statistical nature of the particles. For bosons, the creation (annihilation) operator of a
mode labelled i is denoted ˆa†
i (ˆai). The core rules obeyed by these operators are the
following:
ˆai |ni =
√
n |(n − 1)i , ˆa†
i |ni =
√
n + 1 |(n + 1)i ,
[ˆak, ˆa†
q] = δk,q, [ˆak, ˆaq] = [ˆa†
k, ˆa†
q] = 0 ,
(2.3)
where |ni is a state with n particles in the mode labelled i. Quantum states with a
well-defined number of particles in each mode are called Fock states, or number states,
and the set of all Fock states is called Fock space.
The above defined creation operators can be used to denote many-boson states, that
is, quantum states with more than one indistinguishable boson. If any single boson can
be in one of m modes, an n-boson state r having ri particles in the i-th mode will be
written as
|r ≡ |r1, . . . , rm =
1
√
r!
m
k=1
ˆa†
k
rk
|0 . (2.4)
where the notation r! ≡ r1! · · · rk! has been used. Such a list r of m elements, with each
element equal to the number of particles in a given mode, will be referred to as the
Mode Occupation List (MOL) associated to the quantum state. A many-body state such
that for every k = 1, . . . , m, rk = 1 or rk = 0, is said to be a collision-free state.
Another way to characterize many-body states is through a so-called Mode Assign-
ment List (MAL) R . This is a list of n elements, with the i-th element being the mode
occupied by the i-th particle. It is worth noting that for indistinguishable particles one
cannot talk of “the mode of the i-th particle”, however. Because of this, the order of
the elements of MALs cannot have any physical significance. In other words, MALs are
always defined up to the order of the elements, or, equivalently, they must be always
considered conventionally sorted (for example, in increasing order). Representing the
24
state with a MAL, eq. (2.4) can be rewritten in the following form:
|r ≡ |R =
1
√
r!
n
k=1
ˆa†
Rk
|0 =
1
µ(R)
n
k=1
ˆa†
Rk
|0 , (2.5)
where we denoted with µ(R) the product of the factorials of the occupation numbers
of the state R, that is, µ(R) ≡ r1! · · · rm!.
Yet another way to describe many-body states that will sometimes be useful is to
explicitly list the occupation numbers of each mode. For example, for a state with three
particles, one in the second mode and two in the fourth mode, we write |12, 24 . If
we want to emphasize the absence of particles in, say, the third mode, we write it as
|12, 03, 24 .
Definition 1 (MOL and MAL representations). All of the many-particle quantum
states used in the following will be assumed to have a fixed number of particles n,
with each particle potentially occupying one of m possible modes. Two ways to
represent a state are:
• As the Mode Occupation List (MOL) r ≡ (r1, . . . , rm), i.e. as the m-dimensional
vector whose element rk is the number of particles in the k-th mode. It follows
from this definition that m
k=1 rk = n. We will refer to this representation as
the MOL representation, and denote with Fn,m the set of all MOLs of n photons
into m modes, and with FCF
n,m the set of collision-free MOLs of n photons into
m modes:
Fn,m ≡ (r1, . . . , rm) | ∀i = 1, . . . , m, ri ≥ 0 and
m
i=1
ri = n , (2.6)
FCF
n,m ≡ {(r1, . . . , rm) ∈ Fn,m | ∀i = 1, . . . , m, ri ∈ {0, 1}} . (2.7)
• As the Mode Assignment List (MAL) R ≡ (R1, . . . , Rn), i.e. as the n-dimensional
vector listing the modes occupied by the particles. Given that for indistinguish-
able particles it is not meaningful to assign a specific mode to a specific particle,
the order of the elements of a MAL are conventionally taken to be in increas-
ing order, so to have a one-to-one correspondence between physical states
and MALs. We will refer to this representation as the MAL representation of a
many-particle quantum state and, following the notation of [41], denote with
Gn,m and Qn,m the set of all MALs of n photons into m modes and the set of
collision-free MALs of n photons into m modes, respectively. Equivalently, Gn,m
25
and Qn,m can be defined as particular sets of sequences of n integers, that is,
Gn,m ≡ {(R1, . . . , Rn) | 1 ≤ R1 ≤ · · · ≤ Rn ≤ m} ,
Qn,m ≡ {(R1, . . . , Rn) | 1 ≤ R1 < · · · < Rn ≤ m} .
(2.8)
Clearly, there is a one-to-one relation between Fn,m and Gn,m, and between FCF
n,m and
Qn,m, as these are just different ways to denote the same things:
Fn,m Gn,m, FCF
n,m Qn,m. (2.9)
Another class of sequences that will be useful in the following is Γn,m, that is, the
set of all mn
sequences ω = (ω1, . . . , ωn) of integers such that 1 ≤ ωi ≤ m for each
i = 1, . . . , n. We note that the sets Gn,m and Qn,m defined above can be thought of as
the subsets of non-decreasing and strictly increasing sequences of Γn,m, respectively.
A unitary evolution ˆU acts on creation operators in the following way:
ˆa†
j → ˆUˆb†
j
ˆU†
=
m
k=1
Ujk
ˆb†
k, (2.10)
where Ujk are the scattering amplitudes from the mode j to the mode k, and ˆb†
j are the
creation operators of the output states. See examples 2 and 3 for two simple applications
of eq. (2.10).
Example 2. As an example of the application of eq. (2.10) consider how the unitary
ˆU acts on a single-boson state |j = ˆa†
j |0 , where |0 is the vacuum state:
|j = ˆa†
j |0
ˆU
−−−−→ ˆUˆb†
j
ˆU†
|0 =
m
k=1
Ujk
ˆb†
k |0 =
m
k=1
Ujk |k out , (2.11)
which shows that eq. (2.10) is compatible with the usual rules of evolution of single-
particle states.
Example 3. A less trivial example is provided considering the evolution of a two-
boson state, with the two bosons initially in the modes i and j. The relation between
the creation operators and the two-bosons states is
if i = j, |i j = ˆa†
i ˆa†
j |0 , if i = j, |i j ≡ |i i =
(ˆa†
i )2
√
2
|0 .
26
These equations can be written more coincisely as |i j =
ˆa†
i ˆa†
j
√
µ(i,j)
|0 , where µ(i, j) is
equal to 1 or 2 if i = j and i = j, respectively.
Using this notation, we have
|i j =
ˆa†
i ˆa†
j
µ(i, j)
|0
ˆU
−−−−→
ˆUˆb†
i
ˆU† ˆUˆb†
j
ˆU†
µ(i, j)
|0 =
1
µ(i, j)
m
k=1
m
l=1
UikUjl
ˆb†
k
ˆb†
l |0
=
m
k=1
m
l=1
UikUjl
µ(i, j)µ(k, l)
|k l out .
(2.12)
We consider the four possibilities:
1. If i = j and k0 = l0, we have
out k0 l0|i j =
m
k=1
m
l=1
UikUjl(δk,k0 δl,l0 + δk,l0 δl,k0 ) = Ui,k0 Uj,l0 + Ui,l0 Uj,k0 ,
2. If i = j and k0 = l0,
out k0 l0|i j =
1
√
2
m
k=1
m
l=1
UikUjl(2δk,k0 δl,k0 ) =
√
2 Ui,k0 Uj,k0 ,
3. If i = j and k0 = l0,
out k0 l0|i j =
1
√
2
m
k=1
m
l=1
UikUil(δk,k0 δl,l0 + δk,l0 δl,k0 ) =
√
2 Ui,k0 Ui,l0 ,
4. If i = j and k0 = l0,
out k0 l0|i j =
1
2
m
k=1
m
l=1
UikUil(δk,k0 δl,k0 ) = Ui,k0 Ui,k0 .
All of these four cases are coincisely described by writing, for general values of
i, j, k0, l0,
out k0 l0|i j =
1
µ(i, j)µ(k, l)
perm
Ui,k0 Ui,l0
Uj,k0 Uj,l0
,
where perm(M) is the permanent of the matrix M (see definition 2).
27
Using eq. (2.5) into eq. (2.10) gives
|r
ˆU
−−−−→
1
√
r!
n
k=1


m
j=1
URk,j
ˆb†
j

 |0 =
1
√
r!
m
j1=1
m
j2=1
· · ·
m
jn=1
n
k=1
URk,jk
ˆb†
jk
|0
=
1
√
r! ω∈Γn,m
n
k=1
URk,ω(k)
ˆb†
ω(k) |0 ,
(2.13)
where Γn,m is the set of all sequences of n positive integers lesser than or equal to m.
To compute the scattering amplitudes A(r → s, U) of going from the input r to the
output s ≡ (s1, . . . , sn), we now have to rearrange the terms on the right hand side of
eq. (2.13). To this end we start from the general combinatorial equation (see [41]):
ω∈Γn,m
f(ω1, . . . , ωn) =
ω∈Gn,m
1
µ(ω) σ∈Sn
f(ωσ(1), . . . , ωσ(n)), (2.14)
where f(ω) ≡ f(ω1, . . . , ωn) is any function of n integer numbers, Gn,m is the sequence
of all non-decreasing sequences of n positive integers lesser than or equal to m, given
in definition 1, and Sn is the symmetric group, that is, the set of permutations of n
elements.
Applying eq. (2.14) to eq. (2.13), with f(ω1, . . . , ωn) = n
k=1 URkωk
ˆb†
ωk
, we obtain
|r
ˆU
−−−−→
1
µ(R) ω∈Gn,m
1
µ(ω) σ∈Sn
n
k=1
URk,ωσ(k)
ˆb†
ωσ(k)
|0
=
ω∈Gn,m
1
µ(R)µ(ω)


σ∈Sn
n
k=1
URk,ωσ(k)

 |ω1, . . . , ωn out ,
(2.15)
where in the last step we exploited the commutativity of the product of the creation
operators ˆb†
k.
We thus obtained the following expression for the scattering amplitudes for bosonic
particles:
A(r → s, U) ≡ out s|r ≡ s| ˆU |r =
1
µ(R)µ(S)


σ∈Sn
n
k=1
URk,Sσ(k)

 , (2.16)
in which the factor on the right hand side can be recognised as the permanent of an
appropriate matrix built from U.
28
Definition 2 (Permanent). The permanent of a square matrix, similarly to the de-
terminant, is a function which associates a number to a matrix. It is defined very
similarly to the determinant, with the exception that all minus signs that are present
for the latter become plus signs in the former.
More precisely, the permanent of a squared n×n matrix A = (aij) is
perm(A) =
σ∈Sn
a1,σ(1) · · · an,σ(n) =
σ∈Sn
n
k=1
ak,σ(k), (2.17)
where Sn is the symmetric group, that is, the set of all permutations of n distinct
objects.
To express eq. (2.16) through the above defined permanent function, it is also useful
to introduce the following notations to refer to particular submatrices built from a given
matrix:
Definition 3. Let Mk,l denote the set of all k×l complex-valued matrices. If k = l
we shall write Mk instead of Mk,k. Now, let A ∈ Mk,l, and let α ∈ Gp,k and β ∈ Gq,l.
Then, we shall denote with A[α|β] the p × q dimensional matrix with elements
A[α|β]i,j ≡ Aαi,βj
. If, moreover, α ∈ Qp,k and β ∈ Qq,l, then A[α|β] is a submatrix
of A. If α = β we will simplify the notation to write A[α] instead of A[α|β].
Again, if α ∈ Qp,k and β ∈ Qq,l, we shall denote with A(α|β) the (k − p)×(l − q)
dimensional submatrix of A complementary to A[α|β], that is, the submatrix obtained
from A by deleting rows α and columns β.
Example 4. Consider the 3×4 dimensional matrix A =



1 2 3 0
4 5 6 i
0 4 2 1


 . Then, if
α = (1, 1) ∈ G2,4 and β = (2, 4, 4) ∈ G3,3, we have A[α|β] ≡
2 0 0
2 0 0
. If
instead α = (2, 3) ∈ Q2,4 and β = (1, 2) ∈ Q2,3, we have A[α|β] ≡
4 5
0 4
and
A(α|β) ≡ 3 0 .
Using definitions 2 and 3, we see that for any matrix A ∈ Mm and sequences
α, β ∈ Gn,m, we have
perm(A[α|β]) =
σ∈Sn
n
k=1
Aαk,βσ(k)
,
29
which is just the factor in brackets on the right hand side of eq. (2.16). We conclude that
A(r → s, U) ≡ out s|r ≡ s| ˆU |r =
1
µ(R)µ(S)
perm(U[R|S]). (2.18)
It is worth noting that the symmetric nature of the bosonic creation operators - that
is, the fact that [ˆa†
i , ˆa†
j] = δij - was essential in the above derivation. Analogous
reasonings carried out using fermionic particles - whose creation operators satisfy
{ˆc†
i , ˆc†
j} ≡ c†
i c†
j + c†
jc†
i = δij - would lead to the result
Afermions
(r → s, U) = det(U[R|S]). (2.19)
While eqs. (2.18) and (2.19) may seem very similar at first glance, especially given
the similarities in the definitions of permanents and determinants, they are extremely
different when trying to actually compute these scattering amplitudes. Indeed, while
the determinant of an n dimensional matrix can be efficiently computed, the same is
in general not true for permanents. This is exactly what makes the BosonSampling
problem interesting, and will be described in detail in the next sections.
In conclusion, we can now write down the unitary matrix describing the evolution
of the many-body states, that is, the matrix Uα,β(m, n, U), with α, β ∈ Gn,m, whose
elements contain the permanents (or the determinants, in the case of fermions) of the
corresponding matrix given by eq. (2.18):
Uα,β(m, n, U) ≡
1
µ(α)µ(β)
perm(U[α|β]), α, β ∈ Gn,m. (2.20)
The dependence of U on n, m, and U will often be omitted when clear from the context.
Example 5. Consider the 3×3 unitary matrix Ui,j ≡ ui,j, injected with 2-photon
input states. The resulting many-boson scattering matrix is:
U(3, 2, U) =











u2,3u3,2 + u2,2u3,3 u2,3u3,1 + u2,1u3,3 u2,2u3,1 + u2,1u3,2
√
2u2,3u3,3
√
2u2,2u3,2
√
2u2,1u3,1
u1,3u3,2 + u1,2u3,3 u1,3u3,1 + u1,1u3,3 u1,2u3,1 + u1,1u3,2
√
2u1,3u3,3
√
2u1,2u3,2
√
2u1,1u3,1
u1,3u2,2 + u1,2u2,3 u1,3u2,1 + u1,1u2,3 u1,2u2,1 + u1,1u2,2
√
2u1,3u2,3
√
2u1,2u2,2
√
2u1,1u2,1√
2u3,2u3,3
√
2u3,1u3,3
√
2u3,1u3,2 u2
3,3 u2
3,2 u2
3,1√
2u2,2u2,3
√
2u2,1u2,3
√
2u2,1u2,2 u2
2,3 u2
2,2 u2
2,1√
2u1,2u1,3
√
2u1,1u1,3
√
2u1,1u1,2 u2
1,3 u2
1,2 u2
1,1











2.2 Counting many-body states
Another interesting property of many-body states is their number. Unlike classical
states, it is not meaningful to assign a state to the single particles of a quantum many-
body state. A quantum state is instead described only giving the list of modes occupied
30
by the particles, without reference to which particle is in which mode. The above
example, for indistinguishable particles, would therefore read like the following: “the
two particles are in the modes 1 and 3”. This has the remarkable consequence of
changing the total number of possible quantum states of n particles into m modes, with
respect to the classical case.
The number of classical states of n particles into m modes is easily computed: each
one of the n particles can be in one of m modes, independently of the state of the
others,so we have
m × m × · · · × m
n
= mn
(2.21)
possible states.
On the other hand, if we were to describe quantum many-body states with this
notation (which amounts to using the MAL representation defined in definition 1), we
would have to take into account that MALs differing only for a permutation of the
elements represent the same quantum state. Moreover, in the case of many-fermion
states, Pauli exclusion principle mandates that no more than one particle can occupy the
same mode, posing an additional constraint to the possible many-particle configurations.
A simple example of the differences between the above described types of many-body
states is given in table 2.1, for the case m = 4, n = 2.
classical states
(1,1) (1,2) (1,3) (1,4)
(2,1) (2,2) (2,3) (2,4)
(3,1) (3,2) (3,3) (3,4)
(4,1) (4,2) (4,3) (4,4)
many-boson states
(1,1) (1,2) (1,3) (1,4)
(2,1) (2,2) (2,3) (2,4)
(3,1) (3,2) (3,3) (3,4)
(4,1) (4,2) (4,3) (4,4)
many-fermion states
(1,1) (1,2) (1,3) (1,4)
(2,1) (2,2) (2,3) (2,4)
(3,1) (3,2) (3,3) (3,4)
(4,1) (4,2) (4,3) (4,4)
Table 2.1: Comparison of classical, many-boson,
and many-fermion states. Many-particle states for
n = 2 particles into m = 4 modes, in MAL notation.
The striked out states represent non-physical states. In
the many-boson case, these are due to the indistinguisha-
bility of MALs differing only for the order of the elements.
In the many-fermion case, the additional restrictions im-
posed by Pauli’s principle must be taken into account.
To count the number of many-boson states of n particles into m modes, we will
make use of the MOL representation. The problem is thus to find the number of different
31
sequences of m non-negative integers, with the sum of the integer equal to n. A visual
schematic of this problem is that given in fig. 2.1: each of the m numbers is represented
as the number of marbles (or whatever object) in the corresponding bin. From this
representation it becomes clear that the question can be equivalently stated as: in how
many ways can n objects be distributed among m bins? An easy way to compute this
number is starting from fig. 2.1: if we permute in every possible way all of the marbles
and the inner separators between the bins, we span the whole set of many-boson states.
The number of such permutations is equal to the number of permutations of m − 1 (the
number of inner separators) plus n (the number of marbles) objects, that is, (m−1+n)!.
However, this overcounts the number of many-boson states, having treated as different
configurations with the same occupation numbers but some marbles, or separators,
interchanged. The final number must therefore be normalized over the number of such
“transparent” permutations. We conclude that the number of many-boson states is
|Fn,m| = |Gn,m| =
(m + n − 1)!
(m − 1)!n!
=
m + n − 1
n
. (2.22)
Similarly, the number of many-fermion states, equal to the number of collision-free
many-boson states, is
|FCF
n,m| = |Qn,m| =
m!
n!(m − n)!
=
m
n
. (2.23)
The difference between these numbers increases exponentially with n, as seen in fig. 2.2.
Many-boson states Classical statesa) b)
Figure 2.1: Examples of many-particle states of 4 particles into 6 modes. The disk represent the
particles, and each bin a different mode. (a) In the case of many-boson states, the disks (that is, the
particles) are all identical to each other, and therefore the number of disks in a given bin completely
specifies the state. (b) For classical states, the distinguishability of the particles implies that there can be
more than one state with the same number of particles for each mode.
32
Figure 2.2: Number of classical (blue), many-boson (orange) and many-fermion (green) states, for m = 8
and m = 32, plotted against the number of photons, in logarithmic scale.
2.3 Tools for quantum optics experiments
The basic building blocks used to manipulate light in linear optics are beam splitters,
half- and quarter-wave plates, and phase shifters. For quantum optics experiments,
single photon sources and detectors are also required.
In this section we will give a brief description of these tools, and establish the
mathematical conventions used in the following sections.
2.3.1 Single-mode phase shift
This optical component changes the phase of the electromagnetic field in a given mode,
which means that it acts on the creation operator of a given mode k in the following
way:
ˆa†
in
phaseshifter
−→ ˆa†
out = eiφ ˆNin
ˆa†
ine−iφ ˆNin
= eiφ
ˆa†
in. (2.24)
Physically, a phase shifter can be implemented as a slab of transparent material with
an index of refraction that is different from that of free space.
2.3.2 Beam splitter
A beam splitter (BS), which is a central component of many optical experiments, consists
of a semireflective mirror: when light impinges onto the mirror, part of it will be reflected
and part will be transmitted. In a classical description, a BS can be simply characterized
by the relations
E3 = R31E1 + T31E2,
E4 = T41E1 + R42E2,
(2.25)
where E1, E2 are the amplitudes of the incoming electric fields, E3, E4 the amplitudes
of the outcoming ones, and the complex coefficients Rij, Tij are respectively the re-
33
flectances and transmittances along a particular path through the BS (while we will only
consider the electric fields here, completely analogour relations can be written for the
corresponding magnetic fields). In matrix notation eq. (2.25) are written as
E3
E4
=
R31 T32
T41 R42
E1
E2
, (2.26)
where the above 2x2 matrix is known as the beam splitter matrix. In the ideal case
of a lossless BS, considerations of energy conservation lead to the following relations
between the coefficients of the BS:
|R31|2
+ |T41|2
= |R42|2
+ |T32|2
= 1 and R31T ∗
32 + T41R∗
42 = 0, (2.27)
from which it follows that |R31| = |R42| ≡ |R| and |T41| = |T32| ≡ |T |. Using these
relations we can rewrite eq. (2.26) in the form:
E3
E4
=
cos θ ie−iφ
sin θ
ieiφ
sin θ cos θ
E1
E2
, (2.28)
where θ and φ parametrize the set of all possible 2x2 unitary matrices. The general
structure of the BS can be further simplified by additional assumptions on the forms of
the reflection and transmission coefficients. In the case of symmetrical BSs, for example,
we can assume the coefficients to satisfy
R31 = R42 ≡ R = |R| exp(iφR) and T32 = T41 ≡ T = |T | exp(iφT ), (2.29)
with |R|2
+ |T |2
= 1 and φR − φT = ±π/2, which translates in the condition φ =
0 in the notation of eq. (2.28). In the case of a 50:50 beam splitter with reflection
and transmission coefficients of equal magnitude we have |R| = |T | = 1/
√
2 and
φR − φT = π/2, which translates in the conditions φ = 0 and θ = π/4 in the notation
of eq. (2.28).
The relations in eq. (2.25) convert into analogous relations between the quantized
field operators:
ˆa†
3 = cos θˆa†
1 + ie−iφ
sin θˆa†
2 and ˆa†
4 = ieiφ
sin θˆa†
1 + cos θˆa†
2, (2.30)
or, in the case of 50:50 BSs, in
ˆa†
3 =
1
√
2
ˆa†
1 + iˆa†
2 and ˆa†
4 =
1
√
2
iˆa†
1 + ˆa†
2 . (2.31)
Inverting eq. (2.31) leads to
ˆa†
1 =
1
√
2
ˆa†
3 − iˆa†
4 and ˆa†
2 =
1
√
2
−iˆa†
3 + ˆa†
4 . (2.32)
34
An equivalent way to denote eq. (2.32) is obtained by denoting with ˆa†
1, ˆa†
2 the two input
modes and with ˆb†
1,ˆb†
2 the two output modes, and describing the evolution through the
beam splitter with the same notation used in eq. (2.10):
ˆa†
1 →
1
√
2
ˆb†
1 − iˆb†
2 and ˆa†
2 →
1
√
2
−iˆb†
1 + ˆb†
2 . (2.33)
An important type of BS is the polarizing beam splitter (PBS). This is a device which
distinguishes different polarization states of the incoming light. If the PBS is cut to
separate horizontal and vertical linear polarizations, the transformation of the incoming
modes, ˆa1 and ˆa2, yields the following outcoming modes:
ˆa1,H → ˆa1,H and ˆa1,V → ˆa2,V ,
ˆa2,H → ˆa2,H and ˆa2,V → ˆa1,V ,
(2.34)
that is, it does not change the spatial mode of the horizontally polarized states, but
switches the spatial mode of vertically polarized ones.
E1
E2
E3
E4
Figure 2.3: Schematic repre-
sentation of a lossless beam
splitter, with the notation
used in the text for incom-
ing and outcoming electric
fields.
2.3.3 Wave plates
A waveplate is an optical component which alters the polarization of a light wave
travelling through it. The two most commonly used types of waveplates are half- and
quarter-wave plates. The former rotates the polarization direction of linearly polarized
light, while the latter converts linearly polarized light into circularly polarized light
and vice versa.
35
The same mathematical description given above for BSs applies to the evolution of
light through a waveplate. Instead of having two different spatial modes, ˆa†
1 and ˆa†
2, the
two incoming modes have two different polarizations, that we will denote with ˆax and
ˆay. The equations describing the evolution of the creation operators of the polarization
modes of the field are thus
ˆa†
x = cos θˆa†
x + ie−iφ
sin θˆa†
y and ˆa†
y = ieiφ
sin θˆa†
x + cos θˆa†
y, (2.35)
where x and y are the polarization modes of the light after the evolution through the
waveplate. In the notation of eq. (2.35), half-wave plates correspond to φ = π/2, and
quarter-wave plates to φ = 0.
Figure 2.4: Left: Example of an optical beam splitter as commonly seen in quantum optics laboratories.
Right: Example of a phase shifter, as commonly seen in quantum optics laboratories.
36
2.3.4 Single-photon sources
An ideal single-photon source [42, 43, 44] would be one that
1. Is deterministic (or “on demand”), meaning that it can emit a single photon at any
arbitrary time defined by the user,
2. Has a 100% probability of emitting a single photon and a 0% probability of multiple-
photon emission,
3. Subsequently emitted photons are indistinguishable,
4. The repetition rate is arbitrarily fast.
Given however that no real-world photon source satisfies all of these specifications,
the deviations from the ideal characteristics must be considered when designing exper-
iments.
Single-photon sources are broadly classified into deterministic and probabilistic.
Among the implementations of the former are those based on color centers [45, 46, 47],
quantum dots [48, 49, 50], single atoms [51], single ions [52], single molecules [53],
and atomic ensembles [54], all of which can to some degree emit single photons “on
demand”. On the other hand are the probabilistic single-photon sources. These generally
rely on photons created in pairs via parametric downconversion in bulk crystals and
waveguides, and four-wave mixing in optical fibers. While these sources are probabilistic
- and therefore it is not possible to know exactly when a photon has been emitted -
because the photons are created in pairs, one of the emitted photons can be used to
herald the creation of the other.
While the distinction between deterministic and probabilistic sources is clear in the
abstract, this distinction blurs in practice. This due to the unavoidable experimental
errors that make also “theoretically deterministic sources” be probabilistic in practice.
Although many applications, especially those in the field of quantum-information
science, require an on-demand source of single photons, probabilistic single-photon
sources remain a fundamental tool, and are widely used in many quantum optics
experiments.
Spontaneous parametric downconversion (SPDC) is an important process in quan-
tum optics, typically exploited to generate entangled photon pairs, or heralded single
photons. This is achieved using a nonlinear crystal - that is, a medium in which the
dielectric polarization responds nonlinearly to the electric field - which converts the
photons of a pump beam into pairs of photons of lower energy. A simple model of the
interaction Hamiltonian in such a crystal is
HI ∼ χ(2)
ˆapˆa†
sˆa†
i + hermitian conjugate, (2.36)
37
where χ(2)
is the second-order nonlinear susceptibility of the nonlinear medium. Here
ˆap is the annihilation operator of the pump beam, and ˆa†
s and ˆa†
i are the creation
operators of the signal and idler beams respectively (the names signal and idler are
there for historical reasons, and carry no special significance throughout this thesis).
In the simplest case, with the signal and idler beams initially in vacuum states, a single
photon from the pump beam is converted into two optical photons, one in the signal
beam and the other in the idler:
|1 p |0 s |0 i
SPDC
−−−→ ˆapˆa†
sˆa†
i |1 p |0 s |0 i = |0 p |1 s |1 i . (2.37)
The conditions of energy and momentum conservation pose a contraint on the signal and
idler generated photons, which will depend on the pump beam through the following
relations (see also fig. 2.5):
ωp = ωs + ωi,
kp = ks + ki.
(2.38)
kp
ks ki
Figure 2.5: Phase-matching
condition
There are two main types of SPDC processes, named type I and type II:
In type I, the signal and idler photons have the same polarization, orthogonal to
that of the pump.
In type II downconversion, the polarizations of signal and idler photons are instead
orthogonal to each other. Because of birefringence effects, the generated photons are
emitted along two cones, one for the ordinary wave and another for the extraordinary
wave. At the intersection of the cones, the two photons detected are in a polarization-
entangled state of the form
1
√
2
(|V s |H i + |H s |V i), (2.39)
where |H and |V denote an horizontally and vertically polarized state respectively. A
schematic representation of this process is shown in fig. 2.6.
More generally, the state produced by an SPDC source does not always contain two
photons, but has the form
∞
p=0
gp
|p1 |p2 , (2.40)
38
with |pi = (ˆa†
i )p
/
√
p!, and 0 ≤ g < 1 an appropriate parameter determining the ratio
of generated photons and dependent, among other things, on the strength of the pump
beam. Typically, g 1, so that the probability of generating many photons is low. For
instance, in typical conditions g ∼ 0.1 and the probability of generating a state of the
form |21 |22 is lower of a factor ∼ 102
than the probability of producing a single pair
|11 |12 .
The main advantages of SPDC sources are the high photon indistinguishability, the
collection efficiency, and relatively simple experimental setups. This technique, however,
suffers from two drawbacks. First, since the nonlinear process is nondeterministic, so is
the photon generation, even though it can be heralded. Second, the laser pump power,
and hence the source’s brilliance, has to be kept low to prevent undesired higher-order
terms in the photon generation process.
Figure 2.6: Representation of type II downconversion. The pump beam (red) impinges on the nonlinear
crystal, generating, due to birefringence effects, photons along two cones. On the upper ordinary cone
(orange), the generated photons are horizontally polarized, while on the lower extraordinary cone (green)
the generated photons are vertically aligned. Postselecting on the two intersections of these cones (blue
dots), a pair of polarization-entangled photons is obtained.
2.3.5 Single-photon detectors
Roughly speaking, single-photon detectors are devices which convert single photons
into an electrical signal of some sort [42]. Quantum information science is one of the
field currently driving much of the research toward improved single-photon-detector
technology. For example, many quantum communication protocols rely heavily on
39
detector properties such as detection efficiency.
An ideal single-photon detector [42] would require the following characteristics:
1. The detection efficiency - that is, the probability that a photon incident upon the
detector is successfully detected - is 100%,
2. The dark-count rate - that is, the rate of detector output pulses in the absence of
any incident photons - is zero,
3. The dead time - that is, the time after a photon-detection event during which the
detector is incapable of detecting another photon - is zero,
4. The timing jitter - that is, the variation from event to event in the delay between
the input of the optical signal and the output of the electrical signal - is zero.
Additionally, an ideal single-photon detector would also be able to count the number of
photons in an incident pulse. Detectors able to do this are referred to as photon-counting,
or photon-number resolving, detectors. However, non-photon-number-resolving detec-
tors, which can only distinguish between zero photons and more than zero photons,
are the most commonly used. Indeed, while detecting a single photon is a difficult task,
discriminating the number of incident photons is even more difficult. Examples of non-
photon-number-resolving single-photon detector technologies include single-photon
avalanche photodiodes [55], quantum dots [56], superconducting nanowires [57], and
up-conversion detectors [58, 59, 60].
2.4 Hong-Ou-Mandel effect
The differences between bosons and fermions are not only in the different numbers of
microstates. Their statistical behaviour can differ significantly, as well as be significantly
different from the behaviour of distinguishable particles.
Bosons, roughly speaking, tend to occupy the same state more often than classical
particles, or fermions, do. This behaviour, referred to as bosonic bunching, has been
verified in numerous experimental circumstances, including fundamental ones like
Bose-Einstein condensation [61, 62, 63]. In the context of optical experiments, the most
known effects arising from the symmetric nature of the Bose-Einstein statistics is the
Hong-Ou-Mandel (HOM) effect [64].
In the original experiment, two photons are sent simultaneously through the two
input ports of a symmetric beam splitter. Since no interaction between the two photons
takes place, one would expect no correlation between the detection events at the two
output ports. Instead, the photons are always seen either both on the first output mode,
or both on the second output mode.
40
+ − +
Figure 2.7: Pictorial representation of the suppression of non-bunched events, when two indistinguish-
able bosons evolve through a symmetric beam splitter. Each of the four images represent a possible
evolution of the bosons, with all of them interfering with each other. The two events which would result
in one boson per output port turn out to interfere destructively (note the minus sign), and are suppressed.
This effect is a direct consequence of the quantum interference between the possible
ways two-photon states can evolve. A pictorial representation of this is given in fig. 2.7:
when the photons are injected into two different ports of a symmetric beam splitter,
the scattering amplitudes corresponding to the output photons being in two different
modes interfere destructively.
We can derive this result applying eq. (2.33) to the two-photon input state |11, 12 =
ˆa†
1ˆa†
2 |0 :
|11, 12 = ˆa†
1ˆa†
2 |0 →
1
√
2
ˆb†
1 − iˆb†
2
1
√
2
−iˆb†
1 + ˆb†
2 |0
=
1
2
−i(ˆb†
1)2
+ ˆb†
1
ˆb†
2 − ˆb†
2
ˆb†
1 − i(ˆb†
2)2
=
−i
2
(ˆb†
1)2
+ (ˆb†
2)2
|0 =
−i
√
2
(|21 + |22 ) ,
(2.41)
where in the last steps we used the rules given in eq. (2.3), and in particular the
commutativity of the bosonic creation operators, which implies that ˆb†
1
ˆb†
2 = ˆb†
2
ˆb†
1 (see
also fig. 2.7 for a pictorial representation of how the suppression of non-bunched events
arises). We thus conclude that when two indistinguishable photons enter a symmetric
beam splitter one in each mode, they always come out in the same mode.
This property of photons (or, more generally, of bosons) is highly non-classical, and
is a notable example of how interesting effects can arise when dealing with many-body
quantum states.
In a real world experiment, the two input photons will never be perfectly indistin-
guishable, though. A more careful analysis, taking into account the potentially different
times at which the photons reach the beam splitter, as well as the coherence time of
each photon wave packet, predicts a smooth transition from the classical behaviour to
the antibunching effect described above [65].
In the general case of partially distinguishable particles, the probability PT (s, x) of
detecting an output state s in an HOM experiment with a time delay quantified by x,
becomes an average of the probability PB for bosons and the probability PD assigned
to distinguishable particles, weighted by two factors |c1|2
and |c2|2
which depend on
41
the relative time delay:
PT (s, x) = |c1(x)|2
PB(s) + |c2(x)|2
PD(s). (2.42)
In a typical experimental scenario, with the incoming photons having a Gaussian
frequency distribution around the central frequency, PT (s, x) also has a gaussian profile
[65], as shown in a typical case in fig. 2.8.
In the case of fully distinguishable particles, where PT = PD, no interaction
occurs and the output events of the single photons are not correlated. There are
two possible classical configurations of two photons in the output MOL (1, 1), and
one configuration for both (2, 0) and (0, 2). It follows that PD((1, 1)) = 2/4 and
PD((2, 0)) = PD((0, 2)) = 1/4, as shown in fig. 2.8 for x ≈ ±400µm. On the other
hand, when the particles are fully indistinguishable, PT = PB. The probability of the
various outcomes is now given by eq. (2.41). The output (1, 1) is thus suppressed, while
PB((2, 0)) = PB((0, 2)) = 1/2, as shown in fig. 2.8 for x = 0.
Figure 2.8: Transition to indistinguishability in a HOM experiment. Changing the time delay between
the two input photons a dip in the number of measured output coincidences is seen, corresponding
to the time delay (or, equivalently, the path delay) making the photons indistinguishable. The blue
line is the probability of detecting two photons in one of the two output ports, that is, PT ((2, 0), x) or
equivalently PT ((0, 2), x). The red line is the probability PT ((1, 1), x) of detecting the two photons in
the two different output ports. As expected, the red plot shows a peak corresponding to the antibunching
effect arising when the particles are indistinguishable, while the blue plot show the HOM explained above.
42
Chapter 3
BosonSampling
In this chapter we discuss various aspects of the BosonSampling computational prob-
lem. In section 3.1 the problem of experimentally assessing quantum supremacy is
discussed, in order to appreciate the importance of BosonSampling in the modern
research context. Section 3.2 follows with the description of what the BosonSampling
computational problem is, and its advantages in obtaining experimental evidences
of quantum supremacy. In section 3.3 some issues related to the scalability of boson
sampling implementations are described. The chapter closes with a description of
scattershot boson sampling in section 3.4, as an alternative architecture to scale boson
sampling implementations to higher numbers of photons.
3.1 Importance of BosonSampling
It is currently believed that many quantum mechanical systems cannot be efficiently
simulated with a classical computer [1]. This implies that a quantum device is, to the
best of our knowledge, able to solve problems de facto beyond the capabilities of classical
computers. Exploiting this quantum advantage requires however an high degree of
control over the quantum system, not yet manageable with state of the art technology.
In particular, a post-classical computation with a universal quantum computer will
require an high degree of control of a large number of qubits, and this implies that an
experimental evidence of quantum supremacy [3] with an universal quantum computer
will likely require many years. A notable example is given by the large gap between the
number of qubits that can currently be coherently controlled (∼10), and the number
of qubits required for a calculation such as prime factorization, on a scale that would
challenge classical computers (∼106
). Consequently, there is considerable interest in
non-universal quantum computers and quantum simulators that, while able to only
solve specific problems, might be significantly easier to be implemented experimentally.
Such devices could give the first experimental demonstration of the power of quantum
43
devices over classical computers, and potentially lead to technologically significant
applications.
Moreover, in the context of searching for experimental evidence of quantum supremacy,
the technological difficulties are not the only issue. To show this, we will consider as an
example Shor’s quantum algorithm [30] to efficiently factorize integer numbers. Even if
we were to get past the technological difficulties of implementing Shor’s algorithm with
sufficiently many qubits, it could be easily argued that such an achievement would not
be a conclusive evidence that quantum mechanics allows post-classical computations.
This because we do not have to date a mathematically sound proof that there cannot be a
classical algorithm to efficiently factorize integers. In the language of complexity theory,
this corresponds to the fact that we do not have a proof that Factoring is not in P,
even though this is believed enough to base modern cryptography is based on this
conjecture. More generally, before 2010, there were no instances of problems efficiently
solved by quantum computers, which were proved to not be efficiently solvable with
classical ones.
This changed when, in 2010, Aaronson and Arkhipov (AA) proposed [8] the Boson-
Sampling problem as a way to obtain an easier experimental evidence of quantum
supremacy. BosonSampling is a computational problem that, while hard to solve for
a classical computer, is efficiently solved by a special-purpose quantum device. AA
showed that BosonSampling is naturally implemented using only linear optical ele-
ments, in a photonic platform named a boson sampler. The experimental realization
of a boson sampler, while still challenging with present-day technologies, requires
much less experimental efforts with respect to those required to build a universal
quantum computer. In fact, the AA scheme requires only linear optical elements and
photon-counting detectors, as opposite to, for example, the Knill, Laflamme & Milburn
approach [7, 40] for universal linear optics quantum computing, which requires among
other things an extremely fast feedback control of the detectors.
44
Figure 3.1: Galton board: n identical balls are dropped one by one from the upper corner, and are
randomly scattered into the lower slots. The quantum generalization of this classical “rudimentary
computer” leads to the idea of BosonSampling. Credits: Nicolò Spagnolo.
3.2 The BosonSampling computational problem
Consider the following linear optical experiment: the n-photon state |rAA given by
|rAA ≡ |11, . . . , 1n, 0n+1, . . . , 0m ≡ ˆa†
1 . . . ˆa†
n |01, . . . , 0m , (3.1)
is injected into a passive linear optics network, which implements a unitary map on
the creation operators:
ˆa†
k → ˆUˆb†
k
ˆU†
=
m
j=1
Uk,j
ˆb†
j. (3.2)
with U an Haar-random m×m complex unitary matrix. The evolution induced on
|rAA is
|rAA →
s∈Fn,m
A(r → s, U) |s , (3.3)
where the sum is extended over all many-boson states of n particles into m modes, and
the scattering amplitudes A are, as shown in eq. (2.18), proportional to the permanents
of n×n matrices. AA argued that, for m n, the output of such an apparatus cannot
be efficiently predicted by a classical computer, neither exactly nor approximately [8].
This was rigorously proved in the exact case. The problem of approximately sampling
45
from the output probability distribution of a boson sampling apparatus depends instead
on a series of conjectures, for which strong supporting evidence was provided [8].
This problem, which amounts to that of being able to sample from the output proba-
bility distribution given in eq. (2.18), is referred to as the BosonSampling computational
problem. The constraint m n is essential for the hardness result, as otherwise
semi-classical methods become efficient [66, 67].
Roughly speaking, a boson sampling apparatus is a “quantum version” of a Galton
board. A Galton board, named after the English scientist Sir Francis Galton, is an
upright board with evenly spaced pegs into its upper half, and a number of evenly-
spaced rectangular slots in the lower half (see fig. 3.1). This setup can be imagined to
be a rudimentary “computer”, where n identical balls are dropped one by one from
the upper corner, and are randomly scattered into the lower slots. In the quantum
mechanical version, the n balls are indistinguishable bosons “dropped” simultaneously,
and each peg a unitary transformation, typically implemented as a set of beam splitters
and phase shifters.
More precisely, BosonSampling consists in producing a fair sample of the output
probability distribution P(s |U, rAA) ≡ |A(r → s, U)|2
, where s is an output state of
the n bosons, and rAA the above mentioned input state. The unitary ˆU and the input
state rAA are the input of the BosonSampling problem, while a number of output states
sampled from the correct probability distribution are its solution (see fig. 3.2).
46
Figure 3.2: Conceptual boson sampling apparatus. (a) The input of the BosonSampling problem is
the input many-photon state (in figure the state |0, 0, 1, 1, 0, 1, 0, 0, 0 ), and a suitably chosen unitary U.
The output is a number of outcomes picked according to the bosonic output probability distribution (in
figure, two examples of such states are provided, with MOLs 101000100 and 110000100). Colleting enough
such events allows to reconstruct the probability distribution. This, however, requires an exponentially
increasing (in n) number of events. (b) Injecting an m-mode unitary with n indistinguishable photons,
the output state is a weighted superposition of all possible outcomes. Measuring in which modes the
photons ended up results in the collapsing of this superposition. The probability of finding the photons
in a certain configuration is given by eq. (3.4). Credits: [10].
47
Example 6 (Solution of the BosonSampling problem). Let ˆU be a randomly chosen
unitary transformation, described by the 4×4 matrix
U =
−0.60046+0.220549i −0.108966−0.527875i −0.367627+0.212122i 0.0655053 +0.340358i
−0.576174−0.386133i −0.463806+0.524027i 0.0458648 −0.0209767i 0.148456 −0.0679879i
−0.0337116−0.30791i 0.408837 −0.0733373i −0.0106664+0.578256i 0.543236 −0.319264i
−0.0894887−0.0760134i 0.211627 +0.0494216i 0.680731 +0.139364i 0.0591935 +0.672804i
.
The output probability distribution resulting from the injection of the two-photon
input state with MOL r = (1, 1, 0, 0) is
output state probability amplitude probability
(0,0,0,2) 0.046478 + 0.0651595 i 0.00640597
(0,0,1,1) -0.0300108+0.0707215 i 0.00590217
(0,0,2,0) -0.0175525 + 0.0246647 i 0.000916437
(0,1,0,1) -0.260804-0.194492 i 0.105846
(0,1,1,0) 0.0432791 -0.312955 i 0.099814
(0,2,0,0) 0.462674 + 0.265491 i 0.284553
(1,0,0,1) 0.0195337 -0.147833 i 0.0222362
(1,0,1,0) 0.270811 +0.0424448 i 0.0751401
(1,1,0,0) 0.0218767 -0.0707259 i 0.00548074
(2,0,0,0) 0.609711 + 0.148185 i 0.393706
Measuring the exit modes of the two injected photons at each pulse, we obtain
a series of samples from the above probability distribution. An example of 10 such
samples is the sequence
(0,2,0,0)
(0,1,1,0)
(2,0,0,0)
(2,0,0,0)
(0,2,0,0)
(0,1,1,0)
(2,0,0,0)
(0,1,0,1)
(0,2,0,0)
(0,2,0,0)
The above list is exactly what BosonSampling is all about: obtaining a list of “labels”
distributed according to a particular probability distribution.
In other words, the BosonSampling problem is not solved obtaining the above
listed probabilities, but obtaining a number of samples from this distribution. The
number of such samples is not really important here, even producing a single state
48
from the correct probability distribution would theoretically be enough to achieve
a post-classical computation, though possibly making it harder to experimentally
verify.
The hardness of the BosonSampling problem can be traced back to the #P-hardness
of computing the permanent of a generic complex-valued matrix. Indeed, as shown in
eq. (2.18), the probability P(r → s, U) of an input r evolving into s, is proportional to
the permanent of the matrix U[R|S] (recalling definition 3):
P(r → s, U) = |A(r → s, U)|2
=
1
µ(R)µ(S)
|perm(U[R|S])|2
. (3.4)
Computing the permanent of a n×n matrix with the fastest known classical algorithms
[41, 68] requires a number of operations of the order O(n2n
). This means that, for
example, computing the permanent of a 30×30 complex matrix, corresponding to a
single scattering amplitude for a 30-photon state, requires a number of operations of the
order of ∼ 1010
. If the average time required by a classical computer to perform a single
operation is of the order ∼ 107
, the computation of one such scattering amplitude
will require ∼ 10 minutes. While still clearly manageable by a classical computer,
this already shows the potential advantages of a boson sampling apparatus: if the
experimental problems related to coherently evolve 30 indistinguishable photons inside
an interferometer were to be solved, this would allow to sample from the probability
distribution given by eq. (3.4) without actually knowing the probabilities itselves.
AA demonstrated that, should boson sampling be classically easy to solve, this would
have very strong and undesied consequences in computational complexity theory, and
therefore it is most probable that boson sampling is not classically easy to solve.
It is worth stressing that the BosonSampling problem is not that of finding the per-
manents in eq. (2.18), but only that of sampling from the related probability distribution.
In fact, not even a boson sampler is able to efficiently compute these scattering probabil-
ities. This is due to the fact that, to reconstruct a probability distribution spanning over
a number M of events, roughly speaking, the number of samples is required to be at
least of the order of M. But, as shown in eqs. (2.22) and (2.23), M scales exponentially
with n, implying that the number of experimental samples required to reconstruct the
probability distribution becomes exceedingly large very soon. In figs. 3.3 and 3.4 is
shown that, if the number of samples is not large enough, the reconstructed probability
distribution is different from the real one. In fact, generally speaking, there are strong
arguments against the possibility to compute the permanents of complex-valued matri-
ces by means of quantum experiments [69], although attempts have been reported in
this direction [70].
49
0
0.025
0.05
0.075
0.1
0.125
0.15
States
Probabilitydistribution
0.0
0.5
1.0
1.5
2.0
States
Numberofsamples
0
5
10
15
States
Numberofsamples
0
50
100
150
States
Numberofsamples
a) b)
c) d)
Figure 3.3: Example of boson sampling from a randomly chosen 8×8 unitary matrix. (a) Exact
output probability distribution for two photons injected in the first two modes of a random 8-mode
interferometer. (b), (c), (d) Output states sampled from the output probability distribution, for a number
of samples equal to 10 (b), 100 (c) and 1000 (d). As seen, with a low number of samples, the histogram
may appear different from the true probability distribution.
0
0.01
0.02
0.03
States
Probabilitydistribution
0.0
0.2
0.4
0.6
0.8
1.0
States
Numberofsamples
0
1
2
3
4
5
States
Numberofsamples
0
5
10
15
20
25
30
States
Numberofsamples
a) b)
c) d)
Figure 3.4: Example of boson sampling from a randomly chosen 8×8 unitary matrix. (a) Exact
output probability distribution for four photons injected in the first four modes of a random 8-mode
interferometer. (b), (c), (d) Output states sampled from the output probability distribution, for a number
of samples equal to 10 (b), 100 (c) and 1000 (d). The number of samples required to reliably recover the
original probability distribution is much higher than in fig. 3.3, due to the higher number of many-boson
states, which are here 8+3
4 = 330 against the 8+1
2 = 36 of fig. 3.3.
50
The complexity of BosonSampling makes it an extremely valuable candidate to gain
experimental evidences of the supremacy of quantum devices over classical computers.
Indeed, it presents several advantages in this regard over, for example, Factoring,
which is the paradigmatic problem that would allow quantum computers to perform a
post-classical computation:
1. The BosonSampling problem is even harder than Factoring, being related to
the #P-hard complexity class, and believed to not be in NP.
2. A boson sampler requires significantly less resources to be implemented than a
universal quantum computer. In particular, it does not require adaptive or feed-
forward mechanisms, nor fault-tolerance methods. This relatively simple design
has already prompted a number of small-scale implementations of increasing
complexity [9, 10, 11, 12, 13, 14, 15, 17, 18, 25, 71].
AA suggested [8] that a 400-modes interferometer fed with 20 single photons
is already at the boundary of the simulation powers of present-day classical
computers. While in this regime it would still be possible to carry out a classical
simulation, the quantum device should be able to perform the sampling task faster
than the classical computer.
3. The theoretical evidence of the hardness of BosonSampling is stronger than that
of factoring integers: while in the former case the result only relies on a small
number of conjectures regarding the hardness of some problems [8], in the latter
case there is no compelling evidence for Factoring to not be in P. While known to
be in BQP, Factoring is only believed to be in NP, and strongly believed to not
be in NP-hard.
While the hardness of Factoring is strong enough to build modern cryptography,
it could also happen that a polynomial-time algorithm will be discovered showing
that Factoring is in P as, basically, the sole evidence for its hardness is the fact
that no efficient classical algorithm is yet known.
3.3 Scaling experimental boson sampling implemen-
tations
The hardness of BosonSampling has another potentially important consequence: it
could provide the first experimental evidence against the ECT [8]. This point, however,
is still subject to some debate [67, 72, 73], due to the somewhat informal nature of
the ECT itself. Indeed, the ECT is not a mathematical statement, but a statement about
how the physical world behaves, in a certain asymptotic limit. Because of this it is
51
somewhat ill-defined what exactly would be “enough experimental evidence” to “prove”
or “disprove” such a statement.
While AA argued for the hardness of both the exact and approximate BosonSam-
pling problems, they did not take into account other forms of experimental imperfec-
tions.
More precisely, AA showed that even attempting to estimate the output probability
distribution of a boson-sampler is likely computationally hard, as long as the probability
P of each input state being correctly produced by the sources scales as P > 1/ poly(n),
that is, does not vanish faster than the inverse of a polynomial in n [8]. Further evidence
that even lossy systems or systems with mode mismatch are likely to be classically hard
to solve was later given [74].
Another source of errors which could potentially undermine the scalability of boson
sampling are the unavoidable circuit imperfections, especially taking into account
the fact that a boson sampler cannot likely implement fault-tolerant mechanisms [8].
Therefore, small errors in the calibrations of the elementary optical elements, like
beam splitters and phase-shifters, may accumulate and result in an output probability
distribution significantly different from the ideal one. A thorough analysis of this issue,
reported in [75], leads to the result that as long as the average fidelity of the elementary
gates scales at least like 1 − O(1/n2
), the overall implementation provides a reasonably
good output distribution.
Generally speaking, all of these works seem to reinforce the idea that the Boson-
Sampling problem is scalable even when experimentally plausible conditions are taken
into account. This however, is not definitively enstablished. While the potential for
scaling up the number of modes of network implementations has been fairly well
enstablished, and experiments with 2 photons into up to 21 modes have been reported
[76], scaling the number of photons is much harder and will likely require technologi-
cally improvements, especially on the side of the employed single-photon sources and
detectors.
In particular, most reported boson sampling experiments used SPDC single-photon
sources. With this type of scheme, using a single SPDC source, the probability of
producing n photons decreases exponentially with n, as shown in eq. (2.40). This
means that in order to implement a boson sampler with higher numbers of photons
another kind of architecture is most likely required. One such example is provided by a
scattershot boson sampler, which will be described in the next section.
3.4 Scattershot boson sampling
One of the main difficulties in scaling up the complexity of boson sampling devices is
the requirement of a reliable source of many indistinguishable photons. Indeed, despite
recent advances in photon generation [42] using atoms, molecules, color centers in
52
diamond, and quantum dots, currently the most widely used method remains parametric
downconversion. The drawbacks of SPDC sources, described in section 2.3.4, have
however restricted SPDC implementations of boson sampling experiments to proof-of-
principle demonstrations.
Recently a new scheme, named scattershot boson sampling [19], has been proposed
to make the best use of SPDC sources for photonic boson sampling, greatly enchancing
the rate of n-photon events [19, 20]. As opposite to a boson sampling experiment, in
which a single fixed input state is employed, in a scattershot boson sampler the photons
are injected into randomly chosen modes at each shot. This is achieved adopting
an heralded single-photon source for every mode of the interferometer (see fig. 3.5).
If a single source has a probability p of injecting a photon into the interferometer
for each laser pulse, the probability of n particular sources injecting a photon at the
same time is pn
. However, there are m
n
possible combinations of sources that can
simultaneously inject a photon into the interferometer, so that the resulting probability
of injection of any configuration of n photons is m
n
pn
. For m n2
, which is the
regime in which BosonSampling is hard, the binomial factors dominate over pn
. Thus,
while the efficiency of using a single SPDC source to generate many photons decreses
exponentially with pn
, the overall efficiency of a scattershot boson sampling apparatus
with m SPDC sources increases exponentially with n, in the regime n m, as shown in
fig. 3.6.
While a scattershot boson sampling apparatus would technically solve a problem
different from BosonSampling as defined above, due to the randomly changing input
state, it has been shown [19, 20] that the computational problem solved by a scattershot
boson sampler is at least as hard as BosonSampling, thus making this new scheme as
good as the original one to achieve a post-classical computation.
An experimental implementation of a scattershot boson sampler has been recently
reported [18], using a 13-mode integrated photonic chip and up to six SPDC sources, to
obtain data corresponding to two- and three-photon interference.
53
Figure 3.5: A scattershot boson sampler. An SPDC source is used for every input of the m-mode
interferometer. Each SPDC source generates (ignoring higher order terms) a pair of photons, one of
which is injected into the interferometer, while the other heralds the generation event. Every time a
detector corresponding to an heralding photon “clicks”, we know that a photon has been injected into
the corresponding mode of the interferometer. (image courtesy of [18])
m =8
m =16
m =32
m =64
0 10 20 30 40 50 60
n
100
106
1010
1014
1018
p =0.5
p =0.1
p =0.01
5 10 15 20 25 30
n
1010
1020
1030
1040
1050
11
a) b)
Figure 3.6: (a) Logarithmic plot of the binomial factor m
n as a function of n, for m equal to 8 (blue), 16
(orange), 32 (green) and 64 (red). (b) Logarithmic plot of the probability pn n2
n of injecting n photons
into an m = n2
-modes interferometer with a scattershot boson sampling apparatus, against the number
of photons n, with the single pair generation probability p equal to 0.01 (blue), 0.1 (orange) and 0.5
(green). The plot shows how the probability increases exponentially with n. Moreover, the lower is the
value of p, the higher is the number of photons required to see this advantage, as seen here in the case of
p = 0.01. Credits: [18].
54
Chapter 4
Boson sampling validation
In this chapter we discuss the problem of validating boson sampling experiments.
In section 4.1 the problem is presented in its generality. In section 4.2 we focus on
the validation protocols based on zero-transmission laws. A validation scheme with
Fourier matrices is described, an experimental implementation of which is given in
chapter 6. In section 4.3 we discuss the possibility of validating scattershot boson
sampling experiments with the above described validation scheme for Fourier matrices.
In section 4.4 an alternative validation scheme, based on Sylvester matrices, is presented.
The advantages of this protocol to validate scattershot boson sampling experiments is
discussed.
4.1 Validating boson sampling implementations
The very hardness of BosonSampling, while being its main feature, also poses the
problem of certifying the results of an experimental implementation. This means that
if someone were to present an alleged boson sampling device as a black box, we may
not be able to certify its output. In other words, we may not be able to tell if the output
of the black box is the result of true many-body interference, as opposite, say, of that
of a classical computer giving in output a series of randomly chosen samples. From
a complexity-theoretical point of view, this is due to the fact that BosonSampling
belongs to the #P class [8], and thus not classically verifiable in polynomial time.
More intuitively, we can trace back the causes of this certification problem to two
fundamental factors:
Hardness of computing the bosonic probability distribution: It is not possible
to efficiently calculate with classical resources, and probably neither with quantum
ones [8], the bosonic output probability distribution. While this may naively be thought
of as the main issue, it does not imply per se a fundamental certification problem. As a
counterexample we may think of Factoring: while it is not (as to the current state of
55
knowledge) possible to efficiently find the prime factorization of an integer number, it is
possible to efficiently verify whether a proposed factorization is the right one, simply by
multiplying the numbers together. No such mechanism is, however, currently thought
to exist for BosonSampling.
Hardness of discriminating two probability distributions: Another issue, ar-
guably even more important than the latter, arises from the exponential number of
possible outputs which the bosonic probability distribution covers, and the sampling
nature of BosonSampling. Indeed, even completely disregarding the hardness of com-
puting permanents and scattering amplitudes, the sheer number of samples required
to meaningfully compare two probability distributions extending over an exponential
number of events, can be very problematic (see the simple examples in figs. 3.3 and 3.4).
The boson sampling validation problem can, more precisely, be stated as the problem
of efficiently ruling out that a set of data, produced by the alleged boson sampler, is not
the result of some other probability distribution which may be implemented into the
device.
It is worth noting that one has to specify which alternative probability distributions
one wants to be able to distinguish from the boson sampling one. Indeed, generally
speaking, the more a probability distribution is close to the boson sampling one, the
higher the number of required samples to discriminate them will be, as shown in a
simple case in example 7.
Example 7 (Discriminating between similar probability distributions). Consider two
probability distributions, p1 and p2, defined over the two events 0 and 1 as
p1(x = 0) = 1/2, p1(x = 1) = 1/2, (4.1a)
p2(x = 0) = 1/2 + , p2(x = 1) = 1/2 − . (4.1b)
The average number of samples required to distinguish between p1 and p2, with
confidence α, is given by
Nmin =
log α
x∈events
p1(x) log
p1(x)
p2(x)
. (4.2)
In other words, Nmin is the average length of a dataset D such that the
P(p1 | D)
P(p2 | D)
> α, (4.3)
where P(p | D) is the probability that the observed set of samples D has been gener-
56
ated by the probability distribution p. Applying eq. (4.2) to eq. (4.1) we obtain
Nmin ≈
log α
2 2
+ O(1), for → 0. (4.4)
Thus as expected, decreasing , and thus increasing the “similarity” of the two distri-
butions, the number of required samples increases indefinitely.
From the experimental point of view, this poses the very concrete issue of finding
a way to certify that a built experimental apparatus is working as aspected. From a
more fundamental standpoint instead, an unsolved validation problem threatens to
crucially undermine the theoretical claim of observing a post-classical computation:
if we are given by an untrusted party an alleged boson sampler as a black box, which
efficiently produces a number of samples faster than any classical computer can sample
from the boson sampling distribution, how can we somehow certify that that output is
what it claims to be? Indeed, an issue may arise if the number of samples necessary
to assess that the produced set of data does not come from an alternative probability
distribution increases exponentially with n, thus making an experimental, verifiable
implementation of the BosonSampling problem not scalable.
When the boson sampling validation problem was first raised, it was argued to
be fundamentally unsolvable, thus undermining the possibility of experimentally dis-
criminate a boson sampling distribution even against the trivial uniform one [77].
AA subsequently showed that this was not true, providing an example of an efficient
method to discriminate the boson sampling distribution from the uniform one [21].
This test, however, was not capable of efficiently separating, for example, the output
of a boson sampler from that resulting from the evolution of classical particles into
the interferometer. This validation test was successfully implemented in a number of
experiments [14, 15].
Other tests, able to validate against distinguishable particles, were later proposed
[14, 15, 16, 21, 22, 23] and experimentally implemented [14, 15, 17].
4.2 Fourier suppression law
Among the proposed validation schemes, those based on Zero-Transmission Laws (ZTLs)
[24] (also often referred to as suppression laws) seem to be a promising approach to
discriminate true many-boson evolution from alternative models. A ZTL can be thought
of as a generalization of the HOM effect to many modes and photons. More precisely,
it refers to the capability, occurring in some particular unitary matrices, to efficiently
predict the strict suppression of certain input-output combinations of states. In other
words, a ZTL allows to efficiently predict that for certain pairs of states (r, s), the
scattering amplitude A(r, s) vanishes, without having to compute a permanent.
57
Such suppressions, like the HOM effect, are a pure quantum-mechanical effect, arising
from the interference of many-body states during the evolution in the interferometer,
and as such a strong signature of non-classical evolution.
In the originally proposed BosonSampling scheme, where a random unitary matrix
is employed, no such ZTL can be taken advantage of. However, these turn out to be
useful when thinking to the validation problem in a slightly broader perspective. If
we do not think of a boson sampler as a black box, but approach the more practical
problem of assessing that an actual experimental apparatus is working as advertised,
we can analyse the various possible sources of error. Roughly speaking, the two main
factors potentially leading to an erroneous output probability distribution are: a wrong
implemented unitary evolution, and non-indistinguishable input states. While the
unitary matrix implemented in an interferometer can be efficiently verified [78], the
indistinguishability of the input photons is harder to assess, and it is in this context
that ZTLs can be exploited.
Provided that one can change at will the unitary matrix implemented in the device -
which is in principle possible to do efficiently [79] and has been recently reported in an
experimental setup [17] - then it is possible to first verify the indistinguishability of
the sources by injecting the photons in a specially crafted unitary, whose symmetries
allow to efficiently observe a ZTL, and then on the same device to implement a random
unitary matrix and perform the boson sampling experiment.
The first proposed example of such a ZTL [24] employed Fourier matrices:
Definition 4 (Fourier matrices). Let UF
(m) be the m-dimensional matrix defined
element-wise by
[UF
(m)]j,k ≡
1
√
m
exp
2πi
m
(j − 1)(k − 1) , j, k = 1, . . . , m. (4.5)
We will refer in the following to this matrix as the m×m Fourier matrix. The m
dependence will be omitted when clear from the context.
It was shown [22, 24, 65, 66] that with Fourier matrices, a b-periodic input state r
never evolves in an output state s such that b n
j=1 Sj is not a multiple of the number
of modes m, where S is the MAL introduced in definition 1, and a b-periodic state is
defined as follows:
Definition 5 (Periodic states). Given a state r, we say that r is b-periodic if its MOL
representation has the form
r ≡ (r1, . . . , rb, r1, . . . , rb, . . . , r1, . . . , rb)
m=p·b
, (4.6)
58
√
2 UF
(2) =
1 1
1 −1
√
3 UF
(3) =




1 1 1
1 e
2πi
3 e
−2πi
3
1 e
−2πi
3 e
2πi
3




√
4 UF
(4) =





1 1 1 1
1 i −1 −i
1 −1 1 −1
1 −i −1 i





Table 4.1: Examples of Fourier matrices
where b is the length of each of the identical sublist (bins) composing r, and p = m/b
the number of such sublists. We will refer to b as the periodicity of r.
Being UF
symmetric, that is, UF
i,j = UF
j,i for every i and j, the scattering amplitudes
are also symmetric with respect to input and output states: A(r → s, UF
) = A(s →
r, UF
). In conclusion, the Fourier suppression law [24] can be stated as follows:
Theorem 1 (Fourier suppression law). Let UF
(m) be the m×m Fourier matrix, and
let r and s be two n-boson states, with corresponding MAL representations R and S.
Then, if r is b-periodic and s is such that
mod

b
n
j=1
Rj, m

 = 0, (4.7)
then the following suppression law holds:
A(r → s, UF
) = A(s → r, UF
) = 0. (4.8)
We note that the converse in in general not true. In other words, this condition is
sufficient but not necessary for an input-output pair to be suppressed.
The many-boson matrices corresponding to Fourier matrices with m = 4, 6, 8, and
n = 2, are shown in figs. 4.1 to 4.3. In these figures, the periodic states are highlighted in
green, and the strictly suppressed states are the white squares. The input-output pairs
predicted to be suppressed by theorem 1 are marked by purple circles. The majority
of the suppressed pairs correspond to periodic states, and are predicted by theorem 1.
Figure 4.3 shows however that theorem 1 gives a sufficient but not necessary condition
for a pair to be suppressed, as there are in this case pairs which are suppressed but do
not correspond to periodic states.
59
Fourier matrix Random matrix
a) b)
Figure 4.1: Many-boson matrices for m = 4 and n = 2. Modulus squared of the elements of the
many-boson matrices U(4, 2) corresponding to the 4×4 Fourier matrix and a randomly chosen 4×4
unitary matrix. (a) Periodic states are highlighted in green, and input-output pairs predicted to be
suppressed by theorem 1 are marked by purple circles. All the suppressed pairs are here predicted to
be suppressed. In other words, in this case, the Fourier suppression law completely characterizes the
suppressed input-output pairs. (b) Set of scattering probabilities resulting from a randomly chosen
unitary matrix. No evident structure arises in this case. The calculation of the elements of the matrix
requires the computation of the corresponding permanents, as no efficient method is known to calculate
many-body scattering amplitudes for random matrices. Probabilities lesser than 10−2
are shown as “≈ 0”
for the sake of clarity.
60
Figure 4.2: Modulus squared of the elements of the many-boson matrix U(6, 2, UF
), corresponding to
the 6×6 Fourier matrix injected with 2-photon states. Strictly suppressed input-output pairs are marked
with purple circles, and periodic states highlighted in green. Similarly to fig. 4.1, all of the suppressed
pairs are covered by theorem 1, which is shown here by the fact that the white elements of the matrix
are always marked with purple circles.
61
Figure 4.3: Modulus squared of the elements of the many-boson matrix U(8, 2, UF
), corresponding to
the 8×8 Fourier matrix injected with 2-photon states. Strictly suppressed input-output pairs are marked
with purple circles, and periodic states highlighted in green. Differently from figs. 4.1 and 4.2, not all of
the suppressed pairs are in this case predicted by theorem 1. As an example, there a number of output
suppressed pairs corresponding to the input state (6, 8), which has not any periodicity. This is possible
because theorem 1 gives only a sufficient, but not necessary condition for a pair to be suppressed.
62
The validation scheme which exploits the Fourier suppression law, proposed in [22],
uses a single fixed input state with MAL
RTichy = (1, np−1
+ 1, 2np−1
+ 1, . . . , (n − 1)np−1
+ 1), (4.9)
injected into an inteferometer implementing the m-dimensional Fourier matrix, with
m = np
and p some positive integer. RTichy is an especially crafted n-periodic state,
corresponding to an high fraction of output suppressed states. Counting the number
of states corresponding to a suppressed input-output pair - that is, an input-output
pair that would be suppressed if the evolution was an ideal quantum-mechanical one - it
is possible to quantify how much the observed output probability distribution differs
from the ideal boson sampling one.
More precisely, the violation parameter is defined as
D = Nforbidden/Nruns, (4.10)
where Nforbidden is the number of observed events violating theorem 1, and Nruns is the
total number of events. This parameter has been shown [22] to be a better signature
of quantum many-body statistics than precedently proposed validation schemes. In
particular, it can be used to validate a boson sampler against alternative hypotheses
such as the Mean-Field Sampler [22]. A Mean-Field Sampler is an efficiently evaluable
and physically plausible alternative model [22], which reproduces some aspects of
many-boson dynamics surviving in the semiclassical limit. This model is not distin-
guished from the boson sampler distribution by other validation schemes, making the
Fourier validation test more capable of capturing the signatures of genuine many-boson
quantum interference. A summary of the validation protocol based on Fourier matrices
is given in fig. 4.4.
63
Figure 4.4: Validation scheme with Fourier matrices. Conceptual scheme of the protocol: the
possible configurations of n photons at the output of an m-mode interferometer can be divided into
two categories, unforbidden and forbidden, depending on whether they satisfy or not the suppression
confition of theorem 1, respectively. The pie charts show the expected output statistics with different
classes of particles, where green and red areas represent events with unforbidden and forbidden outputs,
respectively. The injection of a cyclic Fock state (beige box) in an m-mode Fourier interferometer results
in total suppression of forbidden output states. Cyclic states with distinguishable particles (blue box) show
no suppression, being each output combination equally likely to occur. A Mean Field state [22] (purple
box), a semi-classical state reproducing some of the features of bosonic statistics, shows suppression
with highly reduced contrast. Therefore, with a cyclic input the m-mode Fourier interferometer is able
to discriminate, through the measurement of degree of violation D = Nforbidden/Nevents, which of these
three hypotheses the input state belongs to. Credits: [25].
64
4.3 Fourier suppression law and scattershot boson sam-
pling
The original scheme for implementing a boson sampler using a single SPDC source is not
scalable to higher number of photons, as discussed in the earlier sections. Scattershot
boson sampling is, currently, the most promising architecture to implement Boson-
Sampling with higher number of photons. This scheme, however, is essentially based
on the injection of random input states into the interferometer. On the other hand,
the validation scheme based on the Fourier suppression law only works for particular,
periodic input states. Therefore, a scattershot boson sampling experiment cannot be
validated making use of the Fourier suppression law, as originally proposed.
For a validation scheme based on some ZTL to be employable with a scattershot
boson sampler, it is no more sufficient to have an high number of output suppressed
states corresponding to some small class of input states. Instead, it is required that the
fraction of pairs of suppressed input-output states must be a non-negligible fraction
of the total number of states. In other words, considering that all input states in a
scattershot boson sampler are collision-free, and approximating all output states to also
be collision-free (which is justified in the m n limit [80]), we must have
suppressed input-output configurations
m
n
2 > 0 by a statistically significant amount.
(4.11)
In this section we consider whether this condition is verified for the Fourier sup-
pression law, thus making it viable to use in scattershot boson sampling experiments.
We will proceed by counting how many input states have some kind of periodicity,
according to eq. (4.6), and are thus such that theorem 1 can be applied to them. For the
same reasons given above, we will restrict our analysis to the collision-free states.
To count the number N(m, n) of periodic collision-free configurations (periodic
CFCs) of n photons in m modes, we first need to know for which values of b there can
be b-periodic configurations, given m and n. In other words, we want all the integers b
such that the MOL corresponding to the state can be written in the form (4.6).
Given that a partition of the MOL of the form (4.6) can be found for each p = 1
which is a common divisor of m and n, we have that the possible b are all and only the
integers of the form b = m/p for such a p. Formally, we say that the set B(m, n) of all
possible bin lengths b for n photons in m modes is
B(m, n) = {b ∈ N | b = m/p for some p = 1 common divisor of m and n}. (4.12)
Denoting with S(m, n, b) the set of all m-modes, n-photons, b-periodic CFCs, we
65
can now see that
N(m, n) =
b∈B(m,n)
S(m, n, b) . (4.13)
The number of CFCs in S(m, n, b) is equal to the number of lists of length m
composed of only 0 and 1 that are decomposable in p = m/b identical sublists of length
b each. This is in turn equal to the number of ways in which we can choose n/p = nb/m
elements from a set of b elements, i.e. the binomial factor b
nb/m
, and we conclude that
|S(m, n, b)| =
b
nb/m
. (4.14)
A naive approach would be now to say that N(m, n) is equal to the sum of
|S(m, n, b)| for each b ∈ B(m, n). This however overcounts all those states having
multiple periodicities. For example in the case m = 8, n = 4, where B(m, n) = {2, 4},
the states
q = (0, 1, 0, 1, 0, 1, 0, 1) and q = (1, 0, 1, 0, 1, 0, 1, 0) (4.15)
are at the same time 2- and 4-periodic, and would each be counted twice if we were to
say that N(8, 4) = |S(8, 4, 2)| + |S(8, 4, 4)| = 8.
The issue can be traced back to the fact that S(8, 4, 2) ∩ S(8, 4, 4) = ∅. More
generally, the intersection S(m, n, b1) ∩ S(m, n, b2) will be non-empty whenever there
is some integer b0 ∈ B(m, n) common divisor of b1 and b2. The problem of counting
the periodic CFCs is thus reconducted to that of finding the number of elements in the
union of a set of generally not-disjoint subsets, resulting in the general formula
N(m, n) =
∞
k=1
(−1)k+1
(b1,...,bk)
|S1 ∩ · · · ∩ Sk| , (4.16)
where we denoted for brevity Si ≡ S(m, n, bi), and the inner sum is over all unordered
lists of k elements of B(m, n). With reasonings similar to the above ones it is possible
to show that
|S(m, n, b1) ∩ · · · ∩ S(m, n, bk)| = |S(m, n, GCD(b1, . . . , bk))| , (4.17)
where GCD(b1, . . . , bk) denotes the greatest common divisor of b1, . . . , bk ∈ B(m, n).
Equation (4.16), together with (4.17) and (4.14), provides us with a general formula to
compute the exact number of periodic CFCs of given m and n.
As an example of the application of this formula we again take the case m = 8, n = 4,
which now yields the correct result:
N(8, 4) = |S(8, 4, 2)| + |S(8, 4, 4)| − |S(8, 4, 2) ∩ S(8, 4, 4)|
= |S(8, 4, 2)| + |S(8, 4, 4)| − |S(8, 4, 2)| = |S(8, 4, 4)| =
4
2
= 6.
66
While exact, expression (4.16) takes a lot of computational time due to the exponen-
tially growing number of required permutations. In practice, already the computation
of the number of periodic CFCs with m ≈ 120 and n ≈ 60 is very time-consuming.
Computational times can be significantly reduced noting that
b1| b2 =⇒ S(m, n, b1) ⊆ S(m, n, b2) (4.18)
where the vertical bar indicates that b1 divides b2. We can therefore apply eq. (4.16)
with the slight modification of taking bi ∈ ˜B(m, n) instead of bi ∈ B(m, n), where
˜B(m, n) is the subset of elements of B(m, n) which do not divide any other element of
B(m, n), i.e.
˜B(m, n) ≡ {b ∈ B(m, n) : ( b ∈ B(m, n) : b|b )}. (4.19)
Using (4.16) together with (4.19) we can easily compute the number of periodic CFCs
up to and beyond m ≈ 1000. Faster exact results can be achieved in many particular
cases. For example, for m = 2p
for some integer p, and n an even number, reasonings
similar to the above ones lead to (see fig. 4.5)
N(m, n) =
m/2
n/2
. (4.20)
Figure 4.5: Number of periodic states vs number of photons n, for various values of the number of
modes m. The orange line is m/2
n/2
, which exactly reproduces the number of periodic configurations for
all even n as predicted by (4.20).
More generally, for m 1 the number of periodic states is dominated by m/a
n/a
with a = m/ GCD(m, n) (see figs. 4.6 and 4.7).
67
1000 2000 3000 4000
n
200
400
600
800
1000
1200
1400
Figure 4.6: Logarithm of the number of periodic states
of m/2 photons in m modes, plotted as a function of m.
The coloured dashed lines are, in descending slope order,
log m/2
m/4
, log m/3
m/6
, log m/5
m/10
, log m/7
m/14
.
5 10 15 20 25 30 35
n
5
10
15
20
25
30
35
10 20 30 40
n
1000
2000
3000
4000
5000
6000
7000
10 20 30 40 50
n
1×106
2×106
3×106
4×106
5×106
Figure 4.7: Number of periodic states vs number of photons. Blue points: number of periodic
states calculated with eq. (4.16). The continuous green, cyan, purple and red lines are respectively m/2
n/2
,
m/3
n/3
, m/5
n/5
and m/7
n/7
.
Unlike the number of periodic collision-free configurations, an efficient algorithm to
compute the number of suppressed output states predicted by Tichy’s suppression law
for Fourier matrices has not been found yet, except for very particular input states [22].
A brute force computation of the number of suppressed output states for each input
state for various values of m and n is shown in figs. 4.8 and 4.9. These computations
show that only a small fraction of input states results in a significant fraction of output
suppressed states.
However, the above result regarding the number of periodic states is enough to
make the Fourier suppression law unsuitable to validate scattershot boson sampling
experiments. In fact, in the best case scenario in which n = 2p
for some p, the number
of periodic states is ≈ m/2
n/2
(see eq. (4.20)), which is negligibly small compared to m
n
.
In particular, for m = n2
, the fraction n2/2
n/2
/ n2
n
decreases exponentially with n. This
implies that, even if all of these input states were to correspond to an 100% fraction of
output suppressed states, which is far from true, nevertheless in the regime of high n
virtually no suppressed pair would ever be sampled in a actual scattershot experiment,
and the validation scheme would be inefficient.
68
5 10 15 20 25
0
20
40
60
80
Inputs
%ofsuppressedstates
8 modes, 2 photons
0 20 40 60 80 100 120
0
20
40
60
80
Inputs
%ofsuppressedstates
16 modes, 2 photons
0 10 20 30 40 50 60 70
0
20
40
60
80
Inputs
%ofsuppressedstates
8 modes, 4 photons
0 500 1000 1500
0
20
40
60
80
Inputs
%ofsuppressedstates
16 modes, 4 photons
5 10 15 20 25
0
20
40
60
80
Inputs
%ofsuppressedstates
8 modes, 6 photons
0 2000 4000 6000 8000
0
20
40
60
80
Inputs
%ofsuppressedstates
16 modes, 6 photons
a) b)
c) d)
e) f)
Figure 4.8: Ratio of suppressed collision-free output states For each input collision-free state, the
percentage of output suppressed collision-free states is shown for Fourier matrices (blue dots) and
Sylvester matrices (red dots). All the points are calculated using eq. (3.4). In all the cases shown the
number of suppressed output states is significantly higher for the Sylvester matrix. (a) and (b): For
two input photons, all of the output suppressed states are predicted by theorems 1 and 2 in the case of
Fourier and Sylvester matrices, respectively. Green dashed line: exact fraction of suppressions predicted
by eq. (4.24). (c) and (d): For n = 2p
only a subset of suppressed pairs are predicted by theorems 1
and 2. Green dashed line: estimate of the fraction of output suppression states, equal to 1 − 1/n. The
approximation gets closer to the true value for large values of n and m. (e) and (f): For n not equal to a
power of 2, no estimate of the number of output suppressed states is given, neither in [22] nor in [23].
69
0 1000 2000 3000 4000
0
20
40
60
80
Inputs
%ofsuppressedstates
16 modes, 5 photons
0 2000 4000 6000 8000 10 000 12 000
0
20
40
60
80
Inputs
%ofsuppressedstates
16 modes, 8 photons
a) b)
Figure 4.9: Ratio of suppressed collision-free output states for each input in a 16-mode interferometer
injected with 5 (a) and 8 (b) photons. Like fig. 4.8, the ratio of suppressed output states is significantly
higher for almost all input states for the Sylvester matrix than the Fourier matrix.
4.4 Sylvester suppression law
Besides the Fourier matrix, a ZTL has been shown to exist also for another important
class of unitary matrices [23], the so-called Sylvester matrices.
Definition 6 (Sylvester matrices). Let US
be an m-dimensional unitary matrix of
the form US
(m) ≡ H(m)/
√
m, with H(m) defined recursively as
H(2p
) =
H(2p−1
) H(2p−1
)
H(2p−1
) −H(2p−1
)
, (4.21)
for each positive integer p, and with H(1) = 1. We will refer in the following
to US
(m) as normalized Sylvester matrix and to H(m) as Sylvester matrix. The m
dependence of US
and H will be omitted when clear from the context.
An analytic expression for the element (i, j) of a Sylvester matrix can be given in
the form:
[H(2p
)]i,j = (−1)iB jB
, (4.22)
where iB and jB are the binary representations of i and j, with rows and columns
enumerated from zero to m − 1, and is the bitwise dot product. For example, for
m = 8, i = 2, j = 7 we would have
[H(8)]2,7 = (−1)(0,1,0)·(1,1,0)
= (−1)0×1+1×1+0×0
= −1. (4.23)
70
H(2) =
1 1
1 −1
H(4) =





1 1 1 1
1 −1 1 −1
1 1 −1 −1
1 −1 −1 1





H(8) =








1 1 1 1 1 1 1 1
1 −1 1 −1 1 −1 1 −1
1 1 −1 −1 1 1 −1 −1
1 −1 −1 1 1 −1 −1 1
1 1 1 1 −1 −1 −1 −1
1 −1 1 −1 −1 1 −1 1
1 1 −1 −1 −1 −1 1 1
1 −1 −1 1 −1 1 1 −1








Table 4.2: Examples of Sylvester matrices
We will refer in the following to US
(m) as normalized Sylvester matrix and to H(m) as
Sylvester matrix. The m dependence of US
and H will be omitted when clear from the
context.
It was recently proved [23] that injecting two bosons into any pair of (different)
modes of a multi-mode interferometer implementing a Sylvester matrix, the total
fraction of suppressed output states is
Nsupp
Ntot
= m
2(m+1)
. Considering only collision-free
states, the fraction becomes
Nsupp
Ntot,CF
=
m
2(m − 1)
. (4.24)
More generally, the following suppression law holds [23]:
Theorem 2 (Sylvester suppression law). Given an m×m Sylvester matrix, m = 2p
,
n = 2q
, q < p, and an n-boson input state with MAL S of the form
S = (1 + nc, . . . , n + nc), (4.25)
for some integer c such that 0 ≤ c ≤ (m/n − 1), then the strictly suppressed output
states are all and only those R ≡ (R1, R2, . . . , Rn) such that
R1,B ⊕ · · · ⊕ Rn,B = 0 ≡ (0, . . . , 0
q
), (4.26)
where Ri,B is the q-dimensional vector obtained from the binary representation of
Ri − 1 truncated to the q least significant digits, and ⊕ is the bitwise XOR operation
(equivalently the elementwise, modulo 2, sum).
Similarly to the case of Fourier matrices, an estimate of the fraction of strictly
suppressed states predicted by this law is:
V =
Nsupp
Ntot
≈ 1 −
1
n
. (4.27)
71
Theorem 2 predicts an high ratio of suppressed output configurations for all input states,
when n = 2, as also shown in the numerical computations in figs. 4.8 and 4.9, and in the
many-boson matrices shown in figs. 4.10 and 4.11 On the other hand, this suppression
law predicts in the general case of n > 2 photons an high ratio of suppressed output
configurations only for the m/n input states of the form (4.25). This is a negligibly small
fraction of the total number of m
n
collision-free configurations, apparently making
Sylvester matrices even worse than Fourier matrices to validate a scattershot boson
sampler.
However, a brute force computation with eq. (2.18) shows (see figs. 4.8 and 4.9)
that the number of suppressed input-output pairs in a Sylvester matrix is significantly
higher than what predicted by the above described suppression law.
As will be showed in the following chapter, it is however possible to efficiently
predict a greater number of suppressed pairs, by generalizing the ZTL proposed in [23].
72
Fourier matrix, m=8, n=2 Sylvester matrix, m=8, n=2
Fourier matrix, m=8, n=4 Sylvester matrix, m=8, n=4
Figure 4.10: Comparison of Fourier and Sylvester many-boson matrices. Matrices of scattering
amplitudes between all combinations of pairs of input-output collision-free states of 2 and 4 photons
into 8 modes, for Sylvester and Fourier matrices. White: suppressed input-output pair. Orange: non-
suppressed input-output pair. The significantly higher fraction of suppressed pairs of Sylvester matrices
with respect to Fourier matrices is clearly evident in these plots.
73
Figure 4.11: Modulus squared of the elements of the many-boson matrix U(8, 2, US
), corresponding
to the 8×8 Sylvester matrix injected with 2-photon states. The structure of this matrix is much more
simple than that shown in figs. 4.2 and 4.3. Moreover, all of its elements can be efficiently predicted with
Crespi’s suppression law, as opposite to the case of the Fourier matrix, in which only some of the events
corresponding to a periodic input or output can. However, when more than two photons are considered,
not all of the elements of the many-boson matrix are efficiently predicted by theorem 2.
74
Chapter 5
Theoretical results
As discussed in section 4.4, the suppression law given in theorem 2 [23] is not suitable
to validate scattershot boson sampling experiments. On the other hand, as shown
in figs. 4.8 and 4.9, the actual fraction of suppressed pairs is significanly higher, and
some kind of suppression law seems to hold for every input state, not just for those
considered by theorem 2.
In this chapter we propose a generalization of theorem 2, which predicts a much
higher fraction of suppressed pairs, and is thus more suitable to validate scattershot
boson sampling experiments. The proposed suppression law is described in section 5.2,
using notations and lemmas given in section 5.1. In section 5.3 we discuss a validation
scheme to distinguish quantum many-boson interference (up to small errors) from the
output resulting from distinguishable photons, by exploiting the proposed suppression
law. In this context we also derive an analytic expression for the mean number of
samples necessary to experimentally distinguish a nearly quantum evolution from the
classical hypothesis. The chapter closes with a summary of the results in section 5.4.
5.1 Preliminaries
For the results presented in this chapter we will introduce some new notations, on top of
those given in definition 1:
Definition 7 (Binary Matrix representation). Let r = (r1, . . . , rm) be the MOL
of a state of n particles in m = 2p
modes, for some positive integer p, and let
R = (R1, . . . , Rn) be the corresponding MAL. We denote with R
(2)
i the binary repre-
sentation of Ri − 1, padded with zeros on the left so as to make it a vector of length
p, and with Ri,α the α-th element of R
(2)
i . This construction allows us to represent
a many-body state r as a binary matrix (BM) R, i.e. as an n×p-dimensional matrix
75
whose elements are either 0 or 1. We will refer to R as the BM representation (or
simply the BM) of the state r.
Remark 1. We note that for indistinguishable particles it is not meaningful to assign
a specific mode to a particle. This means that the order of the elements of the MAL
describing a state of many indistinguishable particles is not relevant. In other words, if
two MALs R and R differ only for the order of their elements, they represent the same
physical state. Similarly, if two binary matrices R and R can be obtained one from the
other with only a reordering of the rows, they represent the same physical state.
Example 8. Consider the case m = 4 and n = 2, and the state with MOL r ≡
(0, 1, 1, 0) and corresponding MAL R = (2, 3). Following the notations introduced
above we will have R1 = (0, 1), R2 = (1, 0), and the BM of r will be the 2×2 matrix
R =
0 1
1 0
.
Example 9. Consider the case m = 8, n = 4, and the state with MOL
s ≡ (1, 1, 1, 0, 0, 0, 0, 1) and corresponding MAL S = (1, 2, 3, 8). Following the no-
tations introduced above we will have S1 = (0, 0, 0), S2 = (0, 0, 1), S3 = (0, 1, 0),
S4 = (1, 1, 1), and the BM representation of r will be the 4×3 matrix
S =





0 0 0
0 0 1
0 1 0
1 1 1





.
Definition 8. Let R be an n×p BM, and let A be a subset of the p columns of R. Then
we denote with NA
(R) the matrix obtained by negating the columns of R specified
in A. In other words, expliciting the indices, NA
(R) is defined as
[NA
(R)]k,α ≡



Rk,α ⊕ 1, for α ∈ A,
Rk,α, for α /∈ A.
(5.1)
where ⊕ is the sum modulo 2.
Similarly, if r is a MOL representing some state of m = 2p
modes, we denote with
NA
(r) the MOL obtained applying to r the following three operations:
• Convert r to the binary matrix R,
76
• Negate the columns of R specified in A, obtaining the binary matrix NA
(R),
• Find the MOL corresponding to the binary matrix NA
(R).
Example 10. Consider the case m = 8, n = 4 and the state with MOL r =
(1, 1, 1, 1, 0, 0, 0, 0), corresponding to the MAL R = (1, 2, 3, 4), and expressed in
the BM representation as
R =





0 0 0
0 0 1
0 1 0
0 1 1





.
Then, following the notations introduced above, we have when A consists of a single
element,
N{1}
(R) =





1 0 0
1 0 1
1 1 0
1 1 1





, N{2}
(R) =





0 1 0
0 1 1
0 0 0
0 0 1





, N{3}
(R) =





0 0 1
0 0 0
0 1 1
0 1 0





.
From this we see that N{2}
(R) and N{3}
(R) are equal to R, up to a permutation
of the rows, and they therefore represent the same physical state. On the other
hand, N{1}
(R) cannot be made equal to R only rearranging its rows, and therefore
represents a different state. In other words, we have:
N{1}
(r) = (0, 0, 0, 0, 1, 1, 1, 1) = r, N{2}
(r) = r, N{3}
(r) = r.
On the other hand, the set A can contain more than a single element. For example:
N{1,3}
(R) =





1 0 1
1 0 0
1 1 1
1 1 0





, N{1,2,3}
(R) =





1 1 1
1 1 0
1 0 1
1 0 0





,
which, for the same reasons, satisfy
N{1,3}
(r) = (0, 0, 0, 0, 1, 1, 1, 1) = r, N{1,2,3}
(r) = (0, 0, 0, 0, 1, 1, 1, 1) = r.
The following results will also be useful in the following:
Lemma 1. For any two integers a and b we have
(−1)a+b
= (−1)a⊕b
. (5.2)
77
More generally, for any finite set of integer numbers {a1, . . . , aN }, we have
(−1)
N
k=1
ak
= (−1)
N
k=1
ak
. (5.3)
Lemma 2. Let Sn be the set of permutations of {1, . . . , n}, and let τ ∈ Sn be a permuta-
tion different from the identity such that τ2
= 1 (i.e. τ(τ(k)) = k for each k ∈ {1, . . . , n}).
Then we can univocally associate to each σ ∈ Sn another (different) permutation στ ≡ τ◦σ,
where ◦ denotes the composition of permutations.
Lemma 3. Let R and NA
be as in definitions 1 and 8, respectively. Then
• NA
(r) = r if and only if NA
(R) = Rτ for some permutation τ ∈ Sn, where with
Rτ we denote the matrix obtained from R permuting the rows according to τ, i.e.
(Rτ )k,α ≡ Rτ(k),α for each k = 1, . . . , n.
• If τ is such that NA
(R) = Rτ , then
1. τ2
= 1,
2. τ(k) = k for each k = 1, . . . , n,
3. for each σ ∈ Sn, NA
(Rσ) = Rτ◦σ,
4. for all columns α ∈ A, Rα
will have an equal number of 1s and 0s,
5. for all columns α /∈ A, Rα
will have an even number of 1s, and an even
number of 0s.
5.2 Main results
Using the lemmas and definitions listed above, we are now ready to state the main
result of the section:
Theorem 3 (Sufficient condition for suppressed pairs). Let r and s be two many-
boson states of n particles in m = 2p
modes, let R and S be the corresponding BM
representations, and let A be a subset of the columns of R. If the following conditions



NA
(r) = r,
n
k=1 α∈A
Skα = 1.
(5.4a)
(5.4b)
are simultaneously satisfied, then A(r, s, US
) = A(s, r, US
) = 0.
The converse does not in general hold. In other words, this condition is sufficient
78
but not necessary for a pair to be suppressed.
Proof. From now on, we will use the abbreviation A ≡ A(r, s, US
), leaving implicit
the dependence on the states and the unitary evolution considered. The transition
amplitude A can be expressed in terms of a permanent of the form
A =
1
√
r1! . . . rm!s1! . . . sm!
perm US
[R|S] , (5.5)
where perm(U) denotes the permanent of U, and the notation US
[R|S] was given in
definition 3.
Using definition 6 and eq. (4.22), we get from eq. (5.5)
A = C
σ∈Sn
n
k=1
(−1)Rσ(k) Sk
= C
σ∈Sn
(−1)ER,S(σ)
, (5.6)
where we denoted with Rσ(k) Sk the bitwise dot product between Rσ(k) and Sk, defined
as Rσ(k) Sk ≡ p
α=1 Rσ(k),αSk,α, C is a constant factor, and we defined the expression
at the exponent of the last expression (see lemma 1) as
ER,S(σ) ≡
n
k=1
Rσ(k) Sk =
n
k=1
p
α=1
Rσ(k),αSk,α. (5.7)
In the following, we will omit for brevity the dependence of the exponent E from the
states and write the above expression as simply E(σ). Note that the actual value of
E(σ) is not important here: the only relevant detail is its parity.
Since we need to evaluate only the cancellation of the scattering amplitude A,
we will ignore the constant factor C and focus on the sum over the permutations in
eq. (5.6). For this sum to vanish, it is necessary and sufficient that for exactly half of
the permutations we have (−1)E(σ)
= 1. In other words, the scattering amplitude A
vanishes if and only if the number of permutations σ such that E(σ) is even, is equal
to the number of permutations σ such that E(σ ) is odd. A necessary and sufficient
condition for this to hold, is that to each permutation σ we can univocally associate
another different permutation σ such that E(σ ) = 1 ⊕ E(σ).
But if eq. (5.4a) holds, then by lemma 3 and lemma 2 we can univocally associate
to each σ the (different) permutation στ ≡ τ ◦ σ, where τ is a permutation such that
NA
(R) = Rτ . Using στ in eq. (5.7) we have
E(στ ) =
n
k=1
p
α=1
Rτ(σ(k)),αSk,α =
n
k=1


α∈A
Rτ(σ(k)),αSk,α ⊕


α/∈A
Rτ(σ(k)),αSk,α



 .
(5.8)
79
Using now the explicit expression for NA
(R) given in definition 8 and the last result of
lemma 3, we have
NA
(Rσ) = Rτ◦σ ⇐⇒



1 ⊕ Rσ(k),α = Rτ(σ(k)),α, α ∈ A,
Rσ(k),α = Rτ(σ(k)),α, α /∈ A.
(5.9)
Inserting these equations in (5.8), we finally obtain
E(στ ) = E(σ) ⊕
n
k=1 α∈A
Sk,α = E(σ) ⊕ 1, (5.10)
where in the last step we exploited eq. (5.4b) for S. Inserting this last result into eq. (5.6)
we conclude that
A = C
σ∈Sn
(−1)ER,S(σ)
= C
σ∈Sn: E(σ) even
(−1)E(σ)
+ (−1)E(στ )
= 0, (5.11)
which proves that the input/output pair (r, s) is suppressed.
Example 11. Let r be the MOL given in example 10. Then NA
(r) = r holds for
A = {2}, A = {3}. and A = {2, 3}. Theorem 3 predicts that all output states s,
which BM representation S has an odd number of 1s in the second or third column,
are suppressed. Moreover, all output states with BM S having a total number of 1s in
the second and third column, are suppressed. More explicitly, all output states s such
that
n
k=1
Sk,2 = 1 or
n
k=1
Sk,3 = 1 or
n
k=1
[Sk,2 ⊕ Sk,3] = 1,
are predicted to be suppressed. For example, the states S = (3, 6, 7, 8), S = (2, 6, 7, 8),
and S = (4, 6, 7, 8), having BM representations





0 1 0
1 0 1
1 1 0
1 1 1





,





0 0 1
1 0 1
1 1 0
1 1 1





, and





0 1 1
1 0 1
1 1 0
1 1 1





,
respectively, are all suppressed.
Remark 2 (Efficiency). To check if theorem 3 applies to a given input-output pair, one
has to verify condition 5.4 for each of the 2p
− 1 = m − 1 possible (non empty) subsets
of the p columns of R and S. Moreover, the number of elementary operations necessary
80
to verify condition 5.4 for a given set of columns increases polynomially in n. In
conclusion, the amount of computational resources required to verify the hypotheses
of theorem 3 increases polynomially in n and m, and the proposed suppression law is
therefore efficiently verifiable. In fig. 5.1 is shown how the computational times scale
polynomially in n with the suppression law, as opposite to exponentially in n with the
brute-force method of computing the permanents.
4
5
6
7
8p
0
5
10
15
20n
0.001
0.01
0.1
1.
3 4 5 6 7 8
p
0
100
200
n
0.0
0.2
0.4
ss
a) b)
Figure 5.1: (a) Computational times required to assess whether a randomly sampled pair is suppressed,
by using the brute-force method of calculating the permanent (in green) or by using the suppression
law (blue), plotted in logarithmic scale as a function of the number of n and p = log2 m. Continuous
lines: linear fits of the data, showing how the computational time with the brute-force method scales
exponentially in n, as opposite to the linear scaling in n required by the suppression law. (b) Average
number of seconds required to assess if a randomly sampled pair is suppressed using the suppression
law. The linear fits show that the scaling is not only polynomial, but roughly linear in n for each m (we
note that the interesting regime for boson sampling experiments is n m, therefore the irregularities
shown for n ≈ m are negligible).
While eq. (5.4b) gives a sufficient condition for an input-output pair to be suppressed,
this condition is not necessary. For the majority of input states, not all of the suppressed
output states satisfy theorem 3. For a more precise analysis of these quantities the follow-
ing definitions will be useful:
Definition 9. Let U be an m×m unitary matrix. Then we will denote with Ω(U) the
set of input-output (collision-free) suppressed pairs, and with Ωeff(U) those predicted
to be suppressed by theorem 3:
Ω(U) ≡ {(R, S) ∈ Qn,m×Qn,m | perm(U[R|S]) = 0} , (5.12)
Ωeff(U) ≡ (R, S) ∈ Ω(U) | NA
(R) = Rτ and ⊕k,α Sk,α = 1 . (5.13)
In other words, Ω(U) is the set of all n×n submatrices of U whose permanent vanishes,
while Ωeff(U) is the subset of the pairs in Ω(U) for which we can efficiently predict
81
that the permanent vanishes.
Clearly, we have Ωeff(U) ⊂ Ω(U) for any U.
An interesting question regards the evaluation of the ratio Ωeff(US
) / Ω(US
) , i.e.
the amount of suppressed pairs caught by theorem 3, which will be analyzed in some
detail in the next section.
There is however at least one class of inputs for which theorem 3 seems to completely
characterize the output suppressed states (that is, the suppressed output states are all
and only those predicted to be suppressed by theorem 3):
Conjecture 1. Let r and s be two many-boson states of n = 2q
bosons in m = 2p
modes,
let R and S be the corresponding BM representations, let A be a subset of the columns of
R, and let α(r) be the number of subsets A such that condition 5.4a holds, that is
α(r) ≡ {A ⊆ P({1, . . . , p}) : NA
(r) = r} , (5.14)
where we adopted the standard convention of denoting with P(X) the power set of a set
X. If α(r) = n − 1, then the output suppressed states are all and only those satisfying
eq. (5.4b).
Remark 3. Conjecture 1 predicts that for a (rather small) class of input states, the set of
output suppressed states is completely characterized by theorem 3, and has been verified
by brute-force computations for n = 2, 4, 8, 16. We note that this class of states includes
as a subset those considered by theorem 2, for which the result has been proved in [23].
In the next section we will present the results of a series of numerical computations
carried out to determine the dependence of the fraction of predicted suppressed pairs
from m and n.
5.3 Bayesian validation scheme
To assess if a given validation protocol is scalable at increasing n, it is useful to quantify
the number of experimental runs needed to reject an alternative hypothesis with a
given confidence level. Here we propose a simple model to predict the mean number of
samples required to reject the hypothesis C of distinguishable photons, and accept an
hypothesis Q of nearly indistinguishable photons, being a parameter quantifying the
amount of distinguishability introduced by experimental errors.
Let p ≡ Ωeff(US
) / m
n
2
be the fraction of suppressed n-photon input/output pairs
predicted by theorem 3. Given the unbiased nature of Sylvester matrices, p is also the
probability of observing an event in a suppressed pair, when sampling uniformly from
the set of all input-output pairs. The probability of observing an event in a suppressed
pair in the Q hypothesis will be instead quantified by the parameter , and assumed
82
to be a small number pQ( ) = p (the ideal quantum case will therefore correspond to
= 0).
The only assumption of the proposed validation scheme is that for some unknown
value 0, the probability of experimentally detecting an event in a suppressed pair is
given by pQ( 0) = 0p. We will then give a formula to assess if a given set of experimen-
tal data is more likely (with some fixed confidence level) to come from a Q evolution
as opposite to the distinguishable case C. More in particular, the proposed validation
scheme will accept a whole range of values of such that the relative probability of the
data coming from Q as opposed to C will be greater then a confidence level α.
Let M denote the total number of samples collected in an experiment, and let Msupp
be the number of those samples which turned out to be a suppressed pair according to
theorem 3. Then, assuming unbiased prior probabilities, the relative probability V of
the hypothesis Q over the hypothesis C, having observed Msupp efficiently predicted
suppressed pairs over M samples, is
V ≡
P(Q | M, Msupp)
P(C | M, Msupp)
=
P(M, Msupp| Q )
P(M, Msupp| C)
. (5.15)
In the case of completely distinguishable photons the probability is
P(M, Msupp| C) = pMsupp
(1 − p)M−Msupp
(5.16)
while if the dynamic of the system follows Q the probability is
P(M, Msupp| Q ) = ( p)Msupp
(1 − p)M−Msupp
. (5.17)
Using eqs. (5.16) and (5.17) inside eq. (5.15) gives
V = Msupp
(1 − p)M−Msupp
(1 − p)M−Msupp
, (5.18)
and we will consider Q to be successfully distinguished from C, with the confidence
level α, for all the values of such that V ≥ α.
Let’s now denote with 0 the parameter quantifying the real experimental distin-
guishability of the experiment (which is assumed to be unknown for the test). Then, on
average, the number of detected suppressed pairs will be Msupp = 0pM. Plugging this
into eq. (5.18) we obtain
V =

 0p 1 − p
1 − p
1− 0p


M
. (5.19)
In fig. 5.2 is shown an example of application of this test, using eq. (5.19). For a fixed
set of values for M, α, 0, and p, the accepted models Q are those corresponding to
83
the range of values of such that V ≥ α. In fig. 5.2 this is equivalently expressed by
plotting (V )1/M
(blue line) against , and comparing it with the value of α1/M
(purple
line). All values of for which the blue line is above the purple one are accepted, and
thus distinguished from the classical distribution with confidence α. Note how the
accepted values of are those close to 0 (red dotted line), which is the parameter
characerizing the actual distinguishability of the experimental apparatus.
Figure 5.2: Accepted range of values of , for
M = 300, α = 2000, 0 = 0.3, p = 0.1. The
blue line is a fit of (V )1/M
. The solid green line
corresponds to V = 1, the dotted red line to =
0, and the dotted purple line to (V )1/M
= α1/M
.
All the values of corresponding to the blue line
being above the purple one are accepted by the
test, with the confidence α. This range of accepted
values is highlighted in the figure by the solid
purple line.
Inverting eq. (5.19) we can extract the average number M , 0 of experimental runs
needed to reject C and accept Q with confidence α is
M , 0 =
log α
0p log + (1 − 0p) log 1− p
1−p
. (5.20)
As a function of , M , 0 has a positive minimum for = 0, as can be seen for
example in fig. 5.3. Setting thus = 0 in eq. (5.20) we obtain that the average number
of experimental samples required to assess with confidence α that the observed data are
not compatible with the evolution of distinguishable photons, is
M 0, 0 =
log α
0p log 0 + (1 − 0p) log 1− 0p
1−p
. (5.21)
A rough approximation of eq. (5.21) for small 0 is
M 0, 0 ≈
log α
log 1
1−p
+ O( 0), (5.22)
which for small p gives the scaling behaviour M 0, 0 ∼ log α/p. We therefore conclude
that the number of experimental runs needed to validate against distinguishable photons
scales polynomially in n if and only if the condition
p > 1/ poly(n) (5.23)
is satisfied, for small values of 0 and p.
84
0.2 0.4 0.6 0.8 1.0
ϵ
-500
500
1000
Mϵ,ϵ0
Figure 5.3: Needed number of samples
M , 0
(eq. (5.20)) as a function of , for 0 = 0.3,
p = 0.1, and α = 2000. The range of values of
corresponding to V < 1 result in M , 0
< 0.
For these the data is more likely to come from
C than Q . The green dashed line highlights the
value of = 0.
5.4 Conclusions
Further work is needed to assess whether the fraction of suppressed pairs given by
Sylvester matrices satisfies eq. (5.23). As shown in table 5.1 and fig. 5.4, the number
Ω(U) of suppressed pairs is significantly larger for Sylvester matrices than for Fourier
matrices. While it is clear that the former is better suited for scattershot sampling
validation than the latter, the scaling behaviour of Ω(US
) can hardly be deduced from
the data in table 5.1 and fig. 5.4-a.
m n Sylvester Fourier
8 2 4/7 57.14% 12/49 24.49%
8 3 0 8/49 16.33%
8 4 96/175 54.86% 52/245 21.22%
8 5 4/7 57.14% 12/49 24.49%
8 6 4/7 57.14% 24/49 48.98%
8 7 0 0
16 2 8/15 53.33% 32/225 14.22%
16 3 0 64/1225 5.22449%
16 4 480/1183 40.57% 11112/207025 5.37%
16 5 480/1183 40.57% 632/24843 2.54%
16 6 53160/143143 37.14% 256/11011 2.32%
16 8 21952/83655 26.24% 495172/41409225 1.20%
Table 5.1: Fractions of suppressed pairs (i.e. Ω(U)/ m
n
2
) for Sylvester and Fourier matrices, and various
values of the number of modes m and the number of photons n. The values are calculated computing,
with the permanent formula, all the scattering amplitudes between collision-free input-output states.
85
8
16
32
64
128
m
5 10 15 20n
0
20
40
%
8
16
32
64
128
m
5 10 15 20n
0
10
20
30
40
a) b)
%
Figure 5.4: Estimated fraction of suppressed pairs with Fourier and Sylvester matrices. (a)
Fraction of suppressed pairs with Sylvester matrices, that is, Ω(US
)/ m
n
2
, plotted against n for various
values of m. The data is obtained randomly sampling N = 1000 input-output pairs, and computing the
permanent corresponding to each scattering amplitude. The percentage is known to be zero for number
of photons of the form n = 2p
− 1 [81], regardless of m. (b) Fraction of suppressed pairs with Fourier
matrices, that is, Ω(UF
)/ m
n
2
, plotted against n for various values of m. The fraction is in this case
significantly lower than for Sylvester matrices. For most values of n, no suppressed pair was found in
the sampled dataset.
In fig. 5.5 are again given the estimated fractions for Fourier and Sylvester matrices.
The error bars in fig. 5.5 represent the 99.7 % confidence interval, calculated through
Bayesian inference from the number of suppressed pairs found in the sampled datasets.
In particular, in fig. 5.5-a the fraction Ω(US
) depends weakly on m, as opposite to the
strong dependence of UF
on m shown in fig. 5.5-b.
Another relevant matter is the scaling with m and n of Ωeff(US
), that is, the number
of suppressed pairs satisfying the conditions of theorem 3. No efficiently evaluable
analytical expression is known for this number. Brute-force estimates are given in
fig. 5.6, where calculations similar to those in fig. 5.5 are shown, but this time for the
efficiently predictable suppressed pairs. In the presented cases of m = 16, 32, 64, 128,
Ωeff(US
) seems to decrease exponentially in n, for n < m/2.
In conclusion, while theorem 3 represents a significant improvement with respect
to theorem 1 and theorem 2, this fraction is probably still not large enough for an
efficient validation test with Sylvester matrices. However the data on the total fraction
of suppressed states presented in figs. 5.4 and 5.5 leaves open the possibility that, further
86
improving on theorem 3, an efficient validation scheme for Sylvester matrices may be
devised.
m =32
m =64
m =128
m =256
Sylvester
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
n
10
20
30
40
50
%
m =32
m =64
m =128
m =256
Fourier
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
n
2
4
6
8
%a) b)
Figure 5.5: Fraction of suppressed pairs with Fourier and Sylvester matrices. Each color repre-
sents the fraction corresponding to a different number of modes m, plotted as a function of the number
of photons n. The error bars approximately represent the 99.7% confidence interval, estimated from the
sampled data through Bayesian inference. The approximation used breaks for very small number of
found suppressions, which is seen in the points where the error bars span negative values.
2 4 6 8 10 12 14
n
10
20
50
%
m=16
2 4 6 8 10 12 14 16 18 20 22 24 26 28 30
n
0.10
0.50
1
5
10
50
%
m=32
2 10 18 26 34 42 50 58
n
10-4
0.001
0.010
0.100
1
10
100
%
m=64
18 34 50 66 82 98 114
n
0.5
1
5
10
50
%
m=128
Figure 5.6: Estimated fraction of predicted suppressed pairs with Sylvester matrices. Plot of the
percentage of predicted suppressed input-output pairs against n, for various values of m, for Sylvester
matrices. The data is calculated randomly sampling input-output collision-free pairs, and verifying
whether each is suppressed according to theorem 3. The missing data corresponds to values of n at
which no suppression in the sampled pairs was found. For m = 64, the size of the sample for which no
suppressions were found is N = 16 × 103
, while for m = 128 is N = 2000. Larger samples were not
used due to the heavy computational times required.
87
88
Chapter 6
Experimental boson sampling
validation with integrated photonics
In this chapter we report the experimental implementation of the protocol devised
in [22, 24], and explained in section 4.2, and its use to validate quantum many-body
interference against alternative non-trivial hypotheses [25]. To this end, a scalable
approach for the implementation of the fast Fourier transform algorithm, using 3-D
photonic integrated interferometers, is developed. We observe the suppression law for
a large number of output states with 4- and 8-mode optical circuits. The experimental
results demonstrate genuine quantum interference between the injected photons. This
work provides a first experimental realization [25] of the validation protocol with
Fourier matrices [22] discussed in section 4.2 (during this work another experimental
realization of the same validation scheme was reported [17]). The reported experiments
were realized in the Quantum Optics laboratories at University of Rome La Sapienza.
In section 6.1 an overview of the work is given. In section 6.2 the femtosecond
laser-writing technique, employed to build the integrated photonic chips used in the
experiment, is described. In section 6.3 the quantum Fast Fourier Transform is described.
This is a linear optics generalization of the classical Fast Fourier Transform algorithm,
used in this work to implement efficiently the Fourier transformation on the integrated
interferometer. In section 6.4 we give a brief description of the photon generation
methods. In section 6.5 the results of the measurements are presented, focusing on the
observation of the Hong-Ou-Mandel dips and peaks, and the reconstruction of the chip
transformation. In section 6.6 we describe how the experimental results validate the
genuine many-boson quantum interference occuring in the interferometer, as opposite
to alternative models of evolution. Finally, in section 6.7, the results of the experiments
are discussed and the possible future prospects of this line of research.
89
6.1 Overview of the experiment
We here report the experimental observation of the recently theoretically proposed [22]
suppression law for Fourier matrices, and its use to validate quantum many-body inter-
ference against alternative non-trivial hypotheses resulting in similar output probability
distributions.
The Fourier matrices have been implemented with an efficient and reliable approach
by exploiting the quantum version of the Fast Fourier Transform (qFFT), an algorithm
developed by Barak and Ben-Aryeh [26] to optimize the number of optical elements
required to build the Fourier transform over the optical modes.
Here we implement the qFFT on photonic integrated interferometers by exploiting
the 3-D capabilities of femtosecond laser writing [82, 83], which makes it possible to fab-
ricate waveguides arranged in three-dimensional structures with arbitrary layouts [13,
84, 85], by adopting an architecture scalable to a larger number of modes.
The observations have been carried out with two-photon Focks states injected into
4- and 8-mode qFFT interferometers. The peculiar behaviour of Fock states compared to
other kinds of states is investigated, showing in principle the validity of the certification
protocol for the identification of true granular n-particle interference, which is the
source of a rich landscape of quantum effects such as the computational complexity of
BosonSampling.
6.2 Femtosecond laser writing
Building large interferometers out of discrete, bulk optical elements tends to result in
mechanical instabilities which have prevented the demonstration of even a symmetric,
3-mode interferometer that preserves quantum coherence. A more promising approach
to obtain stable multi-mode interferometers involves the fabrication of this network of
linear optical elements by integrated optical waveguides in a glass chip [13, 84].
Waveguides are fabricated using the femtosecond laser micromachining technique
[83, 86], which exploits nonlinear absorption of focused femtosecond pulses to induce a
permanent and localized refractive index increase in transparent materials, as schemat-
ically shown in fig. 6.1. Arbitrary three-dimensional circuits can be directly written
by translating the sample along the desired path, keeping the velocity constant with
respect to the laser beam. This technique allows fast and cost-effective prototyping
of new devices, enabling the implementation of three-dimensional layouts that are
impossible to realize with conventional lithography.
90
Figure 6.1: Laser-writing a waveguide on a
glass chip. The waveguide is fabricated focusing
femtosecond laser pulses to induce a permanent
and localized refractive index increase in transpar-
ent materials.
In the integrated optics approach the role of beam splitters is performed by direc-
tional couplers, devices which bring two waveguides close together to redistribute the
light propagating in them by evanescent field coupling (fig. 6.2-c). The phase shifters
are instead implemented by deforming the S-bent waveguides at the input of each
directional coupler in order to stretch the optical path (fig. 6.2-b). The integrated optics
analogue of a discrete components layout, depicted in fig. 6.3-a, is shown in fig. 6.3-b,
where one can appreciate the one-to-one correspondence between elements in the two
approaches.
TRANSMISSION CONTROL
PHASE-SHIFT CONTROL
0.00 0.05 0.10 0.15 0.20 0.25
0
1
2
3
4
5
d
Phaseshift[rad]
a
b
c
0 1 2 3 4 5 6
0.0
0.2
0.4
0.6
0.8
1.0
Angle [°]
Transmissivity
d [μm]
Transmission
Figure 6.2: Independent control of the phase shift and transmissivity at each directional cou-
pler. (a) The controlled deformation of the S-bent waveguide at the input of each directional coupler
and coupling geometry allows independent control over the phase shift and transmissivity. (b) The
deformation of the S-bent waveguide section is function of a deformation coefficient d. The graph
showed the undeformed S-bend together with a deformed one. (c) Control over the transmissivity of the
directional coupler is performed by modulating the coupling coefficient. This is achieved by changing
the waveguide spacing in the coupling region by rotating one arm of the directional coupler out of the
main circuit plane. A sample dependence of the the transmissivity as a function of the deformation angle
is provided. Credits: [10].
91
t1 t2 t3 t4
t5 t6 t7
t8 t9
t10
IN
1
2
3
4
5
OUT
1
2
3
4
5
t1
t5 t6 t7
t2 t3 t4
t8 t9
t10
IN
1
2
3
4
5 5
4
3
2
1
OUT
a
b
tiαi
βi
tiαi
βi
Figure 6.3: Layout of multimode interferometers. (a) Realization of an arbitrary 5×5 mode trans-
formation via a network of beam splitters with different transmissivities ti. The blue and red boxes stand
for different phase shifters. (b) Implementation of the same scheme adopting integrated photonics. αi
and βi are parameters characterizing the phase shifts acquired by the photons passing through S-bend
waveguide segments. Credits: [10].
6.3 Realization of the quantum Fast Fourier Trans-
form
We here introduce an experimental implementation of the Fourier transformation with
the qFFT approach.
The general method to realize an arbitrary unitary transformation using linear
optics was introduced by Reck et al. [79], who provided a decomposition of an m×m
unitary as a sequence of m(m − 1)/2 beam splitters and phase shifters. However, in
the special case of Fourier matrices, a more efficient method has been proposed [26,
87], which takes advantage of their symmetries to significantly reduce the number of
linear optical elements required.
Based on the classical algorithm of Cooley and Tukey [27], who first introduced
the Fast Fourier Transform algorithm as a more efficient way to compute the discrete
92
Fourier transform, Barak and Ben-Aryeh [26] developed a quantum analogue in the
linear optics domain, leading to the concept of qFFT. This approach, valid for 2p
-
dimensional Fourier matrices, requires only (m/2) log2 m beam splitters and phase
shifters, to be compared with the O(m2
) elements needed for the more general Reck
decomposition, thus enhancing the compactness and scalability of the platform for a
more reliable experimental realization.
The sequential structure arising from the decomposition of the m-dimensional
Fourier matrix using the Barak and Ben-Aryeh algorithm is reproduced by the con-
secutive layers shown in fig. 6.4. The complex arrangement of pairwise interactions
necessary for the qFFT method cannot be easily implemented using a planar architecture.
However, femtosecond laser writing technique allows to overcome this issue exploiting
the third dimension, arranging the waveguides along the bidimensional sections of the
integrated chip.
phase phasedirectional coupler directional coupler
4
3
2
1
4
3
2
1
31
6
42758
L1
L2
L1
L2
L3
P1
P2
P1
3
1
4
2
7
5
8
6
a b
L1 L2
42
31
P1 L1 L2P1 L3P2
⇡/23:
7:
8: 6:
8:
4:
⇡/2
⇡/2
⇡/2
⇡/4
3⇡/442
31
73
51
84
62
73
51
84
62
73
51
84
62
Figure 6.4: Schematic of the structure of the integrated devices. Internal structure of the 4-mode
(a) and 8-mode (b) integrated interferometers implementing the qFFT over the optical modes. The
mode arrangement has been chosen in a way to minimize bending losses. The insets shows the actual
disposition of the waveguides in the cross-section of the devices. The modes coupled together in each
step (Li) of the interferometer are joined by segments. The implemented phase shifts in each step (Pi)
are also indicated. Credits: [25].
The strategy can be outlined as follows: the 2p
modes are ideally placed on the
vertices of a p-dimensional hypercube, and in each step of the algorithm the vertices
connected by parallel edges having one specific direction are made to interact by a
2-mode Hadamard transformation, with proper phase terms. An optical interferometer
implementing this procedure is thus composed of log2 m = p sections, each employing
m/2 balanced beam splitters and phase shifters.
The waveguide interferometers were fabricated realizing the Fourier matrix for m =
4 and m = 8 modes in borosilicate glass chips using femtosecond laser micromachining
[83, 84]. A schematic representation of these two interferometers is given in fig. 6.4.
According to the scheme outlined above and by exploiting the three-dimensional
93
capabilities of the fabrication technique, the waveguides are placed, for what concerns
the cross section of the device, on the vertices of a 2-D projection of the p-dimensional
hypercube. Three-dimensional directional couplers, with proper interaction length
and distance to achieve a balanced splitting, connect in each step the required vertices.
The insets of fig. 6.4 show, at each step i, which modes are connected by directional
couplers (Li) and the amount of phase shift that needs to be introduced in specific
modes (Pi). Phase shifters, where needed, are implemented by geometrical deformation
of the connecting S-bends. Fan-in and fan-out sections at the input and output of the
device allows interfacing with 127-µm spaced single-mode fiber arrays.
We note that in our device geometry, in each step, the vertices to be connected are
all at the same relative distance. This means that, unless geometric deformations are
designed where needed, light travelling in different modes does not acquire undesired
phase delays.
The geometric construction here developed is scalable to an arbitrary number of
modes with a number of elements increasing as m log2 m.
6.4 Photon generation and manipulation
The generation of two-photon states is performed by pumping a 2-mm long BBO
crystal with the second harmonic of a 785 nm wavelength Ti:Sa pulsed laser, with
an average power of 750 mW, which generates photons at 785 nm with a type II
parametric downconversion process. The two photons are spectrally filtered by means
of interferential filters with 3 nm full width at half maximum, and coupled into single
mode fibers. The indistinguishability of the photons is then ensured by a polarization
compensation stage, and by propagation through independent delay lines before the
injection within the interferometer via a single mode fiber array.
After the evolution through the integrated devices, the photons are collected via a
multimode fiber array. The detection system consists of 4 (8) single photon avalanche
photodiodes used for the 4-(8-)mode chip. An electronic data acquisition system allowed
us to detect coincidences between all pairs of output modes. Typical coincidence rates
for each collision-free output combination with distinguishable photons amounted to
∼70-80 Hz for the 4-mode chip, and ∼10-20 Hz for the 8-mode chip.
6.5 One- and two-photon measurements in integrated
Fourier interferometers
The two implemented interferometers of m = 4 and m = 8 modes are fed with one-
and two-photon states. The experimental setup, preparing a biphoton wavepacket to
be injected into the devices, is shown in fig. 6.5.
94
a b
IF
HWP
PBS
PC
PDC output FA
input FA
chip
PBS
HWP
IF
PC
DL
DL
SMF
SMF
input FA
Figure 6.5: Experimental apparatus for input state preparation. (a) The photon source (IF: inter-
ferential filter, HWP: half-wave plate, PBS: polarizing beam splitter, PC: polarization compensator, PDC:
parametric downconversion, DL: delay lines with motorized stages, SMF: single-mode fiber). (b) Photon
injection (extraction) before (after) the evolution through the interferometer (FA: fiber array). Credits:
[25].
To test the validity of the suppression law, we measured the number of coincidences
at each forbidden output combination injecting cyclic inputs with two indistinguishable
photons. The degree of violation D of the suppression law could simply be evaluated
with a counting experiment. Alternatively, the same quantity D can be expressed as a
function of single-photon input-output probabilities and of the HOM visibilities, defined
as
Vi,j =
ND
i,j − NQ
i,j
ND
i,j
, (6.1)
where ND
i,j is the number of detected coincidences for distinguishable photons, NQ
i,j that
for indistinguishable photons, and the subscripts (i, j) are the indices of the two output
modes, for a given input state. The degree of violation can therefore be expressed as:
D =
Nforbidden
Nevents
= Pforbidden =
(i,j)forbidden
PQ
i,j =
(i,j)forbidden
PD
i,j(1 − Vi,j), (6.2)
where PQ
i,j (PD
i,j) are the probabilities of having photons in the outputs i, j in the
case of indistinguishable (distinguishable) particles. Here PD
i,j can be obtained from
single-particle probabilities. The visibilities are measured by recording the number of
coincidences for each output combination as a function of the temporal delay between
the two injected photons.
For the 4-mode device, we measured the full set of 4
2
= 36 collision-free input-
output combinations, that is, those combinations corresponding to the two photons
exiting from two different ports. These contributions have been measured by recording
95
the number of coincidences for each combination of two outputs as a function of the
temporal delay between the two input photons. Due to the suppression law given in
theorem 1, we expect to observe four suppressed outcomes (over six possible output
combinations) for the two cyclic input states (1,3) and (2,4). Since distinguishable
photons exhibit no interference, HOM dips in the coincidence patterns are expected for
the suppressed output states. Conversely, peaks are expected in the non-suppressed
output combinations. The experimental results are shown in fig. 6.6, where the expected
pattern of four suppressions and two enhancements is reported, with average visibilities
of ¯Vsupp = 0.899 ± 0.001 and ¯Venh = −0.951 ± 0.004 for suppression and enhancement
respectively.
For the cyclic inputs, we also measured the interference patterns for the output
contributions where the two photons exit from the same mode. These terms have been
measured by inserting an additional symmetric beam splitter on each output mode,
and by connecting each of its two outputs to a single-photon detector. These cases
correspond to a full-bunching scenario with n = 2, and a HOM peak with V = −1
visibility is expected independently from the input state and from the unitary operation
[71]. This feature has been observed for the tested inputs, where an average visibility
of ¯Vbunch = −0.969 ± 0.024 has been obtained over all full bunching combinations.
The injection of non-cyclic input states has been employed for the complete recon-
struction of the chip action ˜UqFFT
4 , using a data set statistically independent from the
one adopted to observe the suppression law. The adopted reconstruction algorithm,
which exploits knowledge on the internal structure of the interfometers shown in
fig. 6.4, works in two steps. In a first step, the power-splitting ratios measured with
classical light are employed to extrapolate the transmissivities of the directional cou-
plers. In a second step, the two-photon visibilities for the non-cyclic inputs are used to
retrieve the values of the fabrication phases. In both steps, the parameters are obtained
by minimizing a suitable χ2
function. The results are shown in fig. 6.7. The fidelity
between the reconstructed unitary ˜UqFFT
4 and the theoretical Fourier transform UF
4 is
F = 0.9822 ± 0.0001, thus confirming the high quality of the fabrication process. The
error in the estimation of the fidelity is obtained through a Monte Carlo simulation,
properly accounting for the degree of distinguishability of the photons with a rescaling
factor in the visibilities.
96
- 0.642(8)
2500
0.890(2)
2900
0.682(3)
2490
0.525(3)
2860
0.882(2)
2290
- 0.662(7)
2600
0.871(2)
2860
- 0.958(8)
2870
0.871(2)
2920
0.878(1)
3090
- 0.961(9)
2640
0.877(1)
3620
0.603(4)
1640
0.898(1)
4320
- 0.692(11)
1330
- 0.518(9)
1570
0.897(1)
2940
0.619(4)
1630
0.622(3)
3460
0.869(2)
3090
- 0.709(7)
3120
- 0.547(7)
2970
0.898(1)
3160
0.596(3)
2270
0.916(1)
3940
- 0.953(7)
3410
0.917(1)
3670
0.911(1)
3310
- 0.931(8)
2660
0.916(2)
2450
- 0.606(8)
2150
0.898(1)
4380
0.660(3)
3210
0.536(3)
3590
0.869(2)
3230
- 0.628(7)
2990
in (1,2)
in (1,3)
in (1,4)
in (2,3)
in (2,4)
in (3,4)
out (1,2) out (1,3) out (1,4) out (2,3) out (2,4) out (3,4)
Figure 6.6: Suppression law in a 4-mode Fourier integrated chip. Complete set of 36 measured
coincidence patterns (raw experimental data) for all input-output combinations in the 4-mode chip. For
each input-output combination, the measured coincidence pattern as a function of the time delay is
shown (points: experimental data, lines: best-fit curves). Cyclic inputs, highlighted in the picture, exhibit
enhancement (green) and suppression (red) on cyclic and non-cyclic outputs, respectively. For all points,
error bars are due to Poissonian statistics of the events. For each visibility, the error is obtained through a
Monte Carlo simulation by averaging over 3000 simulated data sets. In each plot the zero level coincides
with the baseline, while a dashed line represents the number of coincidence events in the distinguishable
limit. Credits: [25].
97
VrecVqF F T VF
1 2 3 4
1
2
3
4
1 2 3 4
1
2
3
4
OUT
IN
0.4
0.45
4- 1/2
0.55
0.6
a)
b)
Figure 6.7: Suppression law in a 8-mode qFFT integrated chip. (a) HOM visibilities for all 36 input-
output configurations. From left to right: experimental measured visibilities (VqFFT, obtained from
raw experimental data), visibilities calculated from the reconstructed unitary (Vrec), and visibilities
calculated from the theoretical unitary (VF). (b) Representation of the reconstructed experimental
transformation ˜UqFFT
4 , and comparison with UF
4. Colored disks represent the moduli of the reconstructed
matrix elements (all equal to 4−1/2
for UF
4). Arrows represent the phases of the unitary matrix elements
(green: reconstructed unitary, blue: Fourier matrix). Credits: [25].
98
For the 8-mode chip we recorded all the 8
2
= 28 two-photon coincidence patterns,
as a function of the relative delay between the input photons, for each of the 4 collision-
free cyclic inputs and for one non-cyclic input. The results are shown in fig. 6.8.
The reconstruction of the actual unitary transformation ˜UqFFT
8 implemented has been
performed with the same algorithm of the 4-modes, by using the power-splitting ratios
measured with classical light and the two-photon visibilities for one non-cyclic input.
The latter has been chosen in a way to maximize the sensitivity of the measurements
with respect to the five fabrication phases. The results are show in fig. 6.9. The fidelity
between the reconstructed unitary ˜UqFFT
8 and the ideal 8-mode Fourier transform UF
8 is
F = 0.9527 ± 0.0006.
0.800(6)
480
0.977(23)
480
0.844(6)
400
1.019(27)
350
0.772(9)
240
0.971(26)
360
0.850(6)
290
0.813(4)
980
0.856(20)
490
0.755(7)
440
0.895(24)
340
0.755(8)
270
0.838(19)
520
0.850(4)
690
0.890(25)
310
0.800(6)
480
0.960(24)
370
0.822(5)
590
0.790(7)
360
0.980(28)
260
0.824(5)
520
0.959(25)
300
0.796(7)
300
0.938(27)
260
0.765(9)
280
0.795(9)
250
0.875(27)
280
0.847(7)
380
out 1
out 2
out 3
out 4
out 5
out 6
out 7
out 2 out 3 out 4 out 5 out 6 out 7 out 8
Figure 6.8: Suppression law in a 8-mode qFFT integrated chip. Set of 28 measured coincidence
patterns (raw experimental data), corresponding to all collision-free output combinations for the input
(2,6) of the 8-mode interferometer. For each output combination, the measured coincidence pattern
as a function of the time delay is shown (points: experimental data, lines: best-fit curves). Red or
green backgrounds correspond to dips and peaks, respectively. For all points, error bars are due to
the Poissonian statistics of the events. For each visibility, the error is obtained through a Monte Carlo
simulation by averaging over 3000 simulated data sets. In each plot the zero level coincides with the
baseline, while a dashed line represents the number of coincidence events in the distinguishable limit.
Credits: [25].
99
Average visibilities
(1,5) (2,6) (3,7) (4,8)
(1,5) (2,6) (3,7) (4,8)
- 0.25
- 0.5
- 0.75
- 1
0.25
0.5
0.75
1
1 2 3 4 5 6 7 8
1
2
3
4
5
6
7
8
1 2 3 4 5 6 7 8
1
2
3
4
5
6
7
8
OUT
IN
0.3
0.33
8- 1/2
0.38
0.4a) b)
Figure 6.9: Suppression law in a 8-mode qFFT integrated chip. (a) Average visibilities of dips
(red bars) and peaks (green bars) observed for the four collision-free cyclic inputs ((1,5),(2,6),(3,7),(4,8)).
Darker regions correspond to error bars of ±1 standard deviation. (b) Representation of the reconstructed
experimental transformation ˜UqFFT
8 , and comparison with UF
8. Colored disks represent the moduli of the
reconstructed matrix elements (all equal to 8−1/2
for UF
8). Arrows represent the phases of the unitary
matrix elements (green: reconstructed unitary, blue: Fourier matrix). Credits: [25].
6.6 Observation of the suppression law
The suppression of events which do not satisfy theorem 1 is fulfilled only when two
perfectly indistinguishable photons are injected in a cyclic input of a perfect Fourier
interferometer. In such a case, we would have the suppression of all output states such
that the sum of the elements of the corresponding MAL is odd. For the 4-mode (8-mode)
device, this corresponds to 4 (16) suppressed and 2 (12) non-suppressed collision-free
outputs (each one given by two possible arrangements of the two distinguishable
photons), plus 4 (8) terms with two photons in the same output, each one corresponding
to a single possible two-photon path.
The expected violation for distinguishable particles can be obtained from classical
considerations. Let us consider the case with n = 2. The two distinguishable photons
evolve independently from each other, and the output distribution is obtained by
classically mixing single-particle probabilities. All collision-free terms are equally likely
to occur with probability q = 2/m2
, while full-bunching events occur with probability
q = q/2 = 1/m2
. The degree of violation DD can then be obtained by multiplying the
probability q by the number of forbidden output combinations. As a result, we expect a
violation degree of DD = 0.5 for distinguishable two-photon states. The evaluation of
the expected value for a mean field state, which is due to single particle bosonic statistic
effects, require different calculations [22]. It can be shown that for n = 2 the degree of
violation is DMF = 0.25.
For each of the cyclic inputs, we have evaluated here the violation degree Dobs
100
resulting from collected data. By measuring the coincidence pattern as a function of
the path difference ∆x = c∆τ between the two photons, and thus by tuning their
degree of distinguishability, we could address the transition from distinguishable to
indistinguishable particles. The value of Dobs as a function of ∆x has been obtained
as (i,j)forbidden
PD
i,j(N∆x
i,j /ND
i,j), where N∆x
i,j and ND
i,j are the number of measured coin-
cidences for a given value of ∆x and for distinguishable particles respectively. Two
different regions can be identified. For intermediate values of ∆x with respect to the
coherence length of the photons, the measured data fall below the threshold DD, and
hence the hypothesis of distinguishable particles can be ruled out. Then, for smaller
values of the path difference up to ∆x → 0, true two-photon interference can be
certified since both hypothesis of distinguishable particles and mean field state can
be ruled out. The maximum violation occurring at ∆x = 0 delay can be evaluated
using eq. (6.2). The experimental results retrieved from the protocol are shown in the
table of figs. 6.10 and 6.11, in which we compare the values Dobs(0) with the expected
values for distinguishable particles DD and for a Mean Field state DMF. As shown for
our implementation, the robustness of the protocol is ensured by the high number of
standard deviations separating the values in each comparison, thus unambiguously
confirming the success of the certification protocol. In conclusion, the alternative
hypotheses of distinguishable particles and of a mean field state can be ruled out for all
experiments.
101
Figure 6.10: Observed violations Dobs as a function of the path difference |∆x| = c|∆τ| between the
two photons, for the 4-mode interferometer. Blue shaded regions in the plots correspond to the cases
where the hypothesis of distinguishable particles can be ruled out, and true two-particle interference is
present. Blue points: input (1,3). Red points: input (2,4). Blue solid line: theoretical prediction for input
(1,3). Red solid line: theoretical prediction for input (2,4). Black dashed line: theoretical prediction for a
Fourier matrix. (b) Data for the 8-mode interferometer. Blue points: input (1,5). Red points: input (2,6).
Green points: input (3,7). Magenta points: input (4,8). Colored solid lines: corresponding theoretical
predictions for different inputs. Black dashed line: theoretical prediction for a Fourier matrix. Table:
violations Dobs(0) at ∆x = 0 and discrepancies (in sigmas) with the expected values for distinguishable
particles (DD) and Mean Field Sampler (DMF), for the cyclic inputs of the two interferometer. Dobs(0)
are calculated following eq. (6.2), while expected values for the other two cases are DD = 0.5 and
DMF = 0.25. Error bars in all experimental quantities are due to the Poissonian statistics of measured
events. All theoretical predictions in solid lines are calculated from the reconstructed unitaries, obtained
from different sets of experimental data to ensure statistical independence. Credits: [25].
102
Figure 6.11: Observed violations Dobs as a function of the path difference |∆x| = c|∆τ| between the
two photons, for the 8-mode interferometer. Blue shaded regions in the plots correspond to the cases
where the hypothesis of distinguishable particles can be ruled out, and true two-particle interference is
present. Blue points: input (1,5). Red points: input (2,6). Green points: input (3,7). Magenta points: input
(4,8). Colored solid lines: corresponding theoretical predictions for different inputs. Black dashed line:
theoretical prediction for a Fourier matrix. Table: violations Dobs(0) at ∆x = 0 and discrepancies (in
sigmas) with the expected values for distinguishable particles (DD) and Mean Field Sampler (DMF), for
the cyclic inputs of the two interferometer. Dobs(0) are calculated following eq. (6.2), while expected
values for the other two cases are DD = 0.5 and DMF = 0.25. Error bars in all experimental quantities
are due to the Poissonian statistics of measured events. All theoretical predictions in solid lines are
calculated from the reconstructed unitaries, obtained from different sets of experimental data to ensure
statistical independence. Credits: [25].
103
6.7 Discussion
We have reported on the experimental observation of the suppression law on specific
output combinations of a Fourier transformation due to quantum interference between
photons. The observation of the suppression effect allowed us to rule out alternative
hypotheses to the Fock state. The use of a novel implementation architecture, enabled
by the 3-D capabilities of femtosecond laser micromachining, extends the scalability
of this technique to larger systems with lower experimental effort with respect to
other techniques. While the presented architecture is designed to implement a Fourier
matrix for a number of modes equal to m = 2p
, a generalization of the approach can be
obtained by adopting a building block different from the beam splitter. For devices of
odd dimension, for instance, such a tool can be provided by the tritter transformation
[13]. At the same time, the universality of a generalized HOM effect with an arbitrary
number of particles and modes is expected to make it a pivotal tool in the diagnostic and
certification of quantum photonic platforms. BosonSampling represents a key example,
since the scalability of the technique is expected to allow efficient certification of devices
outperforming their classical counterparts. An interesting open problem is whether the
computational hardness of BosonSampling is mantained if the randomness condition
is relaxed, and thus which is the minimal interferometric architecture required for an
evidence of post-classical computation.
Fourier matrices can find application in different contexts. For instance, multiport
beam splitters described by the Fourier matrix can be employed as building blocks for
multiarm interferometers, which can be adopted for quantum-enhanced single and
multiphase estimation protocols [88]. This would also allow the measurement of phase
gradients with precision lower than the shot-noise limit [89]. Other fields where Fourier
matrices are relevant include quantum communication scenarios [90], observation of
two-photon correlations as a function of geometric phase [91], fundamental quantum
information theory including mutually unbiased bases [92], as well as entanglement
generation [93].
104
Conclusions
In this work we presented both theoretical and experimental advances in the context of
validating boson sampling experiments, both with the originally proposed approach
[8] and the recently proposed scattershot boson sampling scheme [19, 20].
From the theoretical point of view, the existing validation techniques have been
analysed in light of the recently proposed, and experimentally implemented, scattershot
boson sampling scheme [18, 19, 20]. We found that the existing methods to validate
genuine many-boson interference cannot be efficiently applied to scattershot boson
sampling experiments. Subsequently, we devised a new Zero-Transmission Law (ZTL)
for Sylvester matrices which has dramatically higher predictive capabilities with respect
to previously proposed ZTLs [23, 24]. We studied the fraction of suppressed input-output
pairs predicted by this ZTL, and its scaling with the number of photons n and the number
of modes m. We found that the number of predicted suppressed input-output pairs,
while significantly higher than that presented by previous ZTLs, still scales badly with n,
decreasing exponentially. However, it is less clear how the total number of suppressed
pairs in Sylvester matrices actually scales. Brute-force computations are not possible
beyond relatively small numbers of n, but the data gathered so far shows a rather
complex behaviour, which leaves open the possibility that this ratio decreases only
polynomially with n and m. Further work is needed in this direction to conclusively
assess whether an efficient validation scheme based on Sylvester matrices is possible,
but what is currently known leaves open this possibility.
From the experimental point of view, we reported on the first implementation of
a validation scheme [22] based on a ZTL for Fourier matrices [24]. This was realized
using integrated photonics technology, with 2 photons injected into 4- and 8-mode in-
terferometers realized on glass chips with femtosecond laser writing technique. A novel
methodology to efficiently implement a 2p
-dimensional Fourier matrix was employed,
exploiting the 3-D capabilities of femtosecond laser writing to implement the Barak and
Ben-Aryeh [26] quantum optics generalization of the Fast Fourier Transform algorithm
[27]. A thorough analysis of the Hong-Ou-Mandel dips and peaks corresponding to
the various possible configurations of 2-photon input-output pairs is reported, and
the transition from distinguishable to indistinguishable photons studied, showing an
excellent agreement with the predictions of quantum mechanics. An algorithm to effi-
105
ciently reconstruct the unitary matrix implemented in the device is presented, showing
a good agreement with the ideal Fourier matrix. Finally, the efficient validation scheme
proposed in [22] is implemented, finding a perfect match with the expected pattern of
suppressions, and ruling out alternative models such as distinguishable particles and
Mean Field sampler [22].
106
Aknowledgments
We aknowledge Fabio Sciarrino and Alessandro Cianchi, who supervised the presented
work, for valuable teachings and useful discussions. We also wish to aknowledge
Nicolò Spagnolo, Marco Bentivegna, Fulvio Flamini, and Niko Viggianiello, who closely
followed every step of this work, for the many tips and discussions.
107
108
Bibliography
[1] Richard P Feynman. “Simulating physics with computers”. In: International jour-
nal of theoretical physics 21.6/7 (1982), pp. 467–488.
[2] John Preskill. Quantum computing and the entanglement frontier. 2012. arXiv:
1203.5813 [quant-ph].
[3] John Preskill. Supremacy Now? 2012. url: http://quantumfrontiers.
com/2012/07/22/supremacy-now/ (visited on 10/27/2015).
[4] John Preskill. Fault-tolerant quantum computation. 1998. arXiv: quant-ph/
9712048.
[5] Emanuel Knill, Raymond Laflamme, and Wojciech H Zurek. “Resilient quantum
computation”. In: Science 279.5349 (1998), pp. 342–345.
[6] Dorit Aharonov and Michael Ben-Or. “Fault-tolerant quantum computation with
constant error rate”. In: SIAM Journal on Computing 38.4 (2008), pp. 1207–1282.
arXiv: quant-ph/9906129.
[7] Emanuel Knill, Raymond Laflamme, and Gerald J Milburn. “A scheme for efficient
quantum computation with linear optics”. In: nature 409.6816 (2001), pp. 46–52.
[8] Scott Aaronson and Alex Arkhipov. “The computational complexity of linear
optics”. In: Proceedings of the forty-third annual ACM symposium on Theory of
computing. ACM. 2011, pp. 333–342. arXiv: 1011.3245 [quant-ph].
[9] Matthew A Broome et al. “Photonic boson sampling in a tunable circuit”. In:
Science 339.6121 (2013), pp. 794–798. arXiv: 1212.2234 [quant-ph].
[10] Andrea Crespi et al. “Integrated multimode interferometers with arbitrary designs
for photonic boson sampling”. In: Nature Photonics 7.7 (2013), pp. 545–549. arXiv:
1212.2783 [quant-ph].
[11] Max Tillmann et al. “Experimental boson sampling”. In: Nature Photonics 7.7
(2013), pp. 540–544. arXiv: 1212.2240 [quant-ph].
[12] Justin B Spring et al. “Boson sampling on a photonic chip”. In: Science 339.6121
(2013), pp. 798–801. arXiv: 1212.2622 [quant-ph].
109
[13] Nicolò Spagnolo et al. “Three-photon bosonic coalescence in an integrated tritter”.
In: Nature communications 4 (2013), p. 1606. arXiv: 1210.6935 [quant-
ph].
[14] Nicolò Spagnolo et al. “Experimental validation of photonic boson sampling”. In:
Nature Photonics 8.8 (2014), pp. 615–620. arXiv: 1311.1622 [quant-ph].
[15] Jacques Carolan et al. “On the experimental verification of quantum complexity
in linear optics”. In: Nature Photonics 8.8 (2014), pp. 621–626. arXiv: 1311.2913
[quant-ph].
[16] Jacques Carolan et al. “Verifying quantum complexity in linear optical experi-
ments”. In: CLEO: QELS_Fundamental Science. Optical Society of America. 2014,
FM2A–7.
[17] Jacques Carolan et al. “Universal linear optics”. In: Science 349.6249 (2015), pp. 711–
716. doi: 10.1126/science.aab3642. arXiv: 1311.2913 [quant-
ph]. url: http://www.sciencemag.org/content/349/6249/
711.abstract.
[18] Marco Bentivegna et al. “Experimental scattershot boson sampling”. In: Science
Advances 1.3 (2015), e1400255. arXiv: 1515.03708 [quant-ph].
[19] Scott Aaronson. Scattershot BosonSampling: A new approach to scalable Boson-
Sampling experiments. 2013. url: http://www.scottaaronson.com/
blog/?p=1579 (visited on 12/22/2015).
[20] AP Lund et al. “Boson sampling from a Gaussian state”. In: Physical review letters
113.10 (2014), p. 100502.
[21] Scott Aaronson and Alex Arkhipov. “Bosonsampling is far from uniform”. In:
Quantum Information & Computation 14.15-16 (2014), pp. 1383–1423.
[22] Malte C Tichy et al. “Stringent and efficient assessment of boson-sampling de-
vices”. In: Physical review letters 113.2 (2014), p. 020502. arXiv: 1312.3080
[quant-ph].
[23] Andrea Crespi. “Suppression laws for multiparticle interference in Sylvester
interferometers”. In: Physical Review A 91.1 (2015), p. 013811. arXiv: 1502.
06372 [quant-ph].
[24] Malte Christopher Tichy et al. “Zero-transmission law for multiport beam split-
ters”. In: Physical review letters 104.22 (2010), p. 220405. arXiv: 1002.5038
[quant-ph].
[25] Andrea Crespi et al. “Quantum suppression law in a 3-D photonic chip implement-
ing the Fast Fourier Transform”. In: (2015). arXiv: 1508.00782 [quant-
ph].
110
[26] Ronen Barak and Yacob Ben-Aryeh. “Quantum fast Fourier transform and quan-
tum computation by linear optics”. In: JOSA B 24.2 (2007), pp. 231–240.
[27] James W Cooley and John W Tukey. “An algorithm for the machine calculation of
complex Fourier series”. In: Mathematics of computation 19.90 (1965), pp. 297–301.
[28] Michael A Nielsen and Isaac L Chuang. Quantum computation and quantum
information. Cambridge university press, 2010.
[29] Robert Solovay and Volker Strassen. “A fast Monte-Carlo test for primality”. In:
SIAM journal on Computing 6.1 (1977), pp. 84–85.
[30] Peter W Shor. “Polynomial-time algorithms for prime factorization and discrete
logarithms on a quantum computer”. In: SIAM journal on computing 26.5 (1997),
pp. 1484–1509. arXiv: quant-ph/9508027.
[31] Lane A Hemaspaandra. “Sigact news complexity theory column 36”. In: ACM
SIGACT News 33.2 (2002), pp. 34–47.
[32] Scott Aaronson. The Scientific Case for P != NP. 2014. url: http://www.
scottaaronson.com/blog/?p=1720 (visited on 11/30/2015).
[33] Scott Aaronson. “Guest column: NP-complete problems and physical reality”. In:
ACM Sigact News 36.1 (2005), pp. 30–52.
[34] Leslie G Valiant. “The complexity of computing the permanent”. In: Theoretical
computer science 8.2 (1979), pp. 189–201.
[35] Charles H Bennett and David P DiVincenzo. “Quantum information and compu-
tation”. In: Nature 404.6775 (2000), pp. 247–255.
[36] Ryszard Horodecki et al. “Quantum entanglement”. In: Reviews of Modern Physics
81.2 (2009), p. 865.
[37] Otfried Gühne and Géza Tóth. “Entanglement detection”. In: Physics Reports 474.1
(2009), pp. 1–75.
[38] Seth Lloyd. “Universal quantum simulators”. In: SCIENCE-NEW YORK THEN
WASHINGTON- (1996), pp. 1073–1077.
[39] Bryan T Gard et al. “An introduction to boson-sampling”. In: arXiv preprint
arXiv:1406.6767 (2014).
[40] Pieter Kok et al. “Linear optical quantum computing with photonic qubits”. In:
Reviews of Modern Physics 79.1 (2007), p. 135. arXiv: quant-ph/0512071.
[41] Henryk Minc. Permanents. Vol. 6. Cambridge University Press, 1984.
[42] MD Eisaman et al. “Invited review article: Single-photon sources and detectors”.
In: Review of Scientific Instruments 82.7 (2011), p. 071101.
111
[43] Brahim Lounis and Michel Orrit. “Single-photon sources”. In: Reports on Progress
in Physics 68.5 (2005), p. 1129.
[44] Christopher Gerry and Peter Knight. Introductory quantum optics. Cambridge
university press, 2005.
[45] Romain Alléaume et al. “Experimental open-air quantum key distribution with a
single-photon source”. In: New Journal of physics 6.1 (2004), p. 92.
[46] T Gaebel et al. “Stable single-photon source in the near infrared”. In: New Journal
of Physics 6.1 (2004), p. 98.
[47] E Wu et al. “Room temperature triggered single-photon source in the near in-
frared”. In: New Journal of Physics 9.12 (2007), p. 434.
[48] Satoshi Kako et al. “A gallium nitride single-photon source operating at 200 K”.
In: Nature materials 5.11 (2006), pp. 887–892.
[49] Andrew J Shields. “Semiconductor quantum light sources”. In: Nature photonics
1.4 (2007), pp. 215–223.
[50] Stefan Strauf et al. “High-frequency single-photon source with polarization
control”. In: Nature photonics 1.12 (2007), pp. 704–708.
[51] Markus Hennrich et al. “Photon statistics of a non-stationary periodically driven
single-photon source”. In: New Journal of Physics 6.1 (2004), p. 86.
[52] Christian Maurer et al. “A single-photon source based on a single Ca ion”. In:
New journal of physics 6 (2004), pp. 94–94.
[53] Mathias Steiner et al. “Highly efficient, tunable single photon source based on
single molecules”. In: Applied physics letters 18 (2007).
[54] Shuai Chen et al. “Deterministic and storable single-photon source based on a
quantum memory”. In: Physical review letters 97.17 (2006), p. 173004.
[55] S Cova et al. “Evolution and prospects for single-photon avalanche diodes and
quenching circuits”. In: Journal of Modern Optics 51.9-10 (2004), pp. 1267–1288.
[56] S Komiyama et al. “A single-photon detector in the far-infrared range”. In: Nature
403.6768 (2000), pp. 405–407.
[57] Kristine M Rosfjord et al. “Nanowire single-photon detector with an integrated
optical cavity and anti-reflection coating”. In: Optics Express 14.2 (2006), pp. 527–
534.
[58] Marius A Albota and Franco NC Wong. “Efficient single-photon counting at
1.55 µm by means of frequency upconversion”. In: Optics letters 29.13 (2004),
pp. 1449–1451.
[59] Hiroki Takesue et al. “1.5-µm single photon counting using polarization-independent
up-conversion detector”. In: Optics express 14.26 (2006), pp. 13067–13072.
112
[60] Robert Thomas Thew et al. “Low jitter up-conversion detectors for telecom
wavelength GHz QKD”. In: New Journal of Physics 8.3 (2006), p. 32.
[61] Kendall B Davis et al. “Bose-Einstein condensation in a gas of sodium atoms”. In:
Physical review letters 75.22 (1995), p. 3969.
[62] Mike H Anderson et al. “Observation of Bose-Einstein condensation in a dilute
atomic vapor”. In: science 269.5221 (1995), pp. 198–201.
[63] Jan Klaers et al. “Bose-Einstein condensation of photons in an optical microcavity”.
In: Nature 468.7323 (2010), pp. 545–548.
[64] CK Hong, ZY Ou, and Leonard Mandel. “Measurement of subpicosecond time
intervals between two photons by interference”. In: Physical Review Letters 59.18
(1987), p. 2044.
[65] Malte C Tichy. “Interference of identical particles from entanglement to boson-
sampling”. In: Journal of Physics B: Atomic, Molecular and Optical Physics 47.10
(2014), p. 103001. arXiv: 1312.4266 [quant-ph].
[66] Malte Christopher Tichy. “Entanglement and interference of identical particles”.
PhD thesis. Universitätsbibliothek Freiburg, 2011.
[67] VS Shchesnovich. “Asymptotic evaluation of bosonic probability amplitudes in
linear unitary networks in the case of large number of bosons”. In: International
Journal of Quantum Information 11.05 (2013), p. 1350045.
[68] David G Glynn. “The permanent of a square matrix”. In: European Journal of
Combinatorics 31.7 (2010), pp. 1887–1891.
[69] Lidror Troyansky and Naftali Tishby. “Permanent uncertainty: On the quantum
evaluation of the determinant and the permanent of a matrix”. In: Proc. 4th
Workshop on Physics and Computation (PhysComp’96). 1996.
[70] Annalisa Marzuoli and Mario Rasetti. “Towards a quantum algorithm for the
permanent”. In: International Journal of Quantum Information 5.01n02 (2007),
pp. 223–228.
[71] Nicolò Spagnolo et al. “General rules for bosonic bunching in multimode inter-
ferometers”. In: Physical review letters 111.13 (2013), p. 130503.
[72] VS Shchesnovich. “Sufficient condition for the mode mismatch of single photons
for scalability of the boson-sampling computer”. In: Physical Review A 89.2 (2014),
p. 022333.
[73] Peter P Rohde et al. “Will boson-sampling ever disprove the Extended Church-
Turing thesis?” In: arXiv preprint arXiv:1401.2199 (2014).
[74] Peter P Rohde and Timothy C Ralph. “Error tolerance of the boson-sampling
model for linear optics quantum computing”. In: Physical Review A 85.2 (2012),
p. 022332.
113
[75] Anthony A Leverrier and Raul Garcia-Patron Sanchez. “Analysis of circuit im-
perfections in BosonSampling”. In: Quantum information & computation 15.5-6
(2015), pp. 489–512.
[76] Alberto Peruzzo et al. “Quantum walks of correlated photons”. In: Science 329.5998
(2010), pp. 1500–1503. arXiv: 1006.4764 [quant-ph].
[77] C Gogolin et al. “Boson-sampling in the light of sample complexity”. In: arXiv
(2013). arXiv: 1306.3995 [quant-ph].
[78] Anthony Laing and Jeremy L O’Brien. “Super-stable tomography of any linear
optical device”. In: arXiv preprint arXiv:1208.2868 (2012).
[79] Michael Reck et al. “Experimental realization of any discrete unitary operator”.
In: Physical Review Letters 73.1 (1994), p. 58.
[80] Alex Arkhipov and Greg Kuperberg. “The bosonic birthday paradox”. In: Geome-
try & Topology Monographs 18 (2012), pp. 1–7.
[81] Rodica Simion and Frank W Schmidt. “On (+ 1,- 1)-matrices with vanishing
permanent”. In: Discrete Mathematics 46.1 (1983), pp. 107–108.
[82] Roberto Osellame et al. “Femtosecond writing of active optical waveguides with
astigmatically shaped beams”. In: JOSA B 20.7 (2003), pp. 1559–1567.
[83] Rafael R Gattass and Eric Mazur. “Femtosecond laser micromachining in trans-
parent materials”. In: Nature photonics 2.4 (2008), pp. 219–225.
[84] Thomas Meany et al. “Non-classical interference in integrated 3D multiports”.
In: Optics express 20.24 (2012), pp. 26895–26905.
[85] Konstantinos Poulios et al. “Quantum walks of correlated photon pairs in two-
dimensional waveguide arrays”. In: Physical review letters 112.14 (2014), p. 143604.
[86] G Della Valle, R Osellame, and P Laporta. “Micromachining of photonic devices
by femtosecond laser pulses”. In: Journal of Optics A: Pure and Applied Optics 11.1
(2009), p. 013001.
[87] P Törmä, I Jex, and S Stenholm. “Beam splitter realizations of totally symmetric
mode couplers”. In: journal of modern optics 43.2 (1996), pp. 245–251.
[88] Nicolò Spagnolo et al. “Quantum interferometry with three-dimensional geome-
try”. In: Scientific reports 2 (2012).
[89] Keith R Motes et al. “Linear Optical Quantum Metrology with Single Photons:
Exploiting Spontaneously Generated Entanglement to Beat the Shot-Noise Limit”.
In: Physical review letters 114.17 (2015), p. 170802.
[90] Saikat Guha. “Structured optical receivers to attain superadditive capacity and
the Holevo limit”. In: Physical review letters 106.24 (2011), p. 240502.
114
[91] Anthony Laing et al. “Observation of quantum interference as a function of
Berry’s phase in a complex Hadamard optical network”. In: Physical review letters
108.26 (2012), p. 260505.
[92] Ingemar Bengtsson et al. “Mutually unbiased bases and Hadamard matrices of
order six”. In: Journal of mathematical physics 48.5 (2007), p. 052106.
[93] Yuan Liang Lim and Almut Beige. “Multiphoton entanglement through a Bell-
multiport beam splitter”. In: Physical Review A 71.6 (2005), p. 062311.
115

thesis

  • 1.
    Universit`a degli Studidi Roma Tor Vergata MACROAREA DI SCIENZE MATEMATICHE FISICHE E NATURALI Corso di Laurea Magistrale in Fisica Tesi di laurea magistrale BosonSampling validation with integrated photonics Candidato: Luca Innocenti Matricola 0206284 Relatore Interno: Dott. Alessandro Cianchi Relatore Esterno: Prof. Fabio Sciarrino Anno Accademico 2014-2015
  • 2.
  • 3.
    List of Acronyms AAAaronson and Arkhipov . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 TM Turing machine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 ECT Extended Church-Turing thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 SPDC Spontaneous parametric downconversion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 HOM Hong-Ou-Mandel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 BS beam splitter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 ZTL Zero-Transmission Law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 MOL Mode Occupation List . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 MAL Mode Assignment List . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 BM binary matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75 qFFT Fast Fourier Transform. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .90 3
  • 4.
  • 5.
    Contents Introduction 7 1 Foundationsof classical and quantum information 11 1.1 Turing machine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1.2 Church-Turing thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 1.3 Complexity theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 1.4 Quantum information and quantum computation . . . . . . . . . . . . 20 2 Quantum and non-linear optics 23 2.1 Description of many-body states . . . . . . . . . . . . . . . . . . . . . . 23 2.2 Counting many-body states . . . . . . . . . . . . . . . . . . . . . . . . 30 2.3 Tools for quantum optics experiments . . . . . . . . . . . . . . . . . . . 33 2.3.1 Single-mode phase shift . . . . . . . . . . . . . . . . . . . . . . 33 2.3.2 Beam splitter . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 2.3.3 Wave plates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 2.3.4 Single-photon sources . . . . . . . . . . . . . . . . . . . . . . . 37 2.3.5 Single-photon detectors . . . . . . . . . . . . . . . . . . . . . . 39 2.4 Hong-Ou-Mandel effect . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 3 BosonSampling 43 3.1 Importance of BosonSampling . . . . . . . . . . . . . . . . . . . . . . 43 3.2 The BosonSampling computational problem . . . . . . . . . . . . . . . 45 3.3 Scaling experimental boson sampling implementations . . . . . . . . . 51 3.4 Scattershot boson sampling . . . . . . . . . . . . . . . . . . . . . . . . . 52 4 Boson sampling validation 55 4.1 Validating boson sampling implementations . . . . . . . . . . . . . . . 55 4.2 Fourier suppression law . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 4.3 Fourier suppression law and scattershot boson sampling . . . . . . . . 65 4.4 Sylvester suppression law . . . . . . . . . . . . . . . . . . . . . . . . . . 70 5
  • 6.
    5 Theoretical results75 5.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75 5.2 Main results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78 5.3 Bayesian validation scheme . . . . . . . . . . . . . . . . . . . . . . . . . 82 5.4 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 6 Experimental boson sampling validation with integrated photonics 89 6.1 Overview of the experiment . . . . . . . . . . . . . . . . . . . . . . . . 90 6.2 Femtosecond laser writing . . . . . . . . . . . . . . . . . . . . . . . . . 90 6.3 Realization of the quantum Fast Fourier Transform . . . . . . . . . . . 92 6.4 Photon generation and manipulation . . . . . . . . . . . . . . . . . . . 94 6.5 One- and two-photon measurements in integrated Fourier interferometers 94 6.6 Observation of the suppression law . . . . . . . . . . . . . . . . . . . . 100 6.7 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 Conclusions 105 6
  • 7.
    Introduction Since the early1980s, it has been argued [1] that simulating quantum systems is a very challenging task. One source of difficulty is the number of parameters needed to characterize a generic quantum system, which grows exponentially with the size of the system. This means that even only storing the state of a large quantum system is not feasible with classical computer memories. Furthermore, the number of operations needed to simulate the temporal evolution of such a system also scales exponentially with the size. Thus, the only way to avoid this exponential overhead in the evolution is the use of approximation methods (such as Monte Carlo methods). However, for many problems of interest, no good approximation scheme are available. It is then still an open problem whether they can be efficiently simulated with a classical approach. Hence, it is widely accepted that classical systems cannot in general efficiently simulate quantum systems. While it is not yet possible to prove it, neither mathematically nor experimentally, there are strong evidences to believe that this is the case. This distinction between classical and quantum world has many implications. One of the most notable examples concerns the possibility that computers exploiting the weirdnesses of quantum mechanics may be able to carry out computations impossible with only classical resources. With the current available technologies, the experimental observation of this quantum advantage (sometimes referred to as quantum supremacy [2, 3]) has proven itself to be rather difficult to achieve. In particular, to observe a post- classical computation with a universal quantum computer one first needs to solve the problem of fault-tolerant quantum computation [4], which is known to be possible in principle [5, 6, 7], but might require decoherence rates that are several orders of magnitude below what achievable today. In the case of linear optics, a number of no-go theorems led to the widespread belief that linear interferometry alone could not provide a path to universal quantum computation. For this reason the result of Aaronson and Arkhipov (AA), that passive linear optical interferometers with many-photon inputs cannot be efficiently simulated by a classical computer [8], represented a significant advance. The related computational problem, that is, sampling from the output proba- bility distribution of such an apparatus, was named by AA the BosonSampling problem. A quantum device able to efficiently solve it is referred to as a boson sampler. More in detail, the BosonSampling computational problem consists in sampling 7
  • 8.
    from the outputprobability distribution resulting from the time-evolution of n indis- tinguishable photons into a random m × m unitary transformation. The hardness of BosonSampling arises from the fact that the scattering amplitude between an input and an output state configuration is proportional to the permanent of a suitable n×n matrix, where the permanent is a particular function, defined similarly to the determinant, which in the general case is known to be hard to compute classically. This immediately suggests an experimental scheme to build a boson sampler using only linear optical elements: just inject n indistinguishable photons into an appropriate linear optical interferometer, and use photon-counting detectors to detect the resulting output states. AA showed that, already with 20 < n < 30 and m n, this would provide direct evidence that a quantum computer can solve a problem faster than what is possible with any classical device. While this regime is far from our current technological capabilities, several implementations of 2- and 3-photon devices have soon been reported [9, 10, 11, 12], and other more complex implementations followed [13, 14, 15, 16, 17, 18]. However, the originally proposed scheme to implement BosonSampling, that is, to generate the input n-photon state through Spontaneous parametric downconver- sion (SPDC), suffers from scalability problems. Indeed, it is unfeasible to generate high numbers of indistinguishable input photons with this method, due to the generation probability decreasing exponentially with n. For this reason, an alternative scheme, named scattershot boson sampling [19], has been devised [20], and subsequently im- plemented [18]. Contrarily to a classical boson sampler, a scattershot boson sampler uses m SPDC sources, one per input mode of the interferometer, to generate random (but known) n-photon input states, with n m. Each SPDC source generates a pair of photons, one of which is injected into the interferometer, while the other is used to herald the SPDC generation event. The use of a scattershot boson sampling scheme results in an exponential increase of the probability of generating n indistinguishable photons, for m and n large enough. While the key part of BosonSampling resides in its simulation complexity, this very hardness also poses a problem of certification of such a device. Indeed, it is believed [8] that, when n is large enough, a classical computer cannot even verify that the device is solving BosonSampling correctly. However, it is still possible to obtain circumstantial evidence of the correct functioning of a device, and efficiently distinguish the output of a boson sampler from that resulting from alternative probability distributions, like the output produced by classical particles evolving through the same interferometer. A number of validation schemes were subsequently devised to validate the output resulting from true many-boson interference [14, 16, 21, 22, 23]. In particular, the tests currently more suitable to identify true many-body interference [22] are those based on Zero-Transmission Laws (ZTLs) [24]. A ZTL, also often referred to as suppression law, is a rule which, for certain particular unitary evolutions, is able to predict that the probability of certain input-output configurations is exactly zero, without having to 8
  • 9.
    compute any permanent. However,for the current validation schemes based on ZTLs, it is mandatory that the input states possess particular symmetries. This requirement may thus be an issue when the ZTLs are applied to validate a scattershot boson sampler. Indeed, the input state in scattershot boson sampling is not fixed, but changes randomly at each n-photon generation event. This mandates for a new ZTL-based validation scheme to be devised, able to efficiently validate a scattershot boson sampling experiment, but still keeping the capability of distinguishing alternative probability distributions. In this thesis we report on both theoretical and experimental advances in the context of validating classical and scattershot boson sampling experiments: • From the theoretical point of view, we devise a new validation scheme, more suitable to validate scattershot boson sampling experiments. This scheme, based on a ZTL valid for a particular class of matrices, the so-called Sylvester matrices, generalizes the ZTL reported in [23], presenting significantly higher predictive capabilities. • From the experimental point of view, we report on the experimental implementa- tion [25] of the validation scheme proposed in [22] based on the ZTL for Fourier matrices [24]. To this end, a scalable methodology to implement the Fourier transformation on integrated photonics was adopted. This approach exploits the 3-D capabilities of femtosecond laser writing technique, together with a recently proposed [26] quantum generalization of the Fast Fourier transform algorithm [27], which allows a significant improvement in the number of elementary optical elements required to implement the desired Fourier transformation. The thesis is structured as follows: Chapter 1 opens with a brief survey of classical and quantum computer science. In chapter 2, after a brief exposition of the theoretical formalism for many-body quantum states, the fundamental tools used in quantum op- tics experiments are presented. In chapter 3 the BosonSampling problem is introduced. The problem of scaling boson sampling experiments is discussed, together with the recently proposed alternative scheme named scattershot boson sampling. In chapter 4 the subject of boson sampling validation is introduced, and an outline of the proposed solutions is provided. In particular, the focus is on the validation schemes based on zero-transmission laws for Fourier and Sylvester matrices, and the possibility of apply- ing them to scattershot boson sampling experiments. In chapter 5 we present a new zero-transmission law for Sylvester matrices. Exploiting this zero-transmission law, we present a scheme to validate scattershot boson sampling experiments. The thesis closes with chapter 6, where we present the experimental implementation of a validation scheme for Fourier matrices. The experimental and technological aspects of the experi- ment are discussed, from the femtosecond laser-written technology employed to build the integrated interferometers, to a novel method to efficiently implement the Fourier 9
  • 10.
    transform on anintegrated photonics chip. A full analysis of the chip reconstruction and the observed suppression effects follows. The chapter closes with a discussion of the usefulness of the presented work, and the possible future improvements. The work presented in this thesis was carried out at the Quantum Information Lab, University of Rome La Sapienza. 10
  • 11.
    Chapter 1 Foundations ofclassical and quantum information Quantum information theory is a relatively recent field, and most of the main concepts have been developed only in the last few decades. It is therefore natural that quantum information relies heavily on ideas developed in the context of classical information theory. These provide on one side tools that can be adapted to the quantum realm to tackle quantum information problems, and on the other a benchmark to measure the advantages provided by quantum resources with respect to their classical counterparts. To meaningfully talk about the efficiency of an algorithm in performing a given task, both in the classical and quantum context, it is necessary to have a mathematically precise notion of what an “algorithm” is, and a mean to quantify its efficiency in solving a given computational problem. The formalization of the concept of an algorithm requires, in turn, the introduction of a computational model. One of the most important computational models is the so-called Turing machine (TM) model of computation, discussed in section 1.1. While a variety of other computational models have been introduced and studied over the years, we will see that, thanks to the Church-Turing thesis, it is enough to just consider the TM model in defining what algorithms can and cannot do. Indeed, until a few decades ago, a much stronger conjecture was believed to be true: the so-called Extended Church-Turing thesis states that any physically plausible model of computation can be efficiently simulated by a TM. This last conjecture is however currently under debate, as we now have reasons to believe that computational devices able to solve certain problems exponentially faster than any classical computer can indeed be devised. The Church-Turing thesis, as well as its extended version, is discussed in section 1.2. To meaningfully assess whether an algorithm is more efficient than another, it is necessary to introduce the idea of complexity classes, studied in the field of computational complexity theory. These are classes of algorithms sharing some common scaling properties, and are discussed in section 1.3. 11
  • 12.
    1.1 Turing machine Whilethe concept of algorithm as a sequence of operations aimed to obtain a given result may seem intuitively obvious, a mathematically precise formulation of the concept was only given in 1930s, thanks to the work of Alonzo Church, Alan Turing, and other pioneers of the computer era. This work resulted in the development of what may arguably be considered as the most important computational model of computer science: the Turing machine (TM) model. A TM captures the notion of an algorithm performing a computational task, and is composed of four main elements: 1. a program, which similarly to an ordinary computer program is a sequence of program lines, each one describing an elementary operation for the TM, 2. a finite state control, which co-ordinates the other operations of the machine, similarly to a modern computer CPU, 3. a tape, which acts like a computer memory, 4. a read-write tape-head, pointing to the position on the tape which is currently readable or writable. The finite state control for a TM consists of a finite set of internal states, q1, . . . , qm. The number of states m can be varied, however it turns out that for m sufficiently large this change does not affect the power of the machine in any essential way, so without loss of generality we may suppose that m is some fixed constant. The finite state control can be thought of as a sort of microprocessor, co-ordinating the TM’s operation. It provides temporary storage off-tape, and a central place where all processing for the machine may be done. In addition to the states q1, . . . , qm, there are also two special internal states, labelled qs and qh. We call these the starting state and the halting state, respectively. The idea is that at the beginning of the computation, the TM is in the starting state qs. The execution of the computation causes the TM’s internal state to change. If the computation ever finishes, the TM ends up in the state qh to indicate that the machine has completed its operation. The TM tape is a one-dimensional object, which stretches off to infinity in one direction. The tape consists of an infinite sequence of tape squares numbered starting from 0. Each tape square contains one symbol drawn from an alphabet, Γ, composed of a finite number of distinct symbols. The read-write tape-head identifies a single square on the TM tape as the square that is currently being accessed by the machine. Summarizing, a TM starts its operation with the finite state control in the state qs, and with the read-write head at the leftmost tape square. The computation then proceeds step by step according to a predefined program. The computation is halted 12
  • 13.
    when the currentstate is qh, and the output of the computation is the current contents of the tape. A program for a TM is a finite ordered list of program lines of the form q, x, q , x , s , where q, q are internal states of the machine, x, x are symbols of the alphabet Γ, and s is equal to −1, +1, or 0. At any given step of the computation, if q and x are the current internal state and the symbol under the read-write tape-head respectively, the TM looks through the list of program lines in order, searching for a line of the form q, x, ·, ·, · . If it does not find such a line, the internal state is changed to qh and the machine halts. If such a line is found, than that program line is executed: the internal state is changed to q , the current symbol on the tape is overwritten by the symbol x , and the tape-head moves left, right, or stands still, depending on whether s is −1, +1, or 0, respectively. Example 1 (Computation of the constant function f(x) = 1). Consider a TM with three internal states, q1, q2, and q3, in addition to the starting state qs and the halting state qh. The used alphabet will contain four symbols: Γ = { , b, 0, 1}, with the zeros and ones used to denote the input number, the to mark the beginning of the input number, and the blanks (b) used on all the other cells of the tape. Finally, we program the TM with the following program lines: 1 : qs, , q1, , +1 , 2 : q1, 0, q1, b, +1 , 3 : q1, 1, q1, b, +1 , 4 : q1, b, q2, b, −1 , 5 : q2, b, q2, b, −1 , 6 : q2, , q3, , +1 , 7 : q3, b, qh, 1, 0 . If we start the TM with the tape containing a number expressed in binary form, surrounded by blank (b) tapes, with the symbol marking the beginning of the input number, and the tape-head starting on the cell containing , like the following: b b b 1 1 1 0 1 b b b b qs start . . .. . . we will get as output the following state: 13
  • 14.
    b b b1 b b b b b b b b qh end . . .. . . To see this, we have to analyse how the program given to the TM acts on the initial tape: starting in the state qs on a tape cell with the symbol , the first line of the program mandates the tape-head to move right and switch to the state q1. In the q1 state, following the lines 2 and 3, the tape-head will move right over-writing all the ones and zeros it finds, until it reaches a blank cell. When a blank cell is reached, according to line 4, the tape-head changes its state to q2 and starts moving left, continuing moving left, following line 5, until it reaches again the cell. When the cell is reached, the state is changed to q3 and the tape-head is moved once on the right. At this point, because of the line 7 of the program, the cell under the tape-head - that is, the cell on the right of the one with - is over-written to 1, and the tape-head state changed to qh, halting the execution. The above analysis shows that this program computes the constant function f(x) = 1. That is, regardless of what number is given in input onto the tape, the TM halts with the number 1 represented onto the tape. In general, a TM can be thought of as computing functions from the non-negative integers to the non-negative integers, with the input to the function represented by the initial state of the tape, and the output of the function by the final state of the tape. The above presented TM is used to formalize the concept of a deterministic algorithm. To also consider non-deterministic algorithms, this model must however be extended. For this purpose, the TM model is generalized to that of a probabilistic TM. In a probabilistic TM, the state transitions are choosen according to some probability distribution, instead of being completely predetermined. A further generalization of TMs provides a theoretical basis for quantum algorithms. These are a special kind of algorithms which, exploiting the properties of quantum mechanics, can potentially outperform any classical algorithm in certain tasks. 1.2 Church-Turing thesis An interesting question is what class of functions is it possible to compute using a Turing machine. Despite its apparent simplicity, the TM model can be used to simulate all the operations performed on a modern computer. Indeed, according to a thesis put forward independently by Church and Turing, the TM model completely captures the notion of computing a function using an algorithm. This is known as the Church-Turing thesis [28]: 14
  • 15.
    Church-Turing thesis: Theclass of functions computable by a Turing ma- chine corresponds exactly to the class of functions which we would naturally regard as being computable by an algorithm. The Church-Turing thesis asserts an equivalence between the rigorous mathematical concept of “function computable by a Turing machine”, and the intuitive concept of what it means for a function to be computable by an algorithm. In this sense it is nothing more than a definition of what we mean when we talk of the “computability” of a function. This thesis is relevant because it makes the study of real-world algorithms amenable to rigorous mathematical analysis. We remark that it is not obvious that every function which we would intuitively regard as computable by an algorithm can be computed using a TM. Indeed, it is conceiv- able that in the future we will discover in Nature a process which computes a function not computable by a TM. Up to now, however, no such process has been observed. Indeed, as will be discussed in more detail in later sections, quantum computers also obey the Church-Turing thesis. That is, quantum computers can compute the same class of functions computable by a TM. A much stronger statement than the Church-Turing thesis is the so-called Extended Church-Turing thesis (ECT) [8] (also sometimes referred to as Strong Church-Turing thesis [28]): Extended Church-Turing thesis: All computational problems that are efficiently solvable by realistic physical devices, are efficiently solvable by a Turing machine. The ECT was however already found to be insufficient to capture all realistic com- putational models in the 1970s, when Solovay and Strassen [29] devised an efficient, probabilistic primality test. As the Solovay-Strassen algorithm relied essentially on randomness, it provided the first evidence that probabilistic Turing machines are capa- ble to solve certain problems more efficiently than deterministic ones. This led to the following ad-hoc modification to the ECT: Extended Probabilistic Church-Turing thesis: All computational prob- lems that are efficiently solvable by realistic physical devices, are efficiently solvable by a probabilistic Turing machine. As this is the form the ECT is currently usually stated as, this is the version we will refer to when talking in the following of “ECT”. However, even in this modified form, the ECT still does seem to be in contrast with the currently accepted physical laws. The first evidence in this direction was given by Shor [30], which proved that two very important problems - the problem of finding the prime factors of an integer, and the so-called discrete logarithm problem - could be solved efficiently on a quantum computer. Since no efficient classical algorithm - neither deterministic nor probabilistic - is currently known to be able to efficiently solve 15
  • 16.
    these problems, Shor’salgorithm strongly suggests that quantum mechanics allows to solve certain problems exponentially faster than any classical computer, and this directly contradicts the ECT. 1.3 Complexity theory Computational complexity theory analyzes the time and space resources required to solve computational problems [28]. Generally speaking, the typical problem faced in computational complexity theory is proving some lower bounds on the resources required by the best possible algorithm for solving a problem, even if that algorithm is not explicitly known. One difficulty in formulating a theory of computational complexity is that different computational models may lead to different resource requirements for the same problem. For instance, multiple-tape TMs can solve many problems significantly faster than single- tape TMs. On the other hand, the strong Church-Turing thesis states that any model of computation can be simulated on a probabilistic TM with at most a polynomial increase in the number of elementary operations required. This means that if we make the coarse distinction between problems which can be solved using resources which are bounded by a polynomial in n, and those whose resource requirements grow faster than any polynomial in n, then this distinction will be well-defined and independent of the considered computational model. This is the chief distinction made in computational complexity. With abuse of the term exponential, the algorithms with resource requirements growing faster than any polynomial in n are said to require an amount of resources scaling exponentially in the problem size. This includes function like nlog n , which grow faster than any polynomial but lower than a true exponential, and are nonetheless said to be scaling exponentially, in this context. A problem is regarded as easy, tractable, or feasible, if an algorithm for solving the problem using polynomial resources exists, and as hard, intractable, or infeasible, if the best possible algorithm requires exponential resources. Many computational problems are formulated as decision problems, that is problems with a yes or no answer. For example, the question is a given number m a prime number or not? is a decision problem. Although most decision problems can easily be stated in simple, familiar language, discussions of the general properties of decision problems are greatly helped by the terminology of formal languages. In this terminology, a language L over the alphabet Σ is a subset of the set Σ∗ of all finite strings of symbols from Σ. For example, if Σ = {0, 1}, then the set of binary representations of even numbers L = {0, 10, 100, 110, . . . } is a language over Σ. A language L is said to be decided by a TM if for every possible input x ∈ Σ∗ , the TM is able to decide whether x belongs to L or not. In other words, the language L is decided if the TM will eventually halt in a 16
  • 17.
    state encoding a“yes” answer if x ∈ L, and eventually halt to a state encoding a “no” answer otherwise. Decision problems are naturally encoded as problems about languages. For instance, the primality decision problem can be encoded using the binary alphabet Σ = {0, 1}, interpreting strings from Σ∗ as non-negative integers, and defining the language L to consist of all binary strings such that the corresponding number is prime. The primality decision problem is then translated to the problem of finding a TM which decides the language L. More generally, to each decision problem is associated a language L over an alphabet Σ∗ , and the problem is translated to that of finding a TM which decides L. To study the relations between computational problems, it is useful to classify them into complexity classes, each one grouping all problems (that is, in the case of decision problems, all languages) sharing some common properties. Most of computational complexity theory is aimed at defining various complexity classes, and at understanding of the relationships between different complexity classes. A brief description of the most important complexity classes for decision problems is provided in the following: • P: We say that a given problem is in TIME(f(n)) if there is a deterministic TM which decides whether a candidate x is in the corresponding language in time O(f(n)), with n the length of x. A problem is said to be solvable in polynomial time if it is in TIME(nk ) for some k. The collection of all languages which are in TIME(nk ), for some k, is denoted P, which is an example of a complexity class Some examples of problems in P are linear programming, the calculation of the greatest common divisors of two numbers, and the problem of determining if a number is prime or not. Not surprisingly, there are lots of problems for which no polynomial-time algo- rithm is known. Proving that a given decision problem is not in P, however, is very difficult. A couple of examples of such problems are 1) given a non-deterministic Turing machine M and an integer n written in binary, does M accept the empty string in at most n steps? and 2) given a pair of regular expressions, do they represent different sets?. Many other problems are believed to not be in P. Among these are notable ones such as Factoring, which is the problem of finding the prime factors decomposition of an integer. This problem is believed to hard problem for classical computers, though no proof, nor compelling evidences for it, are known to date. Factoring is particularly important, since its hardness lies at the heart of wisely used algorithms in cryptography such as the RSA cryptosystem [28]. • NP: An interesting property of the prime factorization problem is that, even if finding the prime factorization of an integer n is very hard, it is easy to check if a proposed set of primes is indeed the correct factorization of n: just multiply the numbers and check if they equal n. The class of decision problems sharing this 17
  • 18.
    property is calledNP. More generally NP, standing for “nondeterministic poly- nomial time”, is the class of all decision problems for which there are efficiently verifiable proofs. A NP problem can often be intuitively stated in the form are there any solutions that satisfy certain constraints? While it is clear that P is a subset of NP, the converse is currently not known. Indeed, whether P equals NP is arguably the most famous open problem in computer science, often abbreviated as the P = NP problem. Many computer scientists believe [31, 32, 33] that P = NP. However, despite decades of work, nobody has been able to prove this, and the possibility that P = NP cannot be excluded. Some implications of either of these possibilities are shown in fig. 1.1. A related complexity class is NP-hard, which groups all decision problems that are, informally, at least as hard as the hardest problems in NP. More precisely, a problem L is NP-hard if every problem in NP can be reduced to L in polynomial time. As a consequence, a polynomial algorithm solving an NP-hard would also automatically provide a polynomial algorithm for all problems in NP. While this is considered highly unlikely, as many NP problems are believed to not be solvable in polynomial time, it has never been proved that this is not the case. Finally, the intersection between NP and NP-hard is the class of the so-called NP-complete problems. • BPP: If we extend our definition of a TM allowing it to have access to a source of randomness, let’s say the ability to flip a fair coin, other complexity classes can be defined. Such a probabilistic Turing machine may only accept or reject inputs with a certain probability, but if the probability of an incorrect accept or reject is low enough, they are as useful as their deterministic counterparts. One of the most important such classes is BPP, which stands for Bounded-error Probabilistic Polynomial time. BPP is the class of decision problems solvable by a probabilistic TM in polynomial time with a probability of error less than 1/3. The choice of 1/3 as error bound is mostly arbitrary, as any error bound strictly less than 1/2 can be reduced to practically zero with only a small increase in the resource requirements. For this reason, problems in BPP are regarded as as efficiently solvable as P problems. In fact, for practical purposes, BPP is considered, even more than P, as the class of problems which are efficiently solvable on a classical computer. 18
  • 19.
    Complexity P ≠ NPP = NP NP-Hard NP-Complete P NP NP-Hard P = NP = NP-Complete Figure 1.1: Relations between the fundamental complexity classes. All the above considered computational classes only took into account classical Turing machines. The advent of quantum mechanics and the conception of quantum computers, however, led to the question of what classes of problems can a quantum computer solve? To try to answer this question, one must study another kind of complexity classes, entering the realm of quantum complexity theory. In this context, arguably the most fundamental complexity class if BQP, standing for Bounded-error Quantum Polynomial time. This is the quantum generalization of BPP, and is defined as the set of decision problems solvable by a quantum computer in polynomial time, with an error probability of at most 1/3 for all instances. Probably the most notable problem which has been shown to be in BQP is Factoring. Indeed, Shor’s algorithm [30] was one of the first devised quantum algorithms able to efficiently solve a problem that the best-known classical counterparts can solve only in exponential time. While only decision problems have been mentioned to this point, these are not the only kind of computational problems. Function problems are a generalized version of decision problems, where the output of the algorithm is not bounded to be a simple YES/NO answer. More formally, a function problem P is defined as a relation R over the cartesian product over strings of an alphabet Σ, that is R ⊂ Σ∗ × Σ∗ . An algorithm is said to solve P if for every pair (x, y) ∈ R, it produces y when given x as input. A class of function problems that will be of interest in the following are the so-called counting problems, which are problems that can be stated as how many X satisfy a certain 19
  • 20.
    property P? Anexample of such a complexity class is #P (pronounced “sharp P”), which is the set of counting problems associated with the decision problems in NP. Intuitively, to each NP problem which can be stated in the form “Are there any solutions having the property P?” is associated a #P problem which can be stated in the form “How many solutions are there which satisfy the property P?”. As can be intuitively deduced from this definition, #P problems are generally believed to be even more difficult than NP problems. A pair of computational classes related to #P are #P-hard and #P-complete. These are defined in a completely analogous way to NP-hard and NP-complete, containing the class of counting problems at least as hard as any #P problem. A notable instance of a #P-hard problem is the calculation of the permanent of a complex-valued matrix. The permanent is a function of matrices defined similarly to the determinant (see the discussions in the following sections, like definition 2), but which, contrarily to the latter, is known to not be computable in polynomial time, for general matrices [34]. Indeed, the problem of computing the permanent of a matrix is known to be a #P-hard problem, and even #P-complete in special circumstances [34]. 1.4 Quantum information and quantum computation Quantum information theory [28, 35, 36, 37] is the study of the information processing tasks that can be accomplished using quantum mechanical systems. One of the main goals of quantum information theory is to investigate how information is stored in the state of a quantum system, how does it differ from that stored in a classical system, and how can this difference be exploited to build quantum devices with capabilities superior to that of their classical counterparts. To this end several concepts and ideas are drawn from other disciplines, such as quantum mechanics, computer science, information theory, and cryptography, and merged with the goal of generalizing the concepts of information and computing to the quantum realm. In the last few decades, information and computation theory have undergone a spurt of new growth, expanding to treat the intact transmission and processing of quantum states, and the interaction of such quantum information with traditional forms of information. We now know that a fully quantum theory of information offers, among other benefits, a brand of cryptography whose security rests on fundamental physics, and a reasonable hope of constructing quantum computers that could dramatically speed- up the solution of certain mathematical problems. Moreover, at a more fundamental level, it has become clear that an information theory based on quantum principles extends and completes classical information theory, much like complex numbers extend and complete the reals. One of the conceptual building blocks of quantum information and quantum computation is that of a qubit. This is the quantum generalization of the classical concept of bit, and the fundamental processing unit of most quantum 20
  • 21.
    devices. While abit can be in one of two states, traditionally referred to as 0 and 1, a qubit is allowed to be in a superposition of these basis states. Properly handling such qubits, quantum computers are able to process information in ways impossible with any classical computer. The first to envisage the notion of a quantum computer was Feynman [1], as a possible solution to the problem of the exponentially increasing amount of resources required to simulate complex quantum systems with classical computers. More than a decade later, Lloyd [38] showed that a quantum computer can indeed act as a universal quantum simulator, where the word universal refers to the fact that the same machine is capable of tackling vastly different problems by simply changing the program it runs. There are a lot of candidate implementations for quantum computation. Among these, in no particular order, are implementations using superconductors, trapped ions, quantum dots, nuclear magnetic resonance, diamond nitrogen vacancies, silicon, linear optics, and many other proposed technologies. Here we will only focus on linear optical implementations of quantum computing, to highlight the difficulties inherent to implement universal quantum computers, as opposite to the relatively much easier demands of boson sampling devices, which will be described in the following sections. Linear optics quantum computation (LOQC) with single photons has the advan- tage that photons have very long decoherence times, which means that the quantum information stored in photons tends to stay there, and that linear optical elements are arguably the simplest building blocks to realize quantum information processing. The downside is that photons do not naturally interact with each other, and in order to apply two-qubit quantum gates, which are necessary to implement universal quantum computation, such interactions are essential. Because of this, effective interactions among photons have to be introduced somehow. The two main methods to implement such interactions among photons are 1) using Kerr nonlinearities, and 2) the use of projective measurements with photodetectors. Unfortunately, present-day nonlinear Kerr media exhibit very poor efficiency [28] and very weak nonlinearities, while projective measurements have the disadvantage of producing probabilistic quantum gates: more often than not these gates fail, destroying the quantum information. In the case of projective-measurements-induced nonlinearities there is however a way to avoid the issue of nondeterministic gates, still mantaning feasible resource requirements: the Knill, Laflamme, and Milburn [7] (KLM) scheme. Introduced in 2001, the KLM protocol allows scalable linear optics quantum computing by using quantum gate teleportation to increase the probability of success of nondeterministic gates [7, 39, 40]. The downside of the KLM scheme is that, for its implementation, it is still necessary to overcome a series of experimental challenges, such as the synchronization of pulses, mode-matching, quickly controllable delay lines, tunable beam splitters and phase shifters, single-photon sources, accurate, fast, single-photon detectors, and extremely 21
  • 22.
    fast feedback controlof these detectors. While most of these features are not terribly unrealistic to implement, the experimental state of the art is simply not at the point at which more complex gate operations such as two-qubit operations can be implemented. On the other hand, a quantum computer is not necessary to implement quantum simulation. Dropping the requirement of being able to simulate any kind of system, special purpose devices can be built to tackle specific problems better than the clas- sical counterparts, in the simpler conceivable case by just emulating, in an analog manner, the behaviour of a complex quantum system on a simpler quantum device. Being generally these special purpose devices easier to implement than full-fledged quantum computers, it is expected that practical quantum simulation will become a reality well before quantum computers. However, despite the undeniable practical usefullness of implementing quantum simulation on a classically intractable quantum system, this would hardly give a definite answer to the question: are there tasks which quantum computers can solve exponentially faster than any classical computer? Indeed, a quantum system that is hard to classically simulate is also typically hard to define as a computational problem. This makes extremely difficult to definitively prove whether a classical algorithm, able to efficiently carry out such a simulation, exists. It is for this reasons that the proposal of a boson computer by Aaronson and Arkhipov [8] gained much interest in the quantum optics community. This kind of special purpose linear optics quantum computer requires only to send n indistinguishable photons through a random unitary evolution, and detect the output photons with standard photodetectors. No teleportation or feedback mechanisms are required, which makes the experimental implementation of such a device much easier than that of a quantum computer following the KLM scheme. Furthermore, the related computational problem is simple enough to be analytically tractable with the tools of computational complexity, allowing to obtain very strong theoretical evidence of its hardness. 22
  • 23.
    Chapter 2 Quantum andnon-linear optics In section 2.1 the formalism required to describe many-body quantum states is intro- duced. In particular, the focus is on many-boson states, and their evolution through a unitary transformation. In section 2.2 a derivation of the total number of physical states according to the bosonic, fermionic, and classical statistics is presented. In sec- tion 2.3 the focus shifts to a more experimental point of view, and the main tools used in quantum optics experiments are presented. The chapter closes with an outline of the Hong-Ou-Mandel effect in section 2.4, which is a striking evidence of how quantum interference results in highly non-classical consequences. 2.1 Description of many-body states To describe a quantum mechanical system with a fixed number of particles, it is enough to use a ket state |Ψ(t) , corresponding to a wave function Ψ(r1, . . . , rn, t) ≡ r1, . . . , rn|Ψ(t) . This is interpreted through Born’s rule as the probability amplitude of the i-th particle being found at the position ri at the time t, that is, as a function such that |Ψ(r1, . . . , rn, t)|2 = probability of finding the i-th particle at the position ri, at the time t. (2.1) Such a wave function has to satisfy certain symmetrization rules, depending on the statistical nature of the particles described, and its time-evolution is characterized by the Schrödinger equation, i ∂ ∂t |Ψ(t) = H |Ψ(t) , (2.2) where H is the Hamiltonian of the system. The Hamiltonian is an operator, correspond- ing to the total energy of the system in most of the cases, which characterizes the dynamics of the system. Equation (2.2), together with an initial condition |Ψ(t0) , is 23
  • 24.
    sufficient to obtainthe wave function at each time t: its solution is, at least formally, given by |Ψ(t) = e−i(t−t0)H/ |Ψ(t0) . The above described formalism is also called first-quantization, to distinguish it from another way of dealing with quantum systems, named second-quantization. The latter differs from the former by a shift in focus: instead of considering the number of particles as a fixed property of the system and using the wave function to describe their states, the system is characterized by the number of particles contained in each possible mode, which are however no longer necessarily fixed. To this end, a creation operator is defined, for each mode, as an operator which acts on a quantum state and produces another quantum state differing from the former for a single quantum added to that mode. The hermitian conjugate of a creation operator is called an annihilation operator (also destruction operator), and instead of adding a quantum to a given mode, it does the opposite, producing a new state with one less particle in the mode. The exact rules followed by creation and annihilation operators depend on the statistical nature of the particles. For bosons, the creation (annihilation) operator of a mode labelled i is denoted ˆa† i (ˆai). The core rules obeyed by these operators are the following: ˆai |ni = √ n |(n − 1)i , ˆa† i |ni = √ n + 1 |(n + 1)i , [ˆak, ˆa† q] = δk,q, [ˆak, ˆaq] = [ˆa† k, ˆa† q] = 0 , (2.3) where |ni is a state with n particles in the mode labelled i. Quantum states with a well-defined number of particles in each mode are called Fock states, or number states, and the set of all Fock states is called Fock space. The above defined creation operators can be used to denote many-boson states, that is, quantum states with more than one indistinguishable boson. If any single boson can be in one of m modes, an n-boson state r having ri particles in the i-th mode will be written as |r ≡ |r1, . . . , rm = 1 √ r! m k=1 ˆa† k rk |0 . (2.4) where the notation r! ≡ r1! · · · rk! has been used. Such a list r of m elements, with each element equal to the number of particles in a given mode, will be referred to as the Mode Occupation List (MOL) associated to the quantum state. A many-body state such that for every k = 1, . . . , m, rk = 1 or rk = 0, is said to be a collision-free state. Another way to characterize many-body states is through a so-called Mode Assign- ment List (MAL) R . This is a list of n elements, with the i-th element being the mode occupied by the i-th particle. It is worth noting that for indistinguishable particles one cannot talk of “the mode of the i-th particle”, however. Because of this, the order of the elements of MALs cannot have any physical significance. In other words, MALs are always defined up to the order of the elements, or, equivalently, they must be always considered conventionally sorted (for example, in increasing order). Representing the 24
  • 25.
    state with aMAL, eq. (2.4) can be rewritten in the following form: |r ≡ |R = 1 √ r! n k=1 ˆa† Rk |0 = 1 µ(R) n k=1 ˆa† Rk |0 , (2.5) where we denoted with µ(R) the product of the factorials of the occupation numbers of the state R, that is, µ(R) ≡ r1! · · · rm!. Yet another way to describe many-body states that will sometimes be useful is to explicitly list the occupation numbers of each mode. For example, for a state with three particles, one in the second mode and two in the fourth mode, we write |12, 24 . If we want to emphasize the absence of particles in, say, the third mode, we write it as |12, 03, 24 . Definition 1 (MOL and MAL representations). All of the many-particle quantum states used in the following will be assumed to have a fixed number of particles n, with each particle potentially occupying one of m possible modes. Two ways to represent a state are: • As the Mode Occupation List (MOL) r ≡ (r1, . . . , rm), i.e. as the m-dimensional vector whose element rk is the number of particles in the k-th mode. It follows from this definition that m k=1 rk = n. We will refer to this representation as the MOL representation, and denote with Fn,m the set of all MOLs of n photons into m modes, and with FCF n,m the set of collision-free MOLs of n photons into m modes: Fn,m ≡ (r1, . . . , rm) | ∀i = 1, . . . , m, ri ≥ 0 and m i=1 ri = n , (2.6) FCF n,m ≡ {(r1, . . . , rm) ∈ Fn,m | ∀i = 1, . . . , m, ri ∈ {0, 1}} . (2.7) • As the Mode Assignment List (MAL) R ≡ (R1, . . . , Rn), i.e. as the n-dimensional vector listing the modes occupied by the particles. Given that for indistinguish- able particles it is not meaningful to assign a specific mode to a specific particle, the order of the elements of a MAL are conventionally taken to be in increas- ing order, so to have a one-to-one correspondence between physical states and MALs. We will refer to this representation as the MAL representation of a many-particle quantum state and, following the notation of [41], denote with Gn,m and Qn,m the set of all MALs of n photons into m modes and the set of collision-free MALs of n photons into m modes, respectively. Equivalently, Gn,m 25
  • 26.
    and Qn,m canbe defined as particular sets of sequences of n integers, that is, Gn,m ≡ {(R1, . . . , Rn) | 1 ≤ R1 ≤ · · · ≤ Rn ≤ m} , Qn,m ≡ {(R1, . . . , Rn) | 1 ≤ R1 < · · · < Rn ≤ m} . (2.8) Clearly, there is a one-to-one relation between Fn,m and Gn,m, and between FCF n,m and Qn,m, as these are just different ways to denote the same things: Fn,m Gn,m, FCF n,m Qn,m. (2.9) Another class of sequences that will be useful in the following is Γn,m, that is, the set of all mn sequences ω = (ω1, . . . , ωn) of integers such that 1 ≤ ωi ≤ m for each i = 1, . . . , n. We note that the sets Gn,m and Qn,m defined above can be thought of as the subsets of non-decreasing and strictly increasing sequences of Γn,m, respectively. A unitary evolution ˆU acts on creation operators in the following way: ˆa† j → ˆUˆb† j ˆU† = m k=1 Ujk ˆb† k, (2.10) where Ujk are the scattering amplitudes from the mode j to the mode k, and ˆb† j are the creation operators of the output states. See examples 2 and 3 for two simple applications of eq. (2.10). Example 2. As an example of the application of eq. (2.10) consider how the unitary ˆU acts on a single-boson state |j = ˆa† j |0 , where |0 is the vacuum state: |j = ˆa† j |0 ˆU −−−−→ ˆUˆb† j ˆU† |0 = m k=1 Ujk ˆb† k |0 = m k=1 Ujk |k out , (2.11) which shows that eq. (2.10) is compatible with the usual rules of evolution of single- particle states. Example 3. A less trivial example is provided considering the evolution of a two- boson state, with the two bosons initially in the modes i and j. The relation between the creation operators and the two-bosons states is if i = j, |i j = ˆa† i ˆa† j |0 , if i = j, |i j ≡ |i i = (ˆa† i )2 √ 2 |0 . 26
  • 27.
    These equations canbe written more coincisely as |i j = ˆa† i ˆa† j √ µ(i,j) |0 , where µ(i, j) is equal to 1 or 2 if i = j and i = j, respectively. Using this notation, we have |i j = ˆa† i ˆa† j µ(i, j) |0 ˆU −−−−→ ˆUˆb† i ˆU† ˆUˆb† j ˆU† µ(i, j) |0 = 1 µ(i, j) m k=1 m l=1 UikUjl ˆb† k ˆb† l |0 = m k=1 m l=1 UikUjl µ(i, j)µ(k, l) |k l out . (2.12) We consider the four possibilities: 1. If i = j and k0 = l0, we have out k0 l0|i j = m k=1 m l=1 UikUjl(δk,k0 δl,l0 + δk,l0 δl,k0 ) = Ui,k0 Uj,l0 + Ui,l0 Uj,k0 , 2. If i = j and k0 = l0, out k0 l0|i j = 1 √ 2 m k=1 m l=1 UikUjl(2δk,k0 δl,k0 ) = √ 2 Ui,k0 Uj,k0 , 3. If i = j and k0 = l0, out k0 l0|i j = 1 √ 2 m k=1 m l=1 UikUil(δk,k0 δl,l0 + δk,l0 δl,k0 ) = √ 2 Ui,k0 Ui,l0 , 4. If i = j and k0 = l0, out k0 l0|i j = 1 2 m k=1 m l=1 UikUil(δk,k0 δl,k0 ) = Ui,k0 Ui,k0 . All of these four cases are coincisely described by writing, for general values of i, j, k0, l0, out k0 l0|i j = 1 µ(i, j)µ(k, l) perm Ui,k0 Ui,l0 Uj,k0 Uj,l0 , where perm(M) is the permanent of the matrix M (see definition 2). 27
  • 28.
    Using eq. (2.5)into eq. (2.10) gives |r ˆU −−−−→ 1 √ r! n k=1   m j=1 URk,j ˆb† j   |0 = 1 √ r! m j1=1 m j2=1 · · · m jn=1 n k=1 URk,jk ˆb† jk |0 = 1 √ r! ω∈Γn,m n k=1 URk,ω(k) ˆb† ω(k) |0 , (2.13) where Γn,m is the set of all sequences of n positive integers lesser than or equal to m. To compute the scattering amplitudes A(r → s, U) of going from the input r to the output s ≡ (s1, . . . , sn), we now have to rearrange the terms on the right hand side of eq. (2.13). To this end we start from the general combinatorial equation (see [41]): ω∈Γn,m f(ω1, . . . , ωn) = ω∈Gn,m 1 µ(ω) σ∈Sn f(ωσ(1), . . . , ωσ(n)), (2.14) where f(ω) ≡ f(ω1, . . . , ωn) is any function of n integer numbers, Gn,m is the sequence of all non-decreasing sequences of n positive integers lesser than or equal to m, given in definition 1, and Sn is the symmetric group, that is, the set of permutations of n elements. Applying eq. (2.14) to eq. (2.13), with f(ω1, . . . , ωn) = n k=1 URkωk ˆb† ωk , we obtain |r ˆU −−−−→ 1 µ(R) ω∈Gn,m 1 µ(ω) σ∈Sn n k=1 URk,ωσ(k) ˆb† ωσ(k) |0 = ω∈Gn,m 1 µ(R)µ(ω)   σ∈Sn n k=1 URk,ωσ(k)   |ω1, . . . , ωn out , (2.15) where in the last step we exploited the commutativity of the product of the creation operators ˆb† k. We thus obtained the following expression for the scattering amplitudes for bosonic particles: A(r → s, U) ≡ out s|r ≡ s| ˆU |r = 1 µ(R)µ(S)   σ∈Sn n k=1 URk,Sσ(k)   , (2.16) in which the factor on the right hand side can be recognised as the permanent of an appropriate matrix built from U. 28
  • 29.
    Definition 2 (Permanent).The permanent of a square matrix, similarly to the de- terminant, is a function which associates a number to a matrix. It is defined very similarly to the determinant, with the exception that all minus signs that are present for the latter become plus signs in the former. More precisely, the permanent of a squared n×n matrix A = (aij) is perm(A) = σ∈Sn a1,σ(1) · · · an,σ(n) = σ∈Sn n k=1 ak,σ(k), (2.17) where Sn is the symmetric group, that is, the set of all permutations of n distinct objects. To express eq. (2.16) through the above defined permanent function, it is also useful to introduce the following notations to refer to particular submatrices built from a given matrix: Definition 3. Let Mk,l denote the set of all k×l complex-valued matrices. If k = l we shall write Mk instead of Mk,k. Now, let A ∈ Mk,l, and let α ∈ Gp,k and β ∈ Gq,l. Then, we shall denote with A[α|β] the p × q dimensional matrix with elements A[α|β]i,j ≡ Aαi,βj . If, moreover, α ∈ Qp,k and β ∈ Qq,l, then A[α|β] is a submatrix of A. If α = β we will simplify the notation to write A[α] instead of A[α|β]. Again, if α ∈ Qp,k and β ∈ Qq,l, we shall denote with A(α|β) the (k − p)×(l − q) dimensional submatrix of A complementary to A[α|β], that is, the submatrix obtained from A by deleting rows α and columns β. Example 4. Consider the 3×4 dimensional matrix A =    1 2 3 0 4 5 6 i 0 4 2 1    . Then, if α = (1, 1) ∈ G2,4 and β = (2, 4, 4) ∈ G3,3, we have A[α|β] ≡ 2 0 0 2 0 0 . If instead α = (2, 3) ∈ Q2,4 and β = (1, 2) ∈ Q2,3, we have A[α|β] ≡ 4 5 0 4 and A(α|β) ≡ 3 0 . Using definitions 2 and 3, we see that for any matrix A ∈ Mm and sequences α, β ∈ Gn,m, we have perm(A[α|β]) = σ∈Sn n k=1 Aαk,βσ(k) , 29
  • 30.
    which is justthe factor in brackets on the right hand side of eq. (2.16). We conclude that A(r → s, U) ≡ out s|r ≡ s| ˆU |r = 1 µ(R)µ(S) perm(U[R|S]). (2.18) It is worth noting that the symmetric nature of the bosonic creation operators - that is, the fact that [ˆa† i , ˆa† j] = δij - was essential in the above derivation. Analogous reasonings carried out using fermionic particles - whose creation operators satisfy {ˆc† i , ˆc† j} ≡ c† i c† j + c† jc† i = δij - would lead to the result Afermions (r → s, U) = det(U[R|S]). (2.19) While eqs. (2.18) and (2.19) may seem very similar at first glance, especially given the similarities in the definitions of permanents and determinants, they are extremely different when trying to actually compute these scattering amplitudes. Indeed, while the determinant of an n dimensional matrix can be efficiently computed, the same is in general not true for permanents. This is exactly what makes the BosonSampling problem interesting, and will be described in detail in the next sections. In conclusion, we can now write down the unitary matrix describing the evolution of the many-body states, that is, the matrix Uα,β(m, n, U), with α, β ∈ Gn,m, whose elements contain the permanents (or the determinants, in the case of fermions) of the corresponding matrix given by eq. (2.18): Uα,β(m, n, U) ≡ 1 µ(α)µ(β) perm(U[α|β]), α, β ∈ Gn,m. (2.20) The dependence of U on n, m, and U will often be omitted when clear from the context. Example 5. Consider the 3×3 unitary matrix Ui,j ≡ ui,j, injected with 2-photon input states. The resulting many-boson scattering matrix is: U(3, 2, U) =            u2,3u3,2 + u2,2u3,3 u2,3u3,1 + u2,1u3,3 u2,2u3,1 + u2,1u3,2 √ 2u2,3u3,3 √ 2u2,2u3,2 √ 2u2,1u3,1 u1,3u3,2 + u1,2u3,3 u1,3u3,1 + u1,1u3,3 u1,2u3,1 + u1,1u3,2 √ 2u1,3u3,3 √ 2u1,2u3,2 √ 2u1,1u3,1 u1,3u2,2 + u1,2u2,3 u1,3u2,1 + u1,1u2,3 u1,2u2,1 + u1,1u2,2 √ 2u1,3u2,3 √ 2u1,2u2,2 √ 2u1,1u2,1√ 2u3,2u3,3 √ 2u3,1u3,3 √ 2u3,1u3,2 u2 3,3 u2 3,2 u2 3,1√ 2u2,2u2,3 √ 2u2,1u2,3 √ 2u2,1u2,2 u2 2,3 u2 2,2 u2 2,1√ 2u1,2u1,3 √ 2u1,1u1,3 √ 2u1,1u1,2 u2 1,3 u2 1,2 u2 1,1            2.2 Counting many-body states Another interesting property of many-body states is their number. Unlike classical states, it is not meaningful to assign a state to the single particles of a quantum many- body state. A quantum state is instead described only giving the list of modes occupied 30
  • 31.
    by the particles,without reference to which particle is in which mode. The above example, for indistinguishable particles, would therefore read like the following: “the two particles are in the modes 1 and 3”. This has the remarkable consequence of changing the total number of possible quantum states of n particles into m modes, with respect to the classical case. The number of classical states of n particles into m modes is easily computed: each one of the n particles can be in one of m modes, independently of the state of the others,so we have m × m × · · · × m n = mn (2.21) possible states. On the other hand, if we were to describe quantum many-body states with this notation (which amounts to using the MAL representation defined in definition 1), we would have to take into account that MALs differing only for a permutation of the elements represent the same quantum state. Moreover, in the case of many-fermion states, Pauli exclusion principle mandates that no more than one particle can occupy the same mode, posing an additional constraint to the possible many-particle configurations. A simple example of the differences between the above described types of many-body states is given in table 2.1, for the case m = 4, n = 2. classical states (1,1) (1,2) (1,3) (1,4) (2,1) (2,2) (2,3) (2,4) (3,1) (3,2) (3,3) (3,4) (4,1) (4,2) (4,3) (4,4) many-boson states (1,1) (1,2) (1,3) (1,4) (2,1) (2,2) (2,3) (2,4) (3,1) (3,2) (3,3) (3,4) (4,1) (4,2) (4,3) (4,4) many-fermion states (1,1) (1,2) (1,3) (1,4) (2,1) (2,2) (2,3) (2,4) (3,1) (3,2) (3,3) (3,4) (4,1) (4,2) (4,3) (4,4) Table 2.1: Comparison of classical, many-boson, and many-fermion states. Many-particle states for n = 2 particles into m = 4 modes, in MAL notation. The striked out states represent non-physical states. In the many-boson case, these are due to the indistinguisha- bility of MALs differing only for the order of the elements. In the many-fermion case, the additional restrictions im- posed by Pauli’s principle must be taken into account. To count the number of many-boson states of n particles into m modes, we will make use of the MOL representation. The problem is thus to find the number of different 31
  • 32.
    sequences of mnon-negative integers, with the sum of the integer equal to n. A visual schematic of this problem is that given in fig. 2.1: each of the m numbers is represented as the number of marbles (or whatever object) in the corresponding bin. From this representation it becomes clear that the question can be equivalently stated as: in how many ways can n objects be distributed among m bins? An easy way to compute this number is starting from fig. 2.1: if we permute in every possible way all of the marbles and the inner separators between the bins, we span the whole set of many-boson states. The number of such permutations is equal to the number of permutations of m − 1 (the number of inner separators) plus n (the number of marbles) objects, that is, (m−1+n)!. However, this overcounts the number of many-boson states, having treated as different configurations with the same occupation numbers but some marbles, or separators, interchanged. The final number must therefore be normalized over the number of such “transparent” permutations. We conclude that the number of many-boson states is |Fn,m| = |Gn,m| = (m + n − 1)! (m − 1)!n! = m + n − 1 n . (2.22) Similarly, the number of many-fermion states, equal to the number of collision-free many-boson states, is |FCF n,m| = |Qn,m| = m! n!(m − n)! = m n . (2.23) The difference between these numbers increases exponentially with n, as seen in fig. 2.2. Many-boson states Classical statesa) b) Figure 2.1: Examples of many-particle states of 4 particles into 6 modes. The disk represent the particles, and each bin a different mode. (a) In the case of many-boson states, the disks (that is, the particles) are all identical to each other, and therefore the number of disks in a given bin completely specifies the state. (b) For classical states, the distinguishability of the particles implies that there can be more than one state with the same number of particles for each mode. 32
  • 33.
    Figure 2.2: Numberof classical (blue), many-boson (orange) and many-fermion (green) states, for m = 8 and m = 32, plotted against the number of photons, in logarithmic scale. 2.3 Tools for quantum optics experiments The basic building blocks used to manipulate light in linear optics are beam splitters, half- and quarter-wave plates, and phase shifters. For quantum optics experiments, single photon sources and detectors are also required. In this section we will give a brief description of these tools, and establish the mathematical conventions used in the following sections. 2.3.1 Single-mode phase shift This optical component changes the phase of the electromagnetic field in a given mode, which means that it acts on the creation operator of a given mode k in the following way: ˆa† in phaseshifter −→ ˆa† out = eiφ ˆNin ˆa† ine−iφ ˆNin = eiφ ˆa† in. (2.24) Physically, a phase shifter can be implemented as a slab of transparent material with an index of refraction that is different from that of free space. 2.3.2 Beam splitter A beam splitter (BS), which is a central component of many optical experiments, consists of a semireflective mirror: when light impinges onto the mirror, part of it will be reflected and part will be transmitted. In a classical description, a BS can be simply characterized by the relations E3 = R31E1 + T31E2, E4 = T41E1 + R42E2, (2.25) where E1, E2 are the amplitudes of the incoming electric fields, E3, E4 the amplitudes of the outcoming ones, and the complex coefficients Rij, Tij are respectively the re- 33
  • 34.
    flectances and transmittancesalong a particular path through the BS (while we will only consider the electric fields here, completely analogour relations can be written for the corresponding magnetic fields). In matrix notation eq. (2.25) are written as E3 E4 = R31 T32 T41 R42 E1 E2 , (2.26) where the above 2x2 matrix is known as the beam splitter matrix. In the ideal case of a lossless BS, considerations of energy conservation lead to the following relations between the coefficients of the BS: |R31|2 + |T41|2 = |R42|2 + |T32|2 = 1 and R31T ∗ 32 + T41R∗ 42 = 0, (2.27) from which it follows that |R31| = |R42| ≡ |R| and |T41| = |T32| ≡ |T |. Using these relations we can rewrite eq. (2.26) in the form: E3 E4 = cos θ ie−iφ sin θ ieiφ sin θ cos θ E1 E2 , (2.28) where θ and φ parametrize the set of all possible 2x2 unitary matrices. The general structure of the BS can be further simplified by additional assumptions on the forms of the reflection and transmission coefficients. In the case of symmetrical BSs, for example, we can assume the coefficients to satisfy R31 = R42 ≡ R = |R| exp(iφR) and T32 = T41 ≡ T = |T | exp(iφT ), (2.29) with |R|2 + |T |2 = 1 and φR − φT = ±π/2, which translates in the condition φ = 0 in the notation of eq. (2.28). In the case of a 50:50 beam splitter with reflection and transmission coefficients of equal magnitude we have |R| = |T | = 1/ √ 2 and φR − φT = π/2, which translates in the conditions φ = 0 and θ = π/4 in the notation of eq. (2.28). The relations in eq. (2.25) convert into analogous relations between the quantized field operators: ˆa† 3 = cos θˆa† 1 + ie−iφ sin θˆa† 2 and ˆa† 4 = ieiφ sin θˆa† 1 + cos θˆa† 2, (2.30) or, in the case of 50:50 BSs, in ˆa† 3 = 1 √ 2 ˆa† 1 + iˆa† 2 and ˆa† 4 = 1 √ 2 iˆa† 1 + ˆa† 2 . (2.31) Inverting eq. (2.31) leads to ˆa† 1 = 1 √ 2 ˆa† 3 − iˆa† 4 and ˆa† 2 = 1 √ 2 −iˆa† 3 + ˆa† 4 . (2.32) 34
  • 35.
    An equivalent wayto denote eq. (2.32) is obtained by denoting with ˆa† 1, ˆa† 2 the two input modes and with ˆb† 1,ˆb† 2 the two output modes, and describing the evolution through the beam splitter with the same notation used in eq. (2.10): ˆa† 1 → 1 √ 2 ˆb† 1 − iˆb† 2 and ˆa† 2 → 1 √ 2 −iˆb† 1 + ˆb† 2 . (2.33) An important type of BS is the polarizing beam splitter (PBS). This is a device which distinguishes different polarization states of the incoming light. If the PBS is cut to separate horizontal and vertical linear polarizations, the transformation of the incoming modes, ˆa1 and ˆa2, yields the following outcoming modes: ˆa1,H → ˆa1,H and ˆa1,V → ˆa2,V , ˆa2,H → ˆa2,H and ˆa2,V → ˆa1,V , (2.34) that is, it does not change the spatial mode of the horizontally polarized states, but switches the spatial mode of vertically polarized ones. E1 E2 E3 E4 Figure 2.3: Schematic repre- sentation of a lossless beam splitter, with the notation used in the text for incom- ing and outcoming electric fields. 2.3.3 Wave plates A waveplate is an optical component which alters the polarization of a light wave travelling through it. The two most commonly used types of waveplates are half- and quarter-wave plates. The former rotates the polarization direction of linearly polarized light, while the latter converts linearly polarized light into circularly polarized light and vice versa. 35
  • 36.
    The same mathematicaldescription given above for BSs applies to the evolution of light through a waveplate. Instead of having two different spatial modes, ˆa† 1 and ˆa† 2, the two incoming modes have two different polarizations, that we will denote with ˆax and ˆay. The equations describing the evolution of the creation operators of the polarization modes of the field are thus ˆa† x = cos θˆa† x + ie−iφ sin θˆa† y and ˆa† y = ieiφ sin θˆa† x + cos θˆa† y, (2.35) where x and y are the polarization modes of the light after the evolution through the waveplate. In the notation of eq. (2.35), half-wave plates correspond to φ = π/2, and quarter-wave plates to φ = 0. Figure 2.4: Left: Example of an optical beam splitter as commonly seen in quantum optics laboratories. Right: Example of a phase shifter, as commonly seen in quantum optics laboratories. 36
  • 37.
    2.3.4 Single-photon sources Anideal single-photon source [42, 43, 44] would be one that 1. Is deterministic (or “on demand”), meaning that it can emit a single photon at any arbitrary time defined by the user, 2. Has a 100% probability of emitting a single photon and a 0% probability of multiple- photon emission, 3. Subsequently emitted photons are indistinguishable, 4. The repetition rate is arbitrarily fast. Given however that no real-world photon source satisfies all of these specifications, the deviations from the ideal characteristics must be considered when designing exper- iments. Single-photon sources are broadly classified into deterministic and probabilistic. Among the implementations of the former are those based on color centers [45, 46, 47], quantum dots [48, 49, 50], single atoms [51], single ions [52], single molecules [53], and atomic ensembles [54], all of which can to some degree emit single photons “on demand”. On the other hand are the probabilistic single-photon sources. These generally rely on photons created in pairs via parametric downconversion in bulk crystals and waveguides, and four-wave mixing in optical fibers. While these sources are probabilistic - and therefore it is not possible to know exactly when a photon has been emitted - because the photons are created in pairs, one of the emitted photons can be used to herald the creation of the other. While the distinction between deterministic and probabilistic sources is clear in the abstract, this distinction blurs in practice. This due to the unavoidable experimental errors that make also “theoretically deterministic sources” be probabilistic in practice. Although many applications, especially those in the field of quantum-information science, require an on-demand source of single photons, probabilistic single-photon sources remain a fundamental tool, and are widely used in many quantum optics experiments. Spontaneous parametric downconversion (SPDC) is an important process in quan- tum optics, typically exploited to generate entangled photon pairs, or heralded single photons. This is achieved using a nonlinear crystal - that is, a medium in which the dielectric polarization responds nonlinearly to the electric field - which converts the photons of a pump beam into pairs of photons of lower energy. A simple model of the interaction Hamiltonian in such a crystal is HI ∼ χ(2) ˆapˆa† sˆa† i + hermitian conjugate, (2.36) 37
  • 38.
    where χ(2) is thesecond-order nonlinear susceptibility of the nonlinear medium. Here ˆap is the annihilation operator of the pump beam, and ˆa† s and ˆa† i are the creation operators of the signal and idler beams respectively (the names signal and idler are there for historical reasons, and carry no special significance throughout this thesis). In the simplest case, with the signal and idler beams initially in vacuum states, a single photon from the pump beam is converted into two optical photons, one in the signal beam and the other in the idler: |1 p |0 s |0 i SPDC −−−→ ˆapˆa† sˆa† i |1 p |0 s |0 i = |0 p |1 s |1 i . (2.37) The conditions of energy and momentum conservation pose a contraint on the signal and idler generated photons, which will depend on the pump beam through the following relations (see also fig. 2.5): ωp = ωs + ωi, kp = ks + ki. (2.38) kp ks ki Figure 2.5: Phase-matching condition There are two main types of SPDC processes, named type I and type II: In type I, the signal and idler photons have the same polarization, orthogonal to that of the pump. In type II downconversion, the polarizations of signal and idler photons are instead orthogonal to each other. Because of birefringence effects, the generated photons are emitted along two cones, one for the ordinary wave and another for the extraordinary wave. At the intersection of the cones, the two photons detected are in a polarization- entangled state of the form 1 √ 2 (|V s |H i + |H s |V i), (2.39) where |H and |V denote an horizontally and vertically polarized state respectively. A schematic representation of this process is shown in fig. 2.6. More generally, the state produced by an SPDC source does not always contain two photons, but has the form ∞ p=0 gp |p1 |p2 , (2.40) 38
  • 39.
    with |pi =(ˆa† i )p / √ p!, and 0 ≤ g < 1 an appropriate parameter determining the ratio of generated photons and dependent, among other things, on the strength of the pump beam. Typically, g 1, so that the probability of generating many photons is low. For instance, in typical conditions g ∼ 0.1 and the probability of generating a state of the form |21 |22 is lower of a factor ∼ 102 than the probability of producing a single pair |11 |12 . The main advantages of SPDC sources are the high photon indistinguishability, the collection efficiency, and relatively simple experimental setups. This technique, however, suffers from two drawbacks. First, since the nonlinear process is nondeterministic, so is the photon generation, even though it can be heralded. Second, the laser pump power, and hence the source’s brilliance, has to be kept low to prevent undesired higher-order terms in the photon generation process. Figure 2.6: Representation of type II downconversion. The pump beam (red) impinges on the nonlinear crystal, generating, due to birefringence effects, photons along two cones. On the upper ordinary cone (orange), the generated photons are horizontally polarized, while on the lower extraordinary cone (green) the generated photons are vertically aligned. Postselecting on the two intersections of these cones (blue dots), a pair of polarization-entangled photons is obtained. 2.3.5 Single-photon detectors Roughly speaking, single-photon detectors are devices which convert single photons into an electrical signal of some sort [42]. Quantum information science is one of the field currently driving much of the research toward improved single-photon-detector technology. For example, many quantum communication protocols rely heavily on 39
  • 40.
    detector properties suchas detection efficiency. An ideal single-photon detector [42] would require the following characteristics: 1. The detection efficiency - that is, the probability that a photon incident upon the detector is successfully detected - is 100%, 2. The dark-count rate - that is, the rate of detector output pulses in the absence of any incident photons - is zero, 3. The dead time - that is, the time after a photon-detection event during which the detector is incapable of detecting another photon - is zero, 4. The timing jitter - that is, the variation from event to event in the delay between the input of the optical signal and the output of the electrical signal - is zero. Additionally, an ideal single-photon detector would also be able to count the number of photons in an incident pulse. Detectors able to do this are referred to as photon-counting, or photon-number resolving, detectors. However, non-photon-number-resolving detec- tors, which can only distinguish between zero photons and more than zero photons, are the most commonly used. Indeed, while detecting a single photon is a difficult task, discriminating the number of incident photons is even more difficult. Examples of non- photon-number-resolving single-photon detector technologies include single-photon avalanche photodiodes [55], quantum dots [56], superconducting nanowires [57], and up-conversion detectors [58, 59, 60]. 2.4 Hong-Ou-Mandel effect The differences between bosons and fermions are not only in the different numbers of microstates. Their statistical behaviour can differ significantly, as well as be significantly different from the behaviour of distinguishable particles. Bosons, roughly speaking, tend to occupy the same state more often than classical particles, or fermions, do. This behaviour, referred to as bosonic bunching, has been verified in numerous experimental circumstances, including fundamental ones like Bose-Einstein condensation [61, 62, 63]. In the context of optical experiments, the most known effects arising from the symmetric nature of the Bose-Einstein statistics is the Hong-Ou-Mandel (HOM) effect [64]. In the original experiment, two photons are sent simultaneously through the two input ports of a symmetric beam splitter. Since no interaction between the two photons takes place, one would expect no correlation between the detection events at the two output ports. Instead, the photons are always seen either both on the first output mode, or both on the second output mode. 40
  • 41.
    + − + Figure2.7: Pictorial representation of the suppression of non-bunched events, when two indistinguish- able bosons evolve through a symmetric beam splitter. Each of the four images represent a possible evolution of the bosons, with all of them interfering with each other. The two events which would result in one boson per output port turn out to interfere destructively (note the minus sign), and are suppressed. This effect is a direct consequence of the quantum interference between the possible ways two-photon states can evolve. A pictorial representation of this is given in fig. 2.7: when the photons are injected into two different ports of a symmetric beam splitter, the scattering amplitudes corresponding to the output photons being in two different modes interfere destructively. We can derive this result applying eq. (2.33) to the two-photon input state |11, 12 = ˆa† 1ˆa† 2 |0 : |11, 12 = ˆa† 1ˆa† 2 |0 → 1 √ 2 ˆb† 1 − iˆb† 2 1 √ 2 −iˆb† 1 + ˆb† 2 |0 = 1 2 −i(ˆb† 1)2 + ˆb† 1 ˆb† 2 − ˆb† 2 ˆb† 1 − i(ˆb† 2)2 = −i 2 (ˆb† 1)2 + (ˆb† 2)2 |0 = −i √ 2 (|21 + |22 ) , (2.41) where in the last steps we used the rules given in eq. (2.3), and in particular the commutativity of the bosonic creation operators, which implies that ˆb† 1 ˆb† 2 = ˆb† 2 ˆb† 1 (see also fig. 2.7 for a pictorial representation of how the suppression of non-bunched events arises). We thus conclude that when two indistinguishable photons enter a symmetric beam splitter one in each mode, they always come out in the same mode. This property of photons (or, more generally, of bosons) is highly non-classical, and is a notable example of how interesting effects can arise when dealing with many-body quantum states. In a real world experiment, the two input photons will never be perfectly indistin- guishable, though. A more careful analysis, taking into account the potentially different times at which the photons reach the beam splitter, as well as the coherence time of each photon wave packet, predicts a smooth transition from the classical behaviour to the antibunching effect described above [65]. In the general case of partially distinguishable particles, the probability PT (s, x) of detecting an output state s in an HOM experiment with a time delay quantified by x, becomes an average of the probability PB for bosons and the probability PD assigned to distinguishable particles, weighted by two factors |c1|2 and |c2|2 which depend on 41
  • 42.
    the relative timedelay: PT (s, x) = |c1(x)|2 PB(s) + |c2(x)|2 PD(s). (2.42) In a typical experimental scenario, with the incoming photons having a Gaussian frequency distribution around the central frequency, PT (s, x) also has a gaussian profile [65], as shown in a typical case in fig. 2.8. In the case of fully distinguishable particles, where PT = PD, no interaction occurs and the output events of the single photons are not correlated. There are two possible classical configurations of two photons in the output MOL (1, 1), and one configuration for both (2, 0) and (0, 2). It follows that PD((1, 1)) = 2/4 and PD((2, 0)) = PD((0, 2)) = 1/4, as shown in fig. 2.8 for x ≈ ±400µm. On the other hand, when the particles are fully indistinguishable, PT = PB. The probability of the various outcomes is now given by eq. (2.41). The output (1, 1) is thus suppressed, while PB((2, 0)) = PB((0, 2)) = 1/2, as shown in fig. 2.8 for x = 0. Figure 2.8: Transition to indistinguishability in a HOM experiment. Changing the time delay between the two input photons a dip in the number of measured output coincidences is seen, corresponding to the time delay (or, equivalently, the path delay) making the photons indistinguishable. The blue line is the probability of detecting two photons in one of the two output ports, that is, PT ((2, 0), x) or equivalently PT ((0, 2), x). The red line is the probability PT ((1, 1), x) of detecting the two photons in the two different output ports. As expected, the red plot shows a peak corresponding to the antibunching effect arising when the particles are indistinguishable, while the blue plot show the HOM explained above. 42
  • 43.
    Chapter 3 BosonSampling In thischapter we discuss various aspects of the BosonSampling computational prob- lem. In section 3.1 the problem of experimentally assessing quantum supremacy is discussed, in order to appreciate the importance of BosonSampling in the modern research context. Section 3.2 follows with the description of what the BosonSampling computational problem is, and its advantages in obtaining experimental evidences of quantum supremacy. In section 3.3 some issues related to the scalability of boson sampling implementations are described. The chapter closes with a description of scattershot boson sampling in section 3.4, as an alternative architecture to scale boson sampling implementations to higher numbers of photons. 3.1 Importance of BosonSampling It is currently believed that many quantum mechanical systems cannot be efficiently simulated with a classical computer [1]. This implies that a quantum device is, to the best of our knowledge, able to solve problems de facto beyond the capabilities of classical computers. Exploiting this quantum advantage requires however an high degree of control over the quantum system, not yet manageable with state of the art technology. In particular, a post-classical computation with a universal quantum computer will require an high degree of control of a large number of qubits, and this implies that an experimental evidence of quantum supremacy [3] with an universal quantum computer will likely require many years. A notable example is given by the large gap between the number of qubits that can currently be coherently controlled (∼10), and the number of qubits required for a calculation such as prime factorization, on a scale that would challenge classical computers (∼106 ). Consequently, there is considerable interest in non-universal quantum computers and quantum simulators that, while able to only solve specific problems, might be significantly easier to be implemented experimentally. Such devices could give the first experimental demonstration of the power of quantum 43
  • 44.
    devices over classicalcomputers, and potentially lead to technologically significant applications. Moreover, in the context of searching for experimental evidence of quantum supremacy, the technological difficulties are not the only issue. To show this, we will consider as an example Shor’s quantum algorithm [30] to efficiently factorize integer numbers. Even if we were to get past the technological difficulties of implementing Shor’s algorithm with sufficiently many qubits, it could be easily argued that such an achievement would not be a conclusive evidence that quantum mechanics allows post-classical computations. This because we do not have to date a mathematically sound proof that there cannot be a classical algorithm to efficiently factorize integers. In the language of complexity theory, this corresponds to the fact that we do not have a proof that Factoring is not in P, even though this is believed enough to base modern cryptography is based on this conjecture. More generally, before 2010, there were no instances of problems efficiently solved by quantum computers, which were proved to not be efficiently solvable with classical ones. This changed when, in 2010, Aaronson and Arkhipov (AA) proposed [8] the Boson- Sampling problem as a way to obtain an easier experimental evidence of quantum supremacy. BosonSampling is a computational problem that, while hard to solve for a classical computer, is efficiently solved by a special-purpose quantum device. AA showed that BosonSampling is naturally implemented using only linear optical ele- ments, in a photonic platform named a boson sampler. The experimental realization of a boson sampler, while still challenging with present-day technologies, requires much less experimental efforts with respect to those required to build a universal quantum computer. In fact, the AA scheme requires only linear optical elements and photon-counting detectors, as opposite to, for example, the Knill, Laflamme & Milburn approach [7, 40] for universal linear optics quantum computing, which requires among other things an extremely fast feedback control of the detectors. 44
  • 45.
    Figure 3.1: Galtonboard: n identical balls are dropped one by one from the upper corner, and are randomly scattered into the lower slots. The quantum generalization of this classical “rudimentary computer” leads to the idea of BosonSampling. Credits: Nicolò Spagnolo. 3.2 The BosonSampling computational problem Consider the following linear optical experiment: the n-photon state |rAA given by |rAA ≡ |11, . . . , 1n, 0n+1, . . . , 0m ≡ ˆa† 1 . . . ˆa† n |01, . . . , 0m , (3.1) is injected into a passive linear optics network, which implements a unitary map on the creation operators: ˆa† k → ˆUˆb† k ˆU† = m j=1 Uk,j ˆb† j. (3.2) with U an Haar-random m×m complex unitary matrix. The evolution induced on |rAA is |rAA → s∈Fn,m A(r → s, U) |s , (3.3) where the sum is extended over all many-boson states of n particles into m modes, and the scattering amplitudes A are, as shown in eq. (2.18), proportional to the permanents of n×n matrices. AA argued that, for m n, the output of such an apparatus cannot be efficiently predicted by a classical computer, neither exactly nor approximately [8]. This was rigorously proved in the exact case. The problem of approximately sampling 45
  • 46.
    from the outputprobability distribution of a boson sampling apparatus depends instead on a series of conjectures, for which strong supporting evidence was provided [8]. This problem, which amounts to that of being able to sample from the output proba- bility distribution given in eq. (2.18), is referred to as the BosonSampling computational problem. The constraint m n is essential for the hardness result, as otherwise semi-classical methods become efficient [66, 67]. Roughly speaking, a boson sampling apparatus is a “quantum version” of a Galton board. A Galton board, named after the English scientist Sir Francis Galton, is an upright board with evenly spaced pegs into its upper half, and a number of evenly- spaced rectangular slots in the lower half (see fig. 3.1). This setup can be imagined to be a rudimentary “computer”, where n identical balls are dropped one by one from the upper corner, and are randomly scattered into the lower slots. In the quantum mechanical version, the n balls are indistinguishable bosons “dropped” simultaneously, and each peg a unitary transformation, typically implemented as a set of beam splitters and phase shifters. More precisely, BosonSampling consists in producing a fair sample of the output probability distribution P(s |U, rAA) ≡ |A(r → s, U)|2 , where s is an output state of the n bosons, and rAA the above mentioned input state. The unitary ˆU and the input state rAA are the input of the BosonSampling problem, while a number of output states sampled from the correct probability distribution are its solution (see fig. 3.2). 46
  • 47.
    Figure 3.2: Conceptualboson sampling apparatus. (a) The input of the BosonSampling problem is the input many-photon state (in figure the state |0, 0, 1, 1, 0, 1, 0, 0, 0 ), and a suitably chosen unitary U. The output is a number of outcomes picked according to the bosonic output probability distribution (in figure, two examples of such states are provided, with MOLs 101000100 and 110000100). Colleting enough such events allows to reconstruct the probability distribution. This, however, requires an exponentially increasing (in n) number of events. (b) Injecting an m-mode unitary with n indistinguishable photons, the output state is a weighted superposition of all possible outcomes. Measuring in which modes the photons ended up results in the collapsing of this superposition. The probability of finding the photons in a certain configuration is given by eq. (3.4). Credits: [10]. 47
  • 48.
    Example 6 (Solutionof the BosonSampling problem). Let ˆU be a randomly chosen unitary transformation, described by the 4×4 matrix U = −0.60046+0.220549i −0.108966−0.527875i −0.367627+0.212122i 0.0655053 +0.340358i −0.576174−0.386133i −0.463806+0.524027i 0.0458648 −0.0209767i 0.148456 −0.0679879i −0.0337116−0.30791i 0.408837 −0.0733373i −0.0106664+0.578256i 0.543236 −0.319264i −0.0894887−0.0760134i 0.211627 +0.0494216i 0.680731 +0.139364i 0.0591935 +0.672804i . The output probability distribution resulting from the injection of the two-photon input state with MOL r = (1, 1, 0, 0) is output state probability amplitude probability (0,0,0,2) 0.046478 + 0.0651595 i 0.00640597 (0,0,1,1) -0.0300108+0.0707215 i 0.00590217 (0,0,2,0) -0.0175525 + 0.0246647 i 0.000916437 (0,1,0,1) -0.260804-0.194492 i 0.105846 (0,1,1,0) 0.0432791 -0.312955 i 0.099814 (0,2,0,0) 0.462674 + 0.265491 i 0.284553 (1,0,0,1) 0.0195337 -0.147833 i 0.0222362 (1,0,1,0) 0.270811 +0.0424448 i 0.0751401 (1,1,0,0) 0.0218767 -0.0707259 i 0.00548074 (2,0,0,0) 0.609711 + 0.148185 i 0.393706 Measuring the exit modes of the two injected photons at each pulse, we obtain a series of samples from the above probability distribution. An example of 10 such samples is the sequence (0,2,0,0) (0,1,1,0) (2,0,0,0) (2,0,0,0) (0,2,0,0) (0,1,1,0) (2,0,0,0) (0,1,0,1) (0,2,0,0) (0,2,0,0) The above list is exactly what BosonSampling is all about: obtaining a list of “labels” distributed according to a particular probability distribution. In other words, the BosonSampling problem is not solved obtaining the above listed probabilities, but obtaining a number of samples from this distribution. The number of such samples is not really important here, even producing a single state 48
  • 49.
    from the correctprobability distribution would theoretically be enough to achieve a post-classical computation, though possibly making it harder to experimentally verify. The hardness of the BosonSampling problem can be traced back to the #P-hardness of computing the permanent of a generic complex-valued matrix. Indeed, as shown in eq. (2.18), the probability P(r → s, U) of an input r evolving into s, is proportional to the permanent of the matrix U[R|S] (recalling definition 3): P(r → s, U) = |A(r → s, U)|2 = 1 µ(R)µ(S) |perm(U[R|S])|2 . (3.4) Computing the permanent of a n×n matrix with the fastest known classical algorithms [41, 68] requires a number of operations of the order O(n2n ). This means that, for example, computing the permanent of a 30×30 complex matrix, corresponding to a single scattering amplitude for a 30-photon state, requires a number of operations of the order of ∼ 1010 . If the average time required by a classical computer to perform a single operation is of the order ∼ 107 , the computation of one such scattering amplitude will require ∼ 10 minutes. While still clearly manageable by a classical computer, this already shows the potential advantages of a boson sampling apparatus: if the experimental problems related to coherently evolve 30 indistinguishable photons inside an interferometer were to be solved, this would allow to sample from the probability distribution given by eq. (3.4) without actually knowing the probabilities itselves. AA demonstrated that, should boson sampling be classically easy to solve, this would have very strong and undesied consequences in computational complexity theory, and therefore it is most probable that boson sampling is not classically easy to solve. It is worth stressing that the BosonSampling problem is not that of finding the per- manents in eq. (2.18), but only that of sampling from the related probability distribution. In fact, not even a boson sampler is able to efficiently compute these scattering probabil- ities. This is due to the fact that, to reconstruct a probability distribution spanning over a number M of events, roughly speaking, the number of samples is required to be at least of the order of M. But, as shown in eqs. (2.22) and (2.23), M scales exponentially with n, implying that the number of experimental samples required to reconstruct the probability distribution becomes exceedingly large very soon. In figs. 3.3 and 3.4 is shown that, if the number of samples is not large enough, the reconstructed probability distribution is different from the real one. In fact, generally speaking, there are strong arguments against the possibility to compute the permanents of complex-valued matri- ces by means of quantum experiments [69], although attempts have been reported in this direction [70]. 49
  • 50.
    0 0.025 0.05 0.075 0.1 0.125 0.15 States Probabilitydistribution 0.0 0.5 1.0 1.5 2.0 States Numberofsamples 0 5 10 15 States Numberofsamples 0 50 100 150 States Numberofsamples a) b) c) d) Figure3.3: Example of boson sampling from a randomly chosen 8×8 unitary matrix. (a) Exact output probability distribution for two photons injected in the first two modes of a random 8-mode interferometer. (b), (c), (d) Output states sampled from the output probability distribution, for a number of samples equal to 10 (b), 100 (c) and 1000 (d). As seen, with a low number of samples, the histogram may appear different from the true probability distribution. 0 0.01 0.02 0.03 States Probabilitydistribution 0.0 0.2 0.4 0.6 0.8 1.0 States Numberofsamples 0 1 2 3 4 5 States Numberofsamples 0 5 10 15 20 25 30 States Numberofsamples a) b) c) d) Figure 3.4: Example of boson sampling from a randomly chosen 8×8 unitary matrix. (a) Exact output probability distribution for four photons injected in the first four modes of a random 8-mode interferometer. (b), (c), (d) Output states sampled from the output probability distribution, for a number of samples equal to 10 (b), 100 (c) and 1000 (d). The number of samples required to reliably recover the original probability distribution is much higher than in fig. 3.3, due to the higher number of many-boson states, which are here 8+3 4 = 330 against the 8+1 2 = 36 of fig. 3.3. 50
  • 51.
    The complexity ofBosonSampling makes it an extremely valuable candidate to gain experimental evidences of the supremacy of quantum devices over classical computers. Indeed, it presents several advantages in this regard over, for example, Factoring, which is the paradigmatic problem that would allow quantum computers to perform a post-classical computation: 1. The BosonSampling problem is even harder than Factoring, being related to the #P-hard complexity class, and believed to not be in NP. 2. A boson sampler requires significantly less resources to be implemented than a universal quantum computer. In particular, it does not require adaptive or feed- forward mechanisms, nor fault-tolerance methods. This relatively simple design has already prompted a number of small-scale implementations of increasing complexity [9, 10, 11, 12, 13, 14, 15, 17, 18, 25, 71]. AA suggested [8] that a 400-modes interferometer fed with 20 single photons is already at the boundary of the simulation powers of present-day classical computers. While in this regime it would still be possible to carry out a classical simulation, the quantum device should be able to perform the sampling task faster than the classical computer. 3. The theoretical evidence of the hardness of BosonSampling is stronger than that of factoring integers: while in the former case the result only relies on a small number of conjectures regarding the hardness of some problems [8], in the latter case there is no compelling evidence for Factoring to not be in P. While known to be in BQP, Factoring is only believed to be in NP, and strongly believed to not be in NP-hard. While the hardness of Factoring is strong enough to build modern cryptography, it could also happen that a polynomial-time algorithm will be discovered showing that Factoring is in P as, basically, the sole evidence for its hardness is the fact that no efficient classical algorithm is yet known. 3.3 Scaling experimental boson sampling implemen- tations The hardness of BosonSampling has another potentially important consequence: it could provide the first experimental evidence against the ECT [8]. This point, however, is still subject to some debate [67, 72, 73], due to the somewhat informal nature of the ECT itself. Indeed, the ECT is not a mathematical statement, but a statement about how the physical world behaves, in a certain asymptotic limit. Because of this it is 51
  • 52.
    somewhat ill-defined whatexactly would be “enough experimental evidence” to “prove” or “disprove” such a statement. While AA argued for the hardness of both the exact and approximate BosonSam- pling problems, they did not take into account other forms of experimental imperfec- tions. More precisely, AA showed that even attempting to estimate the output probability distribution of a boson-sampler is likely computationally hard, as long as the probability P of each input state being correctly produced by the sources scales as P > 1/ poly(n), that is, does not vanish faster than the inverse of a polynomial in n [8]. Further evidence that even lossy systems or systems with mode mismatch are likely to be classically hard to solve was later given [74]. Another source of errors which could potentially undermine the scalability of boson sampling are the unavoidable circuit imperfections, especially taking into account the fact that a boson sampler cannot likely implement fault-tolerant mechanisms [8]. Therefore, small errors in the calibrations of the elementary optical elements, like beam splitters and phase-shifters, may accumulate and result in an output probability distribution significantly different from the ideal one. A thorough analysis of this issue, reported in [75], leads to the result that as long as the average fidelity of the elementary gates scales at least like 1 − O(1/n2 ), the overall implementation provides a reasonably good output distribution. Generally speaking, all of these works seem to reinforce the idea that the Boson- Sampling problem is scalable even when experimentally plausible conditions are taken into account. This however, is not definitively enstablished. While the potential for scaling up the number of modes of network implementations has been fairly well enstablished, and experiments with 2 photons into up to 21 modes have been reported [76], scaling the number of photons is much harder and will likely require technologi- cally improvements, especially on the side of the employed single-photon sources and detectors. In particular, most reported boson sampling experiments used SPDC single-photon sources. With this type of scheme, using a single SPDC source, the probability of producing n photons decreases exponentially with n, as shown in eq. (2.40). This means that in order to implement a boson sampler with higher numbers of photons another kind of architecture is most likely required. One such example is provided by a scattershot boson sampler, which will be described in the next section. 3.4 Scattershot boson sampling One of the main difficulties in scaling up the complexity of boson sampling devices is the requirement of a reliable source of many indistinguishable photons. Indeed, despite recent advances in photon generation [42] using atoms, molecules, color centers in 52
  • 53.
    diamond, and quantumdots, currently the most widely used method remains parametric downconversion. The drawbacks of SPDC sources, described in section 2.3.4, have however restricted SPDC implementations of boson sampling experiments to proof-of- principle demonstrations. Recently a new scheme, named scattershot boson sampling [19], has been proposed to make the best use of SPDC sources for photonic boson sampling, greatly enchancing the rate of n-photon events [19, 20]. As opposite to a boson sampling experiment, in which a single fixed input state is employed, in a scattershot boson sampler the photons are injected into randomly chosen modes at each shot. This is achieved adopting an heralded single-photon source for every mode of the interferometer (see fig. 3.5). If a single source has a probability p of injecting a photon into the interferometer for each laser pulse, the probability of n particular sources injecting a photon at the same time is pn . However, there are m n possible combinations of sources that can simultaneously inject a photon into the interferometer, so that the resulting probability of injection of any configuration of n photons is m n pn . For m n2 , which is the regime in which BosonSampling is hard, the binomial factors dominate over pn . Thus, while the efficiency of using a single SPDC source to generate many photons decreses exponentially with pn , the overall efficiency of a scattershot boson sampling apparatus with m SPDC sources increases exponentially with n, in the regime n m, as shown in fig. 3.6. While a scattershot boson sampling apparatus would technically solve a problem different from BosonSampling as defined above, due to the randomly changing input state, it has been shown [19, 20] that the computational problem solved by a scattershot boson sampler is at least as hard as BosonSampling, thus making this new scheme as good as the original one to achieve a post-classical computation. An experimental implementation of a scattershot boson sampler has been recently reported [18], using a 13-mode integrated photonic chip and up to six SPDC sources, to obtain data corresponding to two- and three-photon interference. 53
  • 54.
    Figure 3.5: Ascattershot boson sampler. An SPDC source is used for every input of the m-mode interferometer. Each SPDC source generates (ignoring higher order terms) a pair of photons, one of which is injected into the interferometer, while the other heralds the generation event. Every time a detector corresponding to an heralding photon “clicks”, we know that a photon has been injected into the corresponding mode of the interferometer. (image courtesy of [18]) m =8 m =16 m =32 m =64 0 10 20 30 40 50 60 n 100 106 1010 1014 1018 p =0.5 p =0.1 p =0.01 5 10 15 20 25 30 n 1010 1020 1030 1040 1050 11 a) b) Figure 3.6: (a) Logarithmic plot of the binomial factor m n as a function of n, for m equal to 8 (blue), 16 (orange), 32 (green) and 64 (red). (b) Logarithmic plot of the probability pn n2 n of injecting n photons into an m = n2 -modes interferometer with a scattershot boson sampling apparatus, against the number of photons n, with the single pair generation probability p equal to 0.01 (blue), 0.1 (orange) and 0.5 (green). The plot shows how the probability increases exponentially with n. Moreover, the lower is the value of p, the higher is the number of photons required to see this advantage, as seen here in the case of p = 0.01. Credits: [18]. 54
  • 55.
    Chapter 4 Boson samplingvalidation In this chapter we discuss the problem of validating boson sampling experiments. In section 4.1 the problem is presented in its generality. In section 4.2 we focus on the validation protocols based on zero-transmission laws. A validation scheme with Fourier matrices is described, an experimental implementation of which is given in chapter 6. In section 4.3 we discuss the possibility of validating scattershot boson sampling experiments with the above described validation scheme for Fourier matrices. In section 4.4 an alternative validation scheme, based on Sylvester matrices, is presented. The advantages of this protocol to validate scattershot boson sampling experiments is discussed. 4.1 Validating boson sampling implementations The very hardness of BosonSampling, while being its main feature, also poses the problem of certifying the results of an experimental implementation. This means that if someone were to present an alleged boson sampling device as a black box, we may not be able to certify its output. In other words, we may not be able to tell if the output of the black box is the result of true many-body interference, as opposite, say, of that of a classical computer giving in output a series of randomly chosen samples. From a complexity-theoretical point of view, this is due to the fact that BosonSampling belongs to the #P class [8], and thus not classically verifiable in polynomial time. More intuitively, we can trace back the causes of this certification problem to two fundamental factors: Hardness of computing the bosonic probability distribution: It is not possible to efficiently calculate with classical resources, and probably neither with quantum ones [8], the bosonic output probability distribution. While this may naively be thought of as the main issue, it does not imply per se a fundamental certification problem. As a counterexample we may think of Factoring: while it is not (as to the current state of 55
  • 56.
    knowledge) possible toefficiently find the prime factorization of an integer number, it is possible to efficiently verify whether a proposed factorization is the right one, simply by multiplying the numbers together. No such mechanism is, however, currently thought to exist for BosonSampling. Hardness of discriminating two probability distributions: Another issue, ar- guably even more important than the latter, arises from the exponential number of possible outputs which the bosonic probability distribution covers, and the sampling nature of BosonSampling. Indeed, even completely disregarding the hardness of com- puting permanents and scattering amplitudes, the sheer number of samples required to meaningfully compare two probability distributions extending over an exponential number of events, can be very problematic (see the simple examples in figs. 3.3 and 3.4). The boson sampling validation problem can, more precisely, be stated as the problem of efficiently ruling out that a set of data, produced by the alleged boson sampler, is not the result of some other probability distribution which may be implemented into the device. It is worth noting that one has to specify which alternative probability distributions one wants to be able to distinguish from the boson sampling one. Indeed, generally speaking, the more a probability distribution is close to the boson sampling one, the higher the number of required samples to discriminate them will be, as shown in a simple case in example 7. Example 7 (Discriminating between similar probability distributions). Consider two probability distributions, p1 and p2, defined over the two events 0 and 1 as p1(x = 0) = 1/2, p1(x = 1) = 1/2, (4.1a) p2(x = 0) = 1/2 + , p2(x = 1) = 1/2 − . (4.1b) The average number of samples required to distinguish between p1 and p2, with confidence α, is given by Nmin = log α x∈events p1(x) log p1(x) p2(x) . (4.2) In other words, Nmin is the average length of a dataset D such that the P(p1 | D) P(p2 | D) > α, (4.3) where P(p | D) is the probability that the observed set of samples D has been gener- 56
  • 57.
    ated by theprobability distribution p. Applying eq. (4.2) to eq. (4.1) we obtain Nmin ≈ log α 2 2 + O(1), for → 0. (4.4) Thus as expected, decreasing , and thus increasing the “similarity” of the two distri- butions, the number of required samples increases indefinitely. From the experimental point of view, this poses the very concrete issue of finding a way to certify that a built experimental apparatus is working as aspected. From a more fundamental standpoint instead, an unsolved validation problem threatens to crucially undermine the theoretical claim of observing a post-classical computation: if we are given by an untrusted party an alleged boson sampler as a black box, which efficiently produces a number of samples faster than any classical computer can sample from the boson sampling distribution, how can we somehow certify that that output is what it claims to be? Indeed, an issue may arise if the number of samples necessary to assess that the produced set of data does not come from an alternative probability distribution increases exponentially with n, thus making an experimental, verifiable implementation of the BosonSampling problem not scalable. When the boson sampling validation problem was first raised, it was argued to be fundamentally unsolvable, thus undermining the possibility of experimentally dis- criminate a boson sampling distribution even against the trivial uniform one [77]. AA subsequently showed that this was not true, providing an example of an efficient method to discriminate the boson sampling distribution from the uniform one [21]. This test, however, was not capable of efficiently separating, for example, the output of a boson sampler from that resulting from the evolution of classical particles into the interferometer. This validation test was successfully implemented in a number of experiments [14, 15]. Other tests, able to validate against distinguishable particles, were later proposed [14, 15, 16, 21, 22, 23] and experimentally implemented [14, 15, 17]. 4.2 Fourier suppression law Among the proposed validation schemes, those based on Zero-Transmission Laws (ZTLs) [24] (also often referred to as suppression laws) seem to be a promising approach to discriminate true many-boson evolution from alternative models. A ZTL can be thought of as a generalization of the HOM effect to many modes and photons. More precisely, it refers to the capability, occurring in some particular unitary matrices, to efficiently predict the strict suppression of certain input-output combinations of states. In other words, a ZTL allows to efficiently predict that for certain pairs of states (r, s), the scattering amplitude A(r, s) vanishes, without having to compute a permanent. 57
  • 58.
    Such suppressions, likethe HOM effect, are a pure quantum-mechanical effect, arising from the interference of many-body states during the evolution in the interferometer, and as such a strong signature of non-classical evolution. In the originally proposed BosonSampling scheme, where a random unitary matrix is employed, no such ZTL can be taken advantage of. However, these turn out to be useful when thinking to the validation problem in a slightly broader perspective. If we do not think of a boson sampler as a black box, but approach the more practical problem of assessing that an actual experimental apparatus is working as advertised, we can analyse the various possible sources of error. Roughly speaking, the two main factors potentially leading to an erroneous output probability distribution are: a wrong implemented unitary evolution, and non-indistinguishable input states. While the unitary matrix implemented in an interferometer can be efficiently verified [78], the indistinguishability of the input photons is harder to assess, and it is in this context that ZTLs can be exploited. Provided that one can change at will the unitary matrix implemented in the device - which is in principle possible to do efficiently [79] and has been recently reported in an experimental setup [17] - then it is possible to first verify the indistinguishability of the sources by injecting the photons in a specially crafted unitary, whose symmetries allow to efficiently observe a ZTL, and then on the same device to implement a random unitary matrix and perform the boson sampling experiment. The first proposed example of such a ZTL [24] employed Fourier matrices: Definition 4 (Fourier matrices). Let UF (m) be the m-dimensional matrix defined element-wise by [UF (m)]j,k ≡ 1 √ m exp 2πi m (j − 1)(k − 1) , j, k = 1, . . . , m. (4.5) We will refer in the following to this matrix as the m×m Fourier matrix. The m dependence will be omitted when clear from the context. It was shown [22, 24, 65, 66] that with Fourier matrices, a b-periodic input state r never evolves in an output state s such that b n j=1 Sj is not a multiple of the number of modes m, where S is the MAL introduced in definition 1, and a b-periodic state is defined as follows: Definition 5 (Periodic states). Given a state r, we say that r is b-periodic if its MOL representation has the form r ≡ (r1, . . . , rb, r1, . . . , rb, . . . , r1, . . . , rb) m=p·b , (4.6) 58
  • 59.
    √ 2 UF (2) = 11 1 −1 √ 3 UF (3) =     1 1 1 1 e 2πi 3 e −2πi 3 1 e −2πi 3 e 2πi 3     √ 4 UF (4) =      1 1 1 1 1 i −1 −i 1 −1 1 −1 1 −i −1 i      Table 4.1: Examples of Fourier matrices where b is the length of each of the identical sublist (bins) composing r, and p = m/b the number of such sublists. We will refer to b as the periodicity of r. Being UF symmetric, that is, UF i,j = UF j,i for every i and j, the scattering amplitudes are also symmetric with respect to input and output states: A(r → s, UF ) = A(s → r, UF ). In conclusion, the Fourier suppression law [24] can be stated as follows: Theorem 1 (Fourier suppression law). Let UF (m) be the m×m Fourier matrix, and let r and s be two n-boson states, with corresponding MAL representations R and S. Then, if r is b-periodic and s is such that mod  b n j=1 Rj, m   = 0, (4.7) then the following suppression law holds: A(r → s, UF ) = A(s → r, UF ) = 0. (4.8) We note that the converse in in general not true. In other words, this condition is sufficient but not necessary for an input-output pair to be suppressed. The many-boson matrices corresponding to Fourier matrices with m = 4, 6, 8, and n = 2, are shown in figs. 4.1 to 4.3. In these figures, the periodic states are highlighted in green, and the strictly suppressed states are the white squares. The input-output pairs predicted to be suppressed by theorem 1 are marked by purple circles. The majority of the suppressed pairs correspond to periodic states, and are predicted by theorem 1. Figure 4.3 shows however that theorem 1 gives a sufficient but not necessary condition for a pair to be suppressed, as there are in this case pairs which are suppressed but do not correspond to periodic states. 59
  • 60.
    Fourier matrix Randommatrix a) b) Figure 4.1: Many-boson matrices for m = 4 and n = 2. Modulus squared of the elements of the many-boson matrices U(4, 2) corresponding to the 4×4 Fourier matrix and a randomly chosen 4×4 unitary matrix. (a) Periodic states are highlighted in green, and input-output pairs predicted to be suppressed by theorem 1 are marked by purple circles. All the suppressed pairs are here predicted to be suppressed. In other words, in this case, the Fourier suppression law completely characterizes the suppressed input-output pairs. (b) Set of scattering probabilities resulting from a randomly chosen unitary matrix. No evident structure arises in this case. The calculation of the elements of the matrix requires the computation of the corresponding permanents, as no efficient method is known to calculate many-body scattering amplitudes for random matrices. Probabilities lesser than 10−2 are shown as “≈ 0” for the sake of clarity. 60
  • 61.
    Figure 4.2: Modulussquared of the elements of the many-boson matrix U(6, 2, UF ), corresponding to the 6×6 Fourier matrix injected with 2-photon states. Strictly suppressed input-output pairs are marked with purple circles, and periodic states highlighted in green. Similarly to fig. 4.1, all of the suppressed pairs are covered by theorem 1, which is shown here by the fact that the white elements of the matrix are always marked with purple circles. 61
  • 62.
    Figure 4.3: Modulussquared of the elements of the many-boson matrix U(8, 2, UF ), corresponding to the 8×8 Fourier matrix injected with 2-photon states. Strictly suppressed input-output pairs are marked with purple circles, and periodic states highlighted in green. Differently from figs. 4.1 and 4.2, not all of the suppressed pairs are in this case predicted by theorem 1. As an example, there a number of output suppressed pairs corresponding to the input state (6, 8), which has not any periodicity. This is possible because theorem 1 gives only a sufficient, but not necessary condition for a pair to be suppressed. 62
  • 63.
    The validation schemewhich exploits the Fourier suppression law, proposed in [22], uses a single fixed input state with MAL RTichy = (1, np−1 + 1, 2np−1 + 1, . . . , (n − 1)np−1 + 1), (4.9) injected into an inteferometer implementing the m-dimensional Fourier matrix, with m = np and p some positive integer. RTichy is an especially crafted n-periodic state, corresponding to an high fraction of output suppressed states. Counting the number of states corresponding to a suppressed input-output pair - that is, an input-output pair that would be suppressed if the evolution was an ideal quantum-mechanical one - it is possible to quantify how much the observed output probability distribution differs from the ideal boson sampling one. More precisely, the violation parameter is defined as D = Nforbidden/Nruns, (4.10) where Nforbidden is the number of observed events violating theorem 1, and Nruns is the total number of events. This parameter has been shown [22] to be a better signature of quantum many-body statistics than precedently proposed validation schemes. In particular, it can be used to validate a boson sampler against alternative hypotheses such as the Mean-Field Sampler [22]. A Mean-Field Sampler is an efficiently evaluable and physically plausible alternative model [22], which reproduces some aspects of many-boson dynamics surviving in the semiclassical limit. This model is not distin- guished from the boson sampler distribution by other validation schemes, making the Fourier validation test more capable of capturing the signatures of genuine many-boson quantum interference. A summary of the validation protocol based on Fourier matrices is given in fig. 4.4. 63
  • 64.
    Figure 4.4: Validationscheme with Fourier matrices. Conceptual scheme of the protocol: the possible configurations of n photons at the output of an m-mode interferometer can be divided into two categories, unforbidden and forbidden, depending on whether they satisfy or not the suppression confition of theorem 1, respectively. The pie charts show the expected output statistics with different classes of particles, where green and red areas represent events with unforbidden and forbidden outputs, respectively. The injection of a cyclic Fock state (beige box) in an m-mode Fourier interferometer results in total suppression of forbidden output states. Cyclic states with distinguishable particles (blue box) show no suppression, being each output combination equally likely to occur. A Mean Field state [22] (purple box), a semi-classical state reproducing some of the features of bosonic statistics, shows suppression with highly reduced contrast. Therefore, with a cyclic input the m-mode Fourier interferometer is able to discriminate, through the measurement of degree of violation D = Nforbidden/Nevents, which of these three hypotheses the input state belongs to. Credits: [25]. 64
  • 65.
    4.3 Fourier suppressionlaw and scattershot boson sam- pling The original scheme for implementing a boson sampler using a single SPDC source is not scalable to higher number of photons, as discussed in the earlier sections. Scattershot boson sampling is, currently, the most promising architecture to implement Boson- Sampling with higher number of photons. This scheme, however, is essentially based on the injection of random input states into the interferometer. On the other hand, the validation scheme based on the Fourier suppression law only works for particular, periodic input states. Therefore, a scattershot boson sampling experiment cannot be validated making use of the Fourier suppression law, as originally proposed. For a validation scheme based on some ZTL to be employable with a scattershot boson sampler, it is no more sufficient to have an high number of output suppressed states corresponding to some small class of input states. Instead, it is required that the fraction of pairs of suppressed input-output states must be a non-negligible fraction of the total number of states. In other words, considering that all input states in a scattershot boson sampler are collision-free, and approximating all output states to also be collision-free (which is justified in the m n limit [80]), we must have suppressed input-output configurations m n 2 > 0 by a statistically significant amount. (4.11) In this section we consider whether this condition is verified for the Fourier sup- pression law, thus making it viable to use in scattershot boson sampling experiments. We will proceed by counting how many input states have some kind of periodicity, according to eq. (4.6), and are thus such that theorem 1 can be applied to them. For the same reasons given above, we will restrict our analysis to the collision-free states. To count the number N(m, n) of periodic collision-free configurations (periodic CFCs) of n photons in m modes, we first need to know for which values of b there can be b-periodic configurations, given m and n. In other words, we want all the integers b such that the MOL corresponding to the state can be written in the form (4.6). Given that a partition of the MOL of the form (4.6) can be found for each p = 1 which is a common divisor of m and n, we have that the possible b are all and only the integers of the form b = m/p for such a p. Formally, we say that the set B(m, n) of all possible bin lengths b for n photons in m modes is B(m, n) = {b ∈ N | b = m/p for some p = 1 common divisor of m and n}. (4.12) Denoting with S(m, n, b) the set of all m-modes, n-photons, b-periodic CFCs, we 65
  • 66.
    can now seethat N(m, n) = b∈B(m,n) S(m, n, b) . (4.13) The number of CFCs in S(m, n, b) is equal to the number of lists of length m composed of only 0 and 1 that are decomposable in p = m/b identical sublists of length b each. This is in turn equal to the number of ways in which we can choose n/p = nb/m elements from a set of b elements, i.e. the binomial factor b nb/m , and we conclude that |S(m, n, b)| = b nb/m . (4.14) A naive approach would be now to say that N(m, n) is equal to the sum of |S(m, n, b)| for each b ∈ B(m, n). This however overcounts all those states having multiple periodicities. For example in the case m = 8, n = 4, where B(m, n) = {2, 4}, the states q = (0, 1, 0, 1, 0, 1, 0, 1) and q = (1, 0, 1, 0, 1, 0, 1, 0) (4.15) are at the same time 2- and 4-periodic, and would each be counted twice if we were to say that N(8, 4) = |S(8, 4, 2)| + |S(8, 4, 4)| = 8. The issue can be traced back to the fact that S(8, 4, 2) ∩ S(8, 4, 4) = ∅. More generally, the intersection S(m, n, b1) ∩ S(m, n, b2) will be non-empty whenever there is some integer b0 ∈ B(m, n) common divisor of b1 and b2. The problem of counting the periodic CFCs is thus reconducted to that of finding the number of elements in the union of a set of generally not-disjoint subsets, resulting in the general formula N(m, n) = ∞ k=1 (−1)k+1 (b1,...,bk) |S1 ∩ · · · ∩ Sk| , (4.16) where we denoted for brevity Si ≡ S(m, n, bi), and the inner sum is over all unordered lists of k elements of B(m, n). With reasonings similar to the above ones it is possible to show that |S(m, n, b1) ∩ · · · ∩ S(m, n, bk)| = |S(m, n, GCD(b1, . . . , bk))| , (4.17) where GCD(b1, . . . , bk) denotes the greatest common divisor of b1, . . . , bk ∈ B(m, n). Equation (4.16), together with (4.17) and (4.14), provides us with a general formula to compute the exact number of periodic CFCs of given m and n. As an example of the application of this formula we again take the case m = 8, n = 4, which now yields the correct result: N(8, 4) = |S(8, 4, 2)| + |S(8, 4, 4)| − |S(8, 4, 2) ∩ S(8, 4, 4)| = |S(8, 4, 2)| + |S(8, 4, 4)| − |S(8, 4, 2)| = |S(8, 4, 4)| = 4 2 = 6. 66
  • 67.
    While exact, expression(4.16) takes a lot of computational time due to the exponen- tially growing number of required permutations. In practice, already the computation of the number of periodic CFCs with m ≈ 120 and n ≈ 60 is very time-consuming. Computational times can be significantly reduced noting that b1| b2 =⇒ S(m, n, b1) ⊆ S(m, n, b2) (4.18) where the vertical bar indicates that b1 divides b2. We can therefore apply eq. (4.16) with the slight modification of taking bi ∈ ˜B(m, n) instead of bi ∈ B(m, n), where ˜B(m, n) is the subset of elements of B(m, n) which do not divide any other element of B(m, n), i.e. ˜B(m, n) ≡ {b ∈ B(m, n) : ( b ∈ B(m, n) : b|b )}. (4.19) Using (4.16) together with (4.19) we can easily compute the number of periodic CFCs up to and beyond m ≈ 1000. Faster exact results can be achieved in many particular cases. For example, for m = 2p for some integer p, and n an even number, reasonings similar to the above ones lead to (see fig. 4.5) N(m, n) = m/2 n/2 . (4.20) Figure 4.5: Number of periodic states vs number of photons n, for various values of the number of modes m. The orange line is m/2 n/2 , which exactly reproduces the number of periodic configurations for all even n as predicted by (4.20). More generally, for m 1 the number of periodic states is dominated by m/a n/a with a = m/ GCD(m, n) (see figs. 4.6 and 4.7). 67
  • 68.
    1000 2000 30004000 n 200 400 600 800 1000 1200 1400 Figure 4.6: Logarithm of the number of periodic states of m/2 photons in m modes, plotted as a function of m. The coloured dashed lines are, in descending slope order, log m/2 m/4 , log m/3 m/6 , log m/5 m/10 , log m/7 m/14 . 5 10 15 20 25 30 35 n 5 10 15 20 25 30 35 10 20 30 40 n 1000 2000 3000 4000 5000 6000 7000 10 20 30 40 50 n 1×106 2×106 3×106 4×106 5×106 Figure 4.7: Number of periodic states vs number of photons. Blue points: number of periodic states calculated with eq. (4.16). The continuous green, cyan, purple and red lines are respectively m/2 n/2 , m/3 n/3 , m/5 n/5 and m/7 n/7 . Unlike the number of periodic collision-free configurations, an efficient algorithm to compute the number of suppressed output states predicted by Tichy’s suppression law for Fourier matrices has not been found yet, except for very particular input states [22]. A brute force computation of the number of suppressed output states for each input state for various values of m and n is shown in figs. 4.8 and 4.9. These computations show that only a small fraction of input states results in a significant fraction of output suppressed states. However, the above result regarding the number of periodic states is enough to make the Fourier suppression law unsuitable to validate scattershot boson sampling experiments. In fact, in the best case scenario in which n = 2p for some p, the number of periodic states is ≈ m/2 n/2 (see eq. (4.20)), which is negligibly small compared to m n . In particular, for m = n2 , the fraction n2/2 n/2 / n2 n decreases exponentially with n. This implies that, even if all of these input states were to correspond to an 100% fraction of output suppressed states, which is far from true, nevertheless in the regime of high n virtually no suppressed pair would ever be sampled in a actual scattershot experiment, and the validation scheme would be inefficient. 68
  • 69.
    5 10 1520 25 0 20 40 60 80 Inputs %ofsuppressedstates 8 modes, 2 photons 0 20 40 60 80 100 120 0 20 40 60 80 Inputs %ofsuppressedstates 16 modes, 2 photons 0 10 20 30 40 50 60 70 0 20 40 60 80 Inputs %ofsuppressedstates 8 modes, 4 photons 0 500 1000 1500 0 20 40 60 80 Inputs %ofsuppressedstates 16 modes, 4 photons 5 10 15 20 25 0 20 40 60 80 Inputs %ofsuppressedstates 8 modes, 6 photons 0 2000 4000 6000 8000 0 20 40 60 80 Inputs %ofsuppressedstates 16 modes, 6 photons a) b) c) d) e) f) Figure 4.8: Ratio of suppressed collision-free output states For each input collision-free state, the percentage of output suppressed collision-free states is shown for Fourier matrices (blue dots) and Sylvester matrices (red dots). All the points are calculated using eq. (3.4). In all the cases shown the number of suppressed output states is significantly higher for the Sylvester matrix. (a) and (b): For two input photons, all of the output suppressed states are predicted by theorems 1 and 2 in the case of Fourier and Sylvester matrices, respectively. Green dashed line: exact fraction of suppressions predicted by eq. (4.24). (c) and (d): For n = 2p only a subset of suppressed pairs are predicted by theorems 1 and 2. Green dashed line: estimate of the fraction of output suppression states, equal to 1 − 1/n. The approximation gets closer to the true value for large values of n and m. (e) and (f): For n not equal to a power of 2, no estimate of the number of output suppressed states is given, neither in [22] nor in [23]. 69
  • 70.
    0 1000 20003000 4000 0 20 40 60 80 Inputs %ofsuppressedstates 16 modes, 5 photons 0 2000 4000 6000 8000 10 000 12 000 0 20 40 60 80 Inputs %ofsuppressedstates 16 modes, 8 photons a) b) Figure 4.9: Ratio of suppressed collision-free output states for each input in a 16-mode interferometer injected with 5 (a) and 8 (b) photons. Like fig. 4.8, the ratio of suppressed output states is significantly higher for almost all input states for the Sylvester matrix than the Fourier matrix. 4.4 Sylvester suppression law Besides the Fourier matrix, a ZTL has been shown to exist also for another important class of unitary matrices [23], the so-called Sylvester matrices. Definition 6 (Sylvester matrices). Let US be an m-dimensional unitary matrix of the form US (m) ≡ H(m)/ √ m, with H(m) defined recursively as H(2p ) = H(2p−1 ) H(2p−1 ) H(2p−1 ) −H(2p−1 ) , (4.21) for each positive integer p, and with H(1) = 1. We will refer in the following to US (m) as normalized Sylvester matrix and to H(m) as Sylvester matrix. The m dependence of US and H will be omitted when clear from the context. An analytic expression for the element (i, j) of a Sylvester matrix can be given in the form: [H(2p )]i,j = (−1)iB jB , (4.22) where iB and jB are the binary representations of i and j, with rows and columns enumerated from zero to m − 1, and is the bitwise dot product. For example, for m = 8, i = 2, j = 7 we would have [H(8)]2,7 = (−1)(0,1,0)·(1,1,0) = (−1)0×1+1×1+0×0 = −1. (4.23) 70
  • 71.
    H(2) = 1 1 1−1 H(4) =      1 1 1 1 1 −1 1 −1 1 1 −1 −1 1 −1 −1 1      H(8) =         1 1 1 1 1 1 1 1 1 −1 1 −1 1 −1 1 −1 1 1 −1 −1 1 1 −1 −1 1 −1 −1 1 1 −1 −1 1 1 1 1 1 −1 −1 −1 −1 1 −1 1 −1 −1 1 −1 1 1 1 −1 −1 −1 −1 1 1 1 −1 −1 1 −1 1 1 −1         Table 4.2: Examples of Sylvester matrices We will refer in the following to US (m) as normalized Sylvester matrix and to H(m) as Sylvester matrix. The m dependence of US and H will be omitted when clear from the context. It was recently proved [23] that injecting two bosons into any pair of (different) modes of a multi-mode interferometer implementing a Sylvester matrix, the total fraction of suppressed output states is Nsupp Ntot = m 2(m+1) . Considering only collision-free states, the fraction becomes Nsupp Ntot,CF = m 2(m − 1) . (4.24) More generally, the following suppression law holds [23]: Theorem 2 (Sylvester suppression law). Given an m×m Sylvester matrix, m = 2p , n = 2q , q < p, and an n-boson input state with MAL S of the form S = (1 + nc, . . . , n + nc), (4.25) for some integer c such that 0 ≤ c ≤ (m/n − 1), then the strictly suppressed output states are all and only those R ≡ (R1, R2, . . . , Rn) such that R1,B ⊕ · · · ⊕ Rn,B = 0 ≡ (0, . . . , 0 q ), (4.26) where Ri,B is the q-dimensional vector obtained from the binary representation of Ri − 1 truncated to the q least significant digits, and ⊕ is the bitwise XOR operation (equivalently the elementwise, modulo 2, sum). Similarly to the case of Fourier matrices, an estimate of the fraction of strictly suppressed states predicted by this law is: V = Nsupp Ntot ≈ 1 − 1 n . (4.27) 71
  • 72.
    Theorem 2 predictsan high ratio of suppressed output configurations for all input states, when n = 2, as also shown in the numerical computations in figs. 4.8 and 4.9, and in the many-boson matrices shown in figs. 4.10 and 4.11 On the other hand, this suppression law predicts in the general case of n > 2 photons an high ratio of suppressed output configurations only for the m/n input states of the form (4.25). This is a negligibly small fraction of the total number of m n collision-free configurations, apparently making Sylvester matrices even worse than Fourier matrices to validate a scattershot boson sampler. However, a brute force computation with eq. (2.18) shows (see figs. 4.8 and 4.9) that the number of suppressed input-output pairs in a Sylvester matrix is significantly higher than what predicted by the above described suppression law. As will be showed in the following chapter, it is however possible to efficiently predict a greater number of suppressed pairs, by generalizing the ZTL proposed in [23]. 72
  • 73.
    Fourier matrix, m=8,n=2 Sylvester matrix, m=8, n=2 Fourier matrix, m=8, n=4 Sylvester matrix, m=8, n=4 Figure 4.10: Comparison of Fourier and Sylvester many-boson matrices. Matrices of scattering amplitudes between all combinations of pairs of input-output collision-free states of 2 and 4 photons into 8 modes, for Sylvester and Fourier matrices. White: suppressed input-output pair. Orange: non- suppressed input-output pair. The significantly higher fraction of suppressed pairs of Sylvester matrices with respect to Fourier matrices is clearly evident in these plots. 73
  • 74.
    Figure 4.11: Modulussquared of the elements of the many-boson matrix U(8, 2, US ), corresponding to the 8×8 Sylvester matrix injected with 2-photon states. The structure of this matrix is much more simple than that shown in figs. 4.2 and 4.3. Moreover, all of its elements can be efficiently predicted with Crespi’s suppression law, as opposite to the case of the Fourier matrix, in which only some of the events corresponding to a periodic input or output can. However, when more than two photons are considered, not all of the elements of the many-boson matrix are efficiently predicted by theorem 2. 74
  • 75.
    Chapter 5 Theoretical results Asdiscussed in section 4.4, the suppression law given in theorem 2 [23] is not suitable to validate scattershot boson sampling experiments. On the other hand, as shown in figs. 4.8 and 4.9, the actual fraction of suppressed pairs is significanly higher, and some kind of suppression law seems to hold for every input state, not just for those considered by theorem 2. In this chapter we propose a generalization of theorem 2, which predicts a much higher fraction of suppressed pairs, and is thus more suitable to validate scattershot boson sampling experiments. The proposed suppression law is described in section 5.2, using notations and lemmas given in section 5.1. In section 5.3 we discuss a validation scheme to distinguish quantum many-boson interference (up to small errors) from the output resulting from distinguishable photons, by exploiting the proposed suppression law. In this context we also derive an analytic expression for the mean number of samples necessary to experimentally distinguish a nearly quantum evolution from the classical hypothesis. The chapter closes with a summary of the results in section 5.4. 5.1 Preliminaries For the results presented in this chapter we will introduce some new notations, on top of those given in definition 1: Definition 7 (Binary Matrix representation). Let r = (r1, . . . , rm) be the MOL of a state of n particles in m = 2p modes, for some positive integer p, and let R = (R1, . . . , Rn) be the corresponding MAL. We denote with R (2) i the binary repre- sentation of Ri − 1, padded with zeros on the left so as to make it a vector of length p, and with Ri,α the α-th element of R (2) i . This construction allows us to represent a many-body state r as a binary matrix (BM) R, i.e. as an n×p-dimensional matrix 75
  • 76.
    whose elements areeither 0 or 1. We will refer to R as the BM representation (or simply the BM) of the state r. Remark 1. We note that for indistinguishable particles it is not meaningful to assign a specific mode to a particle. This means that the order of the elements of the MAL describing a state of many indistinguishable particles is not relevant. In other words, if two MALs R and R differ only for the order of their elements, they represent the same physical state. Similarly, if two binary matrices R and R can be obtained one from the other with only a reordering of the rows, they represent the same physical state. Example 8. Consider the case m = 4 and n = 2, and the state with MOL r ≡ (0, 1, 1, 0) and corresponding MAL R = (2, 3). Following the notations introduced above we will have R1 = (0, 1), R2 = (1, 0), and the BM of r will be the 2×2 matrix R = 0 1 1 0 . Example 9. Consider the case m = 8, n = 4, and the state with MOL s ≡ (1, 1, 1, 0, 0, 0, 0, 1) and corresponding MAL S = (1, 2, 3, 8). Following the no- tations introduced above we will have S1 = (0, 0, 0), S2 = (0, 0, 1), S3 = (0, 1, 0), S4 = (1, 1, 1), and the BM representation of r will be the 4×3 matrix S =      0 0 0 0 0 1 0 1 0 1 1 1      . Definition 8. Let R be an n×p BM, and let A be a subset of the p columns of R. Then we denote with NA (R) the matrix obtained by negating the columns of R specified in A. In other words, expliciting the indices, NA (R) is defined as [NA (R)]k,α ≡    Rk,α ⊕ 1, for α ∈ A, Rk,α, for α /∈ A. (5.1) where ⊕ is the sum modulo 2. Similarly, if r is a MOL representing some state of m = 2p modes, we denote with NA (r) the MOL obtained applying to r the following three operations: • Convert r to the binary matrix R, 76
  • 77.
    • Negate thecolumns of R specified in A, obtaining the binary matrix NA (R), • Find the MOL corresponding to the binary matrix NA (R). Example 10. Consider the case m = 8, n = 4 and the state with MOL r = (1, 1, 1, 1, 0, 0, 0, 0), corresponding to the MAL R = (1, 2, 3, 4), and expressed in the BM representation as R =      0 0 0 0 0 1 0 1 0 0 1 1      . Then, following the notations introduced above, we have when A consists of a single element, N{1} (R) =      1 0 0 1 0 1 1 1 0 1 1 1      , N{2} (R) =      0 1 0 0 1 1 0 0 0 0 0 1      , N{3} (R) =      0 0 1 0 0 0 0 1 1 0 1 0      . From this we see that N{2} (R) and N{3} (R) are equal to R, up to a permutation of the rows, and they therefore represent the same physical state. On the other hand, N{1} (R) cannot be made equal to R only rearranging its rows, and therefore represents a different state. In other words, we have: N{1} (r) = (0, 0, 0, 0, 1, 1, 1, 1) = r, N{2} (r) = r, N{3} (r) = r. On the other hand, the set A can contain more than a single element. For example: N{1,3} (R) =      1 0 1 1 0 0 1 1 1 1 1 0      , N{1,2,3} (R) =      1 1 1 1 1 0 1 0 1 1 0 0      , which, for the same reasons, satisfy N{1,3} (r) = (0, 0, 0, 0, 1, 1, 1, 1) = r, N{1,2,3} (r) = (0, 0, 0, 0, 1, 1, 1, 1) = r. The following results will also be useful in the following: Lemma 1. For any two integers a and b we have (−1)a+b = (−1)a⊕b . (5.2) 77
  • 78.
    More generally, forany finite set of integer numbers {a1, . . . , aN }, we have (−1) N k=1 ak = (−1) N k=1 ak . (5.3) Lemma 2. Let Sn be the set of permutations of {1, . . . , n}, and let τ ∈ Sn be a permuta- tion different from the identity such that τ2 = 1 (i.e. τ(τ(k)) = k for each k ∈ {1, . . . , n}). Then we can univocally associate to each σ ∈ Sn another (different) permutation στ ≡ τ◦σ, where ◦ denotes the composition of permutations. Lemma 3. Let R and NA be as in definitions 1 and 8, respectively. Then • NA (r) = r if and only if NA (R) = Rτ for some permutation τ ∈ Sn, where with Rτ we denote the matrix obtained from R permuting the rows according to τ, i.e. (Rτ )k,α ≡ Rτ(k),α for each k = 1, . . . , n. • If τ is such that NA (R) = Rτ , then 1. τ2 = 1, 2. τ(k) = k for each k = 1, . . . , n, 3. for each σ ∈ Sn, NA (Rσ) = Rτ◦σ, 4. for all columns α ∈ A, Rα will have an equal number of 1s and 0s, 5. for all columns α /∈ A, Rα will have an even number of 1s, and an even number of 0s. 5.2 Main results Using the lemmas and definitions listed above, we are now ready to state the main result of the section: Theorem 3 (Sufficient condition for suppressed pairs). Let r and s be two many- boson states of n particles in m = 2p modes, let R and S be the corresponding BM representations, and let A be a subset of the columns of R. If the following conditions    NA (r) = r, n k=1 α∈A Skα = 1. (5.4a) (5.4b) are simultaneously satisfied, then A(r, s, US ) = A(s, r, US ) = 0. The converse does not in general hold. In other words, this condition is sufficient 78
  • 79.
    but not necessaryfor a pair to be suppressed. Proof. From now on, we will use the abbreviation A ≡ A(r, s, US ), leaving implicit the dependence on the states and the unitary evolution considered. The transition amplitude A can be expressed in terms of a permanent of the form A = 1 √ r1! . . . rm!s1! . . . sm! perm US [R|S] , (5.5) where perm(U) denotes the permanent of U, and the notation US [R|S] was given in definition 3. Using definition 6 and eq. (4.22), we get from eq. (5.5) A = C σ∈Sn n k=1 (−1)Rσ(k) Sk = C σ∈Sn (−1)ER,S(σ) , (5.6) where we denoted with Rσ(k) Sk the bitwise dot product between Rσ(k) and Sk, defined as Rσ(k) Sk ≡ p α=1 Rσ(k),αSk,α, C is a constant factor, and we defined the expression at the exponent of the last expression (see lemma 1) as ER,S(σ) ≡ n k=1 Rσ(k) Sk = n k=1 p α=1 Rσ(k),αSk,α. (5.7) In the following, we will omit for brevity the dependence of the exponent E from the states and write the above expression as simply E(σ). Note that the actual value of E(σ) is not important here: the only relevant detail is its parity. Since we need to evaluate only the cancellation of the scattering amplitude A, we will ignore the constant factor C and focus on the sum over the permutations in eq. (5.6). For this sum to vanish, it is necessary and sufficient that for exactly half of the permutations we have (−1)E(σ) = 1. In other words, the scattering amplitude A vanishes if and only if the number of permutations σ such that E(σ) is even, is equal to the number of permutations σ such that E(σ ) is odd. A necessary and sufficient condition for this to hold, is that to each permutation σ we can univocally associate another different permutation σ such that E(σ ) = 1 ⊕ E(σ). But if eq. (5.4a) holds, then by lemma 3 and lemma 2 we can univocally associate to each σ the (different) permutation στ ≡ τ ◦ σ, where τ is a permutation such that NA (R) = Rτ . Using στ in eq. (5.7) we have E(στ ) = n k=1 p α=1 Rτ(σ(k)),αSk,α = n k=1   α∈A Rτ(σ(k)),αSk,α ⊕   α/∈A Rτ(σ(k)),αSk,α     . (5.8) 79
  • 80.
    Using now theexplicit expression for NA (R) given in definition 8 and the last result of lemma 3, we have NA (Rσ) = Rτ◦σ ⇐⇒    1 ⊕ Rσ(k),α = Rτ(σ(k)),α, α ∈ A, Rσ(k),α = Rτ(σ(k)),α, α /∈ A. (5.9) Inserting these equations in (5.8), we finally obtain E(στ ) = E(σ) ⊕ n k=1 α∈A Sk,α = E(σ) ⊕ 1, (5.10) where in the last step we exploited eq. (5.4b) for S. Inserting this last result into eq. (5.6) we conclude that A = C σ∈Sn (−1)ER,S(σ) = C σ∈Sn: E(σ) even (−1)E(σ) + (−1)E(στ ) = 0, (5.11) which proves that the input/output pair (r, s) is suppressed. Example 11. Let r be the MOL given in example 10. Then NA (r) = r holds for A = {2}, A = {3}. and A = {2, 3}. Theorem 3 predicts that all output states s, which BM representation S has an odd number of 1s in the second or third column, are suppressed. Moreover, all output states with BM S having a total number of 1s in the second and third column, are suppressed. More explicitly, all output states s such that n k=1 Sk,2 = 1 or n k=1 Sk,3 = 1 or n k=1 [Sk,2 ⊕ Sk,3] = 1, are predicted to be suppressed. For example, the states S = (3, 6, 7, 8), S = (2, 6, 7, 8), and S = (4, 6, 7, 8), having BM representations      0 1 0 1 0 1 1 1 0 1 1 1      ,      0 0 1 1 0 1 1 1 0 1 1 1      , and      0 1 1 1 0 1 1 1 0 1 1 1      , respectively, are all suppressed. Remark 2 (Efficiency). To check if theorem 3 applies to a given input-output pair, one has to verify condition 5.4 for each of the 2p − 1 = m − 1 possible (non empty) subsets of the p columns of R and S. Moreover, the number of elementary operations necessary 80
  • 81.
    to verify condition5.4 for a given set of columns increases polynomially in n. In conclusion, the amount of computational resources required to verify the hypotheses of theorem 3 increases polynomially in n and m, and the proposed suppression law is therefore efficiently verifiable. In fig. 5.1 is shown how the computational times scale polynomially in n with the suppression law, as opposite to exponentially in n with the brute-force method of computing the permanents. 4 5 6 7 8p 0 5 10 15 20n 0.001 0.01 0.1 1. 3 4 5 6 7 8 p 0 100 200 n 0.0 0.2 0.4 ss a) b) Figure 5.1: (a) Computational times required to assess whether a randomly sampled pair is suppressed, by using the brute-force method of calculating the permanent (in green) or by using the suppression law (blue), plotted in logarithmic scale as a function of the number of n and p = log2 m. Continuous lines: linear fits of the data, showing how the computational time with the brute-force method scales exponentially in n, as opposite to the linear scaling in n required by the suppression law. (b) Average number of seconds required to assess if a randomly sampled pair is suppressed using the suppression law. The linear fits show that the scaling is not only polynomial, but roughly linear in n for each m (we note that the interesting regime for boson sampling experiments is n m, therefore the irregularities shown for n ≈ m are negligible). While eq. (5.4b) gives a sufficient condition for an input-output pair to be suppressed, this condition is not necessary. For the majority of input states, not all of the suppressed output states satisfy theorem 3. For a more precise analysis of these quantities the follow- ing definitions will be useful: Definition 9. Let U be an m×m unitary matrix. Then we will denote with Ω(U) the set of input-output (collision-free) suppressed pairs, and with Ωeff(U) those predicted to be suppressed by theorem 3: Ω(U) ≡ {(R, S) ∈ Qn,m×Qn,m | perm(U[R|S]) = 0} , (5.12) Ωeff(U) ≡ (R, S) ∈ Ω(U) | NA (R) = Rτ and ⊕k,α Sk,α = 1 . (5.13) In other words, Ω(U) is the set of all n×n submatrices of U whose permanent vanishes, while Ωeff(U) is the subset of the pairs in Ω(U) for which we can efficiently predict 81
  • 82.
    that the permanentvanishes. Clearly, we have Ωeff(U) ⊂ Ω(U) for any U. An interesting question regards the evaluation of the ratio Ωeff(US ) / Ω(US ) , i.e. the amount of suppressed pairs caught by theorem 3, which will be analyzed in some detail in the next section. There is however at least one class of inputs for which theorem 3 seems to completely characterize the output suppressed states (that is, the suppressed output states are all and only those predicted to be suppressed by theorem 3): Conjecture 1. Let r and s be two many-boson states of n = 2q bosons in m = 2p modes, let R and S be the corresponding BM representations, let A be a subset of the columns of R, and let α(r) be the number of subsets A such that condition 5.4a holds, that is α(r) ≡ {A ⊆ P({1, . . . , p}) : NA (r) = r} , (5.14) where we adopted the standard convention of denoting with P(X) the power set of a set X. If α(r) = n − 1, then the output suppressed states are all and only those satisfying eq. (5.4b). Remark 3. Conjecture 1 predicts that for a (rather small) class of input states, the set of output suppressed states is completely characterized by theorem 3, and has been verified by brute-force computations for n = 2, 4, 8, 16. We note that this class of states includes as a subset those considered by theorem 2, for which the result has been proved in [23]. In the next section we will present the results of a series of numerical computations carried out to determine the dependence of the fraction of predicted suppressed pairs from m and n. 5.3 Bayesian validation scheme To assess if a given validation protocol is scalable at increasing n, it is useful to quantify the number of experimental runs needed to reject an alternative hypothesis with a given confidence level. Here we propose a simple model to predict the mean number of samples required to reject the hypothesis C of distinguishable photons, and accept an hypothesis Q of nearly indistinguishable photons, being a parameter quantifying the amount of distinguishability introduced by experimental errors. Let p ≡ Ωeff(US ) / m n 2 be the fraction of suppressed n-photon input/output pairs predicted by theorem 3. Given the unbiased nature of Sylvester matrices, p is also the probability of observing an event in a suppressed pair, when sampling uniformly from the set of all input-output pairs. The probability of observing an event in a suppressed pair in the Q hypothesis will be instead quantified by the parameter , and assumed 82
  • 83.
    to be asmall number pQ( ) = p (the ideal quantum case will therefore correspond to = 0). The only assumption of the proposed validation scheme is that for some unknown value 0, the probability of experimentally detecting an event in a suppressed pair is given by pQ( 0) = 0p. We will then give a formula to assess if a given set of experimen- tal data is more likely (with some fixed confidence level) to come from a Q evolution as opposite to the distinguishable case C. More in particular, the proposed validation scheme will accept a whole range of values of such that the relative probability of the data coming from Q as opposed to C will be greater then a confidence level α. Let M denote the total number of samples collected in an experiment, and let Msupp be the number of those samples which turned out to be a suppressed pair according to theorem 3. Then, assuming unbiased prior probabilities, the relative probability V of the hypothesis Q over the hypothesis C, having observed Msupp efficiently predicted suppressed pairs over M samples, is V ≡ P(Q | M, Msupp) P(C | M, Msupp) = P(M, Msupp| Q ) P(M, Msupp| C) . (5.15) In the case of completely distinguishable photons the probability is P(M, Msupp| C) = pMsupp (1 − p)M−Msupp (5.16) while if the dynamic of the system follows Q the probability is P(M, Msupp| Q ) = ( p)Msupp (1 − p)M−Msupp . (5.17) Using eqs. (5.16) and (5.17) inside eq. (5.15) gives V = Msupp (1 − p)M−Msupp (1 − p)M−Msupp , (5.18) and we will consider Q to be successfully distinguished from C, with the confidence level α, for all the values of such that V ≥ α. Let’s now denote with 0 the parameter quantifying the real experimental distin- guishability of the experiment (which is assumed to be unknown for the test). Then, on average, the number of detected suppressed pairs will be Msupp = 0pM. Plugging this into eq. (5.18) we obtain V =   0p 1 − p 1 − p 1− 0p   M . (5.19) In fig. 5.2 is shown an example of application of this test, using eq. (5.19). For a fixed set of values for M, α, 0, and p, the accepted models Q are those corresponding to 83
  • 84.
    the range ofvalues of such that V ≥ α. In fig. 5.2 this is equivalently expressed by plotting (V )1/M (blue line) against , and comparing it with the value of α1/M (purple line). All values of for which the blue line is above the purple one are accepted, and thus distinguished from the classical distribution with confidence α. Note how the accepted values of are those close to 0 (red dotted line), which is the parameter characerizing the actual distinguishability of the experimental apparatus. Figure 5.2: Accepted range of values of , for M = 300, α = 2000, 0 = 0.3, p = 0.1. The blue line is a fit of (V )1/M . The solid green line corresponds to V = 1, the dotted red line to = 0, and the dotted purple line to (V )1/M = α1/M . All the values of corresponding to the blue line being above the purple one are accepted by the test, with the confidence α. This range of accepted values is highlighted in the figure by the solid purple line. Inverting eq. (5.19) we can extract the average number M , 0 of experimental runs needed to reject C and accept Q with confidence α is M , 0 = log α 0p log + (1 − 0p) log 1− p 1−p . (5.20) As a function of , M , 0 has a positive minimum for = 0, as can be seen for example in fig. 5.3. Setting thus = 0 in eq. (5.20) we obtain that the average number of experimental samples required to assess with confidence α that the observed data are not compatible with the evolution of distinguishable photons, is M 0, 0 = log α 0p log 0 + (1 − 0p) log 1− 0p 1−p . (5.21) A rough approximation of eq. (5.21) for small 0 is M 0, 0 ≈ log α log 1 1−p + O( 0), (5.22) which for small p gives the scaling behaviour M 0, 0 ∼ log α/p. We therefore conclude that the number of experimental runs needed to validate against distinguishable photons scales polynomially in n if and only if the condition p > 1/ poly(n) (5.23) is satisfied, for small values of 0 and p. 84
  • 85.
    0.2 0.4 0.60.8 1.0 ϵ -500 500 1000 Mϵ,ϵ0 Figure 5.3: Needed number of samples M , 0 (eq. (5.20)) as a function of , for 0 = 0.3, p = 0.1, and α = 2000. The range of values of corresponding to V < 1 result in M , 0 < 0. For these the data is more likely to come from C than Q . The green dashed line highlights the value of = 0. 5.4 Conclusions Further work is needed to assess whether the fraction of suppressed pairs given by Sylvester matrices satisfies eq. (5.23). As shown in table 5.1 and fig. 5.4, the number Ω(U) of suppressed pairs is significantly larger for Sylvester matrices than for Fourier matrices. While it is clear that the former is better suited for scattershot sampling validation than the latter, the scaling behaviour of Ω(US ) can hardly be deduced from the data in table 5.1 and fig. 5.4-a. m n Sylvester Fourier 8 2 4/7 57.14% 12/49 24.49% 8 3 0 8/49 16.33% 8 4 96/175 54.86% 52/245 21.22% 8 5 4/7 57.14% 12/49 24.49% 8 6 4/7 57.14% 24/49 48.98% 8 7 0 0 16 2 8/15 53.33% 32/225 14.22% 16 3 0 64/1225 5.22449% 16 4 480/1183 40.57% 11112/207025 5.37% 16 5 480/1183 40.57% 632/24843 2.54% 16 6 53160/143143 37.14% 256/11011 2.32% 16 8 21952/83655 26.24% 495172/41409225 1.20% Table 5.1: Fractions of suppressed pairs (i.e. Ω(U)/ m n 2 ) for Sylvester and Fourier matrices, and various values of the number of modes m and the number of photons n. The values are calculated computing, with the permanent formula, all the scattering amplitudes between collision-free input-output states. 85
  • 86.
    8 16 32 64 128 m 5 10 1520n 0 20 40 % 8 16 32 64 128 m 5 10 15 20n 0 10 20 30 40 a) b) % Figure 5.4: Estimated fraction of suppressed pairs with Fourier and Sylvester matrices. (a) Fraction of suppressed pairs with Sylvester matrices, that is, Ω(US )/ m n 2 , plotted against n for various values of m. The data is obtained randomly sampling N = 1000 input-output pairs, and computing the permanent corresponding to each scattering amplitude. The percentage is known to be zero for number of photons of the form n = 2p − 1 [81], regardless of m. (b) Fraction of suppressed pairs with Fourier matrices, that is, Ω(UF )/ m n 2 , plotted against n for various values of m. The fraction is in this case significantly lower than for Sylvester matrices. For most values of n, no suppressed pair was found in the sampled dataset. In fig. 5.5 are again given the estimated fractions for Fourier and Sylvester matrices. The error bars in fig. 5.5 represent the 99.7 % confidence interval, calculated through Bayesian inference from the number of suppressed pairs found in the sampled datasets. In particular, in fig. 5.5-a the fraction Ω(US ) depends weakly on m, as opposite to the strong dependence of UF on m shown in fig. 5.5-b. Another relevant matter is the scaling with m and n of Ωeff(US ), that is, the number of suppressed pairs satisfying the conditions of theorem 3. No efficiently evaluable analytical expression is known for this number. Brute-force estimates are given in fig. 5.6, where calculations similar to those in fig. 5.5 are shown, but this time for the efficiently predictable suppressed pairs. In the presented cases of m = 16, 32, 64, 128, Ωeff(US ) seems to decrease exponentially in n, for n < m/2. In conclusion, while theorem 3 represents a significant improvement with respect to theorem 1 and theorem 2, this fraction is probably still not large enough for an efficient validation test with Sylvester matrices. However the data on the total fraction of suppressed states presented in figs. 5.4 and 5.5 leaves open the possibility that, further 86
  • 87.
    improving on theorem3, an efficient validation scheme for Sylvester matrices may be devised. m =32 m =64 m =128 m =256 Sylvester 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 n 10 20 30 40 50 % m =32 m =64 m =128 m =256 Fourier 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 n 2 4 6 8 %a) b) Figure 5.5: Fraction of suppressed pairs with Fourier and Sylvester matrices. Each color repre- sents the fraction corresponding to a different number of modes m, plotted as a function of the number of photons n. The error bars approximately represent the 99.7% confidence interval, estimated from the sampled data through Bayesian inference. The approximation used breaks for very small number of found suppressions, which is seen in the points where the error bars span negative values. 2 4 6 8 10 12 14 n 10 20 50 % m=16 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 n 0.10 0.50 1 5 10 50 % m=32 2 10 18 26 34 42 50 58 n 10-4 0.001 0.010 0.100 1 10 100 % m=64 18 34 50 66 82 98 114 n 0.5 1 5 10 50 % m=128 Figure 5.6: Estimated fraction of predicted suppressed pairs with Sylvester matrices. Plot of the percentage of predicted suppressed input-output pairs against n, for various values of m, for Sylvester matrices. The data is calculated randomly sampling input-output collision-free pairs, and verifying whether each is suppressed according to theorem 3. The missing data corresponds to values of n at which no suppression in the sampled pairs was found. For m = 64, the size of the sample for which no suppressions were found is N = 16 × 103 , while for m = 128 is N = 2000. Larger samples were not used due to the heavy computational times required. 87
  • 88.
  • 89.
    Chapter 6 Experimental bosonsampling validation with integrated photonics In this chapter we report the experimental implementation of the protocol devised in [22, 24], and explained in section 4.2, and its use to validate quantum many-body interference against alternative non-trivial hypotheses [25]. To this end, a scalable approach for the implementation of the fast Fourier transform algorithm, using 3-D photonic integrated interferometers, is developed. We observe the suppression law for a large number of output states with 4- and 8-mode optical circuits. The experimental results demonstrate genuine quantum interference between the injected photons. This work provides a first experimental realization [25] of the validation protocol with Fourier matrices [22] discussed in section 4.2 (during this work another experimental realization of the same validation scheme was reported [17]). The reported experiments were realized in the Quantum Optics laboratories at University of Rome La Sapienza. In section 6.1 an overview of the work is given. In section 6.2 the femtosecond laser-writing technique, employed to build the integrated photonic chips used in the experiment, is described. In section 6.3 the quantum Fast Fourier Transform is described. This is a linear optics generalization of the classical Fast Fourier Transform algorithm, used in this work to implement efficiently the Fourier transformation on the integrated interferometer. In section 6.4 we give a brief description of the photon generation methods. In section 6.5 the results of the measurements are presented, focusing on the observation of the Hong-Ou-Mandel dips and peaks, and the reconstruction of the chip transformation. In section 6.6 we describe how the experimental results validate the genuine many-boson quantum interference occuring in the interferometer, as opposite to alternative models of evolution. Finally, in section 6.7, the results of the experiments are discussed and the possible future prospects of this line of research. 89
  • 90.
    6.1 Overview ofthe experiment We here report the experimental observation of the recently theoretically proposed [22] suppression law for Fourier matrices, and its use to validate quantum many-body inter- ference against alternative non-trivial hypotheses resulting in similar output probability distributions. The Fourier matrices have been implemented with an efficient and reliable approach by exploiting the quantum version of the Fast Fourier Transform (qFFT), an algorithm developed by Barak and Ben-Aryeh [26] to optimize the number of optical elements required to build the Fourier transform over the optical modes. Here we implement the qFFT on photonic integrated interferometers by exploiting the 3-D capabilities of femtosecond laser writing [82, 83], which makes it possible to fab- ricate waveguides arranged in three-dimensional structures with arbitrary layouts [13, 84, 85], by adopting an architecture scalable to a larger number of modes. The observations have been carried out with two-photon Focks states injected into 4- and 8-mode qFFT interferometers. The peculiar behaviour of Fock states compared to other kinds of states is investigated, showing in principle the validity of the certification protocol for the identification of true granular n-particle interference, which is the source of a rich landscape of quantum effects such as the computational complexity of BosonSampling. 6.2 Femtosecond laser writing Building large interferometers out of discrete, bulk optical elements tends to result in mechanical instabilities which have prevented the demonstration of even a symmetric, 3-mode interferometer that preserves quantum coherence. A more promising approach to obtain stable multi-mode interferometers involves the fabrication of this network of linear optical elements by integrated optical waveguides in a glass chip [13, 84]. Waveguides are fabricated using the femtosecond laser micromachining technique [83, 86], which exploits nonlinear absorption of focused femtosecond pulses to induce a permanent and localized refractive index increase in transparent materials, as schemat- ically shown in fig. 6.1. Arbitrary three-dimensional circuits can be directly written by translating the sample along the desired path, keeping the velocity constant with respect to the laser beam. This technique allows fast and cost-effective prototyping of new devices, enabling the implementation of three-dimensional layouts that are impossible to realize with conventional lithography. 90
  • 91.
    Figure 6.1: Laser-writinga waveguide on a glass chip. The waveguide is fabricated focusing femtosecond laser pulses to induce a permanent and localized refractive index increase in transpar- ent materials. In the integrated optics approach the role of beam splitters is performed by direc- tional couplers, devices which bring two waveguides close together to redistribute the light propagating in them by evanescent field coupling (fig. 6.2-c). The phase shifters are instead implemented by deforming the S-bent waveguides at the input of each directional coupler in order to stretch the optical path (fig. 6.2-b). The integrated optics analogue of a discrete components layout, depicted in fig. 6.3-a, is shown in fig. 6.3-b, where one can appreciate the one-to-one correspondence between elements in the two approaches. TRANSMISSION CONTROL PHASE-SHIFT CONTROL 0.00 0.05 0.10 0.15 0.20 0.25 0 1 2 3 4 5 d Phaseshift[rad] a b c 0 1 2 3 4 5 6 0.0 0.2 0.4 0.6 0.8 1.0 Angle [°] Transmissivity d [μm] Transmission Figure 6.2: Independent control of the phase shift and transmissivity at each directional cou- pler. (a) The controlled deformation of the S-bent waveguide at the input of each directional coupler and coupling geometry allows independent control over the phase shift and transmissivity. (b) The deformation of the S-bent waveguide section is function of a deformation coefficient d. The graph showed the undeformed S-bend together with a deformed one. (c) Control over the transmissivity of the directional coupler is performed by modulating the coupling coefficient. This is achieved by changing the waveguide spacing in the coupling region by rotating one arm of the directional coupler out of the main circuit plane. A sample dependence of the the transmissivity as a function of the deformation angle is provided. Credits: [10]. 91
  • 92.
    t1 t2 t3t4 t5 t6 t7 t8 t9 t10 IN 1 2 3 4 5 OUT 1 2 3 4 5 t1 t5 t6 t7 t2 t3 t4 t8 t9 t10 IN 1 2 3 4 5 5 4 3 2 1 OUT a b tiαi βi tiαi βi Figure 6.3: Layout of multimode interferometers. (a) Realization of an arbitrary 5×5 mode trans- formation via a network of beam splitters with different transmissivities ti. The blue and red boxes stand for different phase shifters. (b) Implementation of the same scheme adopting integrated photonics. αi and βi are parameters characterizing the phase shifts acquired by the photons passing through S-bend waveguide segments. Credits: [10]. 6.3 Realization of the quantum Fast Fourier Trans- form We here introduce an experimental implementation of the Fourier transformation with the qFFT approach. The general method to realize an arbitrary unitary transformation using linear optics was introduced by Reck et al. [79], who provided a decomposition of an m×m unitary as a sequence of m(m − 1)/2 beam splitters and phase shifters. However, in the special case of Fourier matrices, a more efficient method has been proposed [26, 87], which takes advantage of their symmetries to significantly reduce the number of linear optical elements required. Based on the classical algorithm of Cooley and Tukey [27], who first introduced the Fast Fourier Transform algorithm as a more efficient way to compute the discrete 92
  • 93.
    Fourier transform, Barakand Ben-Aryeh [26] developed a quantum analogue in the linear optics domain, leading to the concept of qFFT. This approach, valid for 2p - dimensional Fourier matrices, requires only (m/2) log2 m beam splitters and phase shifters, to be compared with the O(m2 ) elements needed for the more general Reck decomposition, thus enhancing the compactness and scalability of the platform for a more reliable experimental realization. The sequential structure arising from the decomposition of the m-dimensional Fourier matrix using the Barak and Ben-Aryeh algorithm is reproduced by the con- secutive layers shown in fig. 6.4. The complex arrangement of pairwise interactions necessary for the qFFT method cannot be easily implemented using a planar architecture. However, femtosecond laser writing technique allows to overcome this issue exploiting the third dimension, arranging the waveguides along the bidimensional sections of the integrated chip. phase phasedirectional coupler directional coupler 4 3 2 1 4 3 2 1 31 6 42758 L1 L2 L1 L2 L3 P1 P2 P1 3 1 4 2 7 5 8 6 a b L1 L2 42 31 P1 L1 L2P1 L3P2 ⇡/23: 7: 8: 6: 8: 4: ⇡/2 ⇡/2 ⇡/2 ⇡/4 3⇡/442 31 73 51 84 62 73 51 84 62 73 51 84 62 Figure 6.4: Schematic of the structure of the integrated devices. Internal structure of the 4-mode (a) and 8-mode (b) integrated interferometers implementing the qFFT over the optical modes. The mode arrangement has been chosen in a way to minimize bending losses. The insets shows the actual disposition of the waveguides in the cross-section of the devices. The modes coupled together in each step (Li) of the interferometer are joined by segments. The implemented phase shifts in each step (Pi) are also indicated. Credits: [25]. The strategy can be outlined as follows: the 2p modes are ideally placed on the vertices of a p-dimensional hypercube, and in each step of the algorithm the vertices connected by parallel edges having one specific direction are made to interact by a 2-mode Hadamard transformation, with proper phase terms. An optical interferometer implementing this procedure is thus composed of log2 m = p sections, each employing m/2 balanced beam splitters and phase shifters. The waveguide interferometers were fabricated realizing the Fourier matrix for m = 4 and m = 8 modes in borosilicate glass chips using femtosecond laser micromachining [83, 84]. A schematic representation of these two interferometers is given in fig. 6.4. According to the scheme outlined above and by exploiting the three-dimensional 93
  • 94.
    capabilities of thefabrication technique, the waveguides are placed, for what concerns the cross section of the device, on the vertices of a 2-D projection of the p-dimensional hypercube. Three-dimensional directional couplers, with proper interaction length and distance to achieve a balanced splitting, connect in each step the required vertices. The insets of fig. 6.4 show, at each step i, which modes are connected by directional couplers (Li) and the amount of phase shift that needs to be introduced in specific modes (Pi). Phase shifters, where needed, are implemented by geometrical deformation of the connecting S-bends. Fan-in and fan-out sections at the input and output of the device allows interfacing with 127-µm spaced single-mode fiber arrays. We note that in our device geometry, in each step, the vertices to be connected are all at the same relative distance. This means that, unless geometric deformations are designed where needed, light travelling in different modes does not acquire undesired phase delays. The geometric construction here developed is scalable to an arbitrary number of modes with a number of elements increasing as m log2 m. 6.4 Photon generation and manipulation The generation of two-photon states is performed by pumping a 2-mm long BBO crystal with the second harmonic of a 785 nm wavelength Ti:Sa pulsed laser, with an average power of 750 mW, which generates photons at 785 nm with a type II parametric downconversion process. The two photons are spectrally filtered by means of interferential filters with 3 nm full width at half maximum, and coupled into single mode fibers. The indistinguishability of the photons is then ensured by a polarization compensation stage, and by propagation through independent delay lines before the injection within the interferometer via a single mode fiber array. After the evolution through the integrated devices, the photons are collected via a multimode fiber array. The detection system consists of 4 (8) single photon avalanche photodiodes used for the 4-(8-)mode chip. An electronic data acquisition system allowed us to detect coincidences between all pairs of output modes. Typical coincidence rates for each collision-free output combination with distinguishable photons amounted to ∼70-80 Hz for the 4-mode chip, and ∼10-20 Hz for the 8-mode chip. 6.5 One- and two-photon measurements in integrated Fourier interferometers The two implemented interferometers of m = 4 and m = 8 modes are fed with one- and two-photon states. The experimental setup, preparing a biphoton wavepacket to be injected into the devices, is shown in fig. 6.5. 94
  • 95.
    a b IF HWP PBS PC PDC outputFA input FA chip PBS HWP IF PC DL DL SMF SMF input FA Figure 6.5: Experimental apparatus for input state preparation. (a) The photon source (IF: inter- ferential filter, HWP: half-wave plate, PBS: polarizing beam splitter, PC: polarization compensator, PDC: parametric downconversion, DL: delay lines with motorized stages, SMF: single-mode fiber). (b) Photon injection (extraction) before (after) the evolution through the interferometer (FA: fiber array). Credits: [25]. To test the validity of the suppression law, we measured the number of coincidences at each forbidden output combination injecting cyclic inputs with two indistinguishable photons. The degree of violation D of the suppression law could simply be evaluated with a counting experiment. Alternatively, the same quantity D can be expressed as a function of single-photon input-output probabilities and of the HOM visibilities, defined as Vi,j = ND i,j − NQ i,j ND i,j , (6.1) where ND i,j is the number of detected coincidences for distinguishable photons, NQ i,j that for indistinguishable photons, and the subscripts (i, j) are the indices of the two output modes, for a given input state. The degree of violation can therefore be expressed as: D = Nforbidden Nevents = Pforbidden = (i,j)forbidden PQ i,j = (i,j)forbidden PD i,j(1 − Vi,j), (6.2) where PQ i,j (PD i,j) are the probabilities of having photons in the outputs i, j in the case of indistinguishable (distinguishable) particles. Here PD i,j can be obtained from single-particle probabilities. The visibilities are measured by recording the number of coincidences for each output combination as a function of the temporal delay between the two injected photons. For the 4-mode device, we measured the full set of 4 2 = 36 collision-free input- output combinations, that is, those combinations corresponding to the two photons exiting from two different ports. These contributions have been measured by recording 95
  • 96.
    the number ofcoincidences for each combination of two outputs as a function of the temporal delay between the two input photons. Due to the suppression law given in theorem 1, we expect to observe four suppressed outcomes (over six possible output combinations) for the two cyclic input states (1,3) and (2,4). Since distinguishable photons exhibit no interference, HOM dips in the coincidence patterns are expected for the suppressed output states. Conversely, peaks are expected in the non-suppressed output combinations. The experimental results are shown in fig. 6.6, where the expected pattern of four suppressions and two enhancements is reported, with average visibilities of ¯Vsupp = 0.899 ± 0.001 and ¯Venh = −0.951 ± 0.004 for suppression and enhancement respectively. For the cyclic inputs, we also measured the interference patterns for the output contributions where the two photons exit from the same mode. These terms have been measured by inserting an additional symmetric beam splitter on each output mode, and by connecting each of its two outputs to a single-photon detector. These cases correspond to a full-bunching scenario with n = 2, and a HOM peak with V = −1 visibility is expected independently from the input state and from the unitary operation [71]. This feature has been observed for the tested inputs, where an average visibility of ¯Vbunch = −0.969 ± 0.024 has been obtained over all full bunching combinations. The injection of non-cyclic input states has been employed for the complete recon- struction of the chip action ˜UqFFT 4 , using a data set statistically independent from the one adopted to observe the suppression law. The adopted reconstruction algorithm, which exploits knowledge on the internal structure of the interfometers shown in fig. 6.4, works in two steps. In a first step, the power-splitting ratios measured with classical light are employed to extrapolate the transmissivities of the directional cou- plers. In a second step, the two-photon visibilities for the non-cyclic inputs are used to retrieve the values of the fabrication phases. In both steps, the parameters are obtained by minimizing a suitable χ2 function. The results are shown in fig. 6.7. The fidelity between the reconstructed unitary ˜UqFFT 4 and the theoretical Fourier transform UF 4 is F = 0.9822 ± 0.0001, thus confirming the high quality of the fabrication process. The error in the estimation of the fidelity is obtained through a Monte Carlo simulation, properly accounting for the degree of distinguishability of the photons with a rescaling factor in the visibilities. 96
  • 97.
    - 0.642(8) 2500 0.890(2) 2900 0.682(3) 2490 0.525(3) 2860 0.882(2) 2290 - 0.662(7) 2600 0.871(2) 2860 -0.958(8) 2870 0.871(2) 2920 0.878(1) 3090 - 0.961(9) 2640 0.877(1) 3620 0.603(4) 1640 0.898(1) 4320 - 0.692(11) 1330 - 0.518(9) 1570 0.897(1) 2940 0.619(4) 1630 0.622(3) 3460 0.869(2) 3090 - 0.709(7) 3120 - 0.547(7) 2970 0.898(1) 3160 0.596(3) 2270 0.916(1) 3940 - 0.953(7) 3410 0.917(1) 3670 0.911(1) 3310 - 0.931(8) 2660 0.916(2) 2450 - 0.606(8) 2150 0.898(1) 4380 0.660(3) 3210 0.536(3) 3590 0.869(2) 3230 - 0.628(7) 2990 in (1,2) in (1,3) in (1,4) in (2,3) in (2,4) in (3,4) out (1,2) out (1,3) out (1,4) out (2,3) out (2,4) out (3,4) Figure 6.6: Suppression law in a 4-mode Fourier integrated chip. Complete set of 36 measured coincidence patterns (raw experimental data) for all input-output combinations in the 4-mode chip. For each input-output combination, the measured coincidence pattern as a function of the time delay is shown (points: experimental data, lines: best-fit curves). Cyclic inputs, highlighted in the picture, exhibit enhancement (green) and suppression (red) on cyclic and non-cyclic outputs, respectively. For all points, error bars are due to Poissonian statistics of the events. For each visibility, the error is obtained through a Monte Carlo simulation by averaging over 3000 simulated data sets. In each plot the zero level coincides with the baseline, while a dashed line represents the number of coincidence events in the distinguishable limit. Credits: [25]. 97
  • 98.
    VrecVqF F TVF 1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4 OUT IN 0.4 0.45 4- 1/2 0.55 0.6 a) b) Figure 6.7: Suppression law in a 8-mode qFFT integrated chip. (a) HOM visibilities for all 36 input- output configurations. From left to right: experimental measured visibilities (VqFFT, obtained from raw experimental data), visibilities calculated from the reconstructed unitary (Vrec), and visibilities calculated from the theoretical unitary (VF). (b) Representation of the reconstructed experimental transformation ˜UqFFT 4 , and comparison with UF 4. Colored disks represent the moduli of the reconstructed matrix elements (all equal to 4−1/2 for UF 4). Arrows represent the phases of the unitary matrix elements (green: reconstructed unitary, blue: Fourier matrix). Credits: [25]. 98
  • 99.
    For the 8-modechip we recorded all the 8 2 = 28 two-photon coincidence patterns, as a function of the relative delay between the input photons, for each of the 4 collision- free cyclic inputs and for one non-cyclic input. The results are shown in fig. 6.8. The reconstruction of the actual unitary transformation ˜UqFFT 8 implemented has been performed with the same algorithm of the 4-modes, by using the power-splitting ratios measured with classical light and the two-photon visibilities for one non-cyclic input. The latter has been chosen in a way to maximize the sensitivity of the measurements with respect to the five fabrication phases. The results are show in fig. 6.9. The fidelity between the reconstructed unitary ˜UqFFT 8 and the ideal 8-mode Fourier transform UF 8 is F = 0.9527 ± 0.0006. 0.800(6) 480 0.977(23) 480 0.844(6) 400 1.019(27) 350 0.772(9) 240 0.971(26) 360 0.850(6) 290 0.813(4) 980 0.856(20) 490 0.755(7) 440 0.895(24) 340 0.755(8) 270 0.838(19) 520 0.850(4) 690 0.890(25) 310 0.800(6) 480 0.960(24) 370 0.822(5) 590 0.790(7) 360 0.980(28) 260 0.824(5) 520 0.959(25) 300 0.796(7) 300 0.938(27) 260 0.765(9) 280 0.795(9) 250 0.875(27) 280 0.847(7) 380 out 1 out 2 out 3 out 4 out 5 out 6 out 7 out 2 out 3 out 4 out 5 out 6 out 7 out 8 Figure 6.8: Suppression law in a 8-mode qFFT integrated chip. Set of 28 measured coincidence patterns (raw experimental data), corresponding to all collision-free output combinations for the input (2,6) of the 8-mode interferometer. For each output combination, the measured coincidence pattern as a function of the time delay is shown (points: experimental data, lines: best-fit curves). Red or green backgrounds correspond to dips and peaks, respectively. For all points, error bars are due to the Poissonian statistics of the events. For each visibility, the error is obtained through a Monte Carlo simulation by averaging over 3000 simulated data sets. In each plot the zero level coincides with the baseline, while a dashed line represents the number of coincidence events in the distinguishable limit. Credits: [25]. 99
  • 100.
    Average visibilities (1,5) (2,6)(3,7) (4,8) (1,5) (2,6) (3,7) (4,8) - 0.25 - 0.5 - 0.75 - 1 0.25 0.5 0.75 1 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 OUT IN 0.3 0.33 8- 1/2 0.38 0.4a) b) Figure 6.9: Suppression law in a 8-mode qFFT integrated chip. (a) Average visibilities of dips (red bars) and peaks (green bars) observed for the four collision-free cyclic inputs ((1,5),(2,6),(3,7),(4,8)). Darker regions correspond to error bars of ±1 standard deviation. (b) Representation of the reconstructed experimental transformation ˜UqFFT 8 , and comparison with UF 8. Colored disks represent the moduli of the reconstructed matrix elements (all equal to 8−1/2 for UF 8). Arrows represent the phases of the unitary matrix elements (green: reconstructed unitary, blue: Fourier matrix). Credits: [25]. 6.6 Observation of the suppression law The suppression of events which do not satisfy theorem 1 is fulfilled only when two perfectly indistinguishable photons are injected in a cyclic input of a perfect Fourier interferometer. In such a case, we would have the suppression of all output states such that the sum of the elements of the corresponding MAL is odd. For the 4-mode (8-mode) device, this corresponds to 4 (16) suppressed and 2 (12) non-suppressed collision-free outputs (each one given by two possible arrangements of the two distinguishable photons), plus 4 (8) terms with two photons in the same output, each one corresponding to a single possible two-photon path. The expected violation for distinguishable particles can be obtained from classical considerations. Let us consider the case with n = 2. The two distinguishable photons evolve independently from each other, and the output distribution is obtained by classically mixing single-particle probabilities. All collision-free terms are equally likely to occur with probability q = 2/m2 , while full-bunching events occur with probability q = q/2 = 1/m2 . The degree of violation DD can then be obtained by multiplying the probability q by the number of forbidden output combinations. As a result, we expect a violation degree of DD = 0.5 for distinguishable two-photon states. The evaluation of the expected value for a mean field state, which is due to single particle bosonic statistic effects, require different calculations [22]. It can be shown that for n = 2 the degree of violation is DMF = 0.25. For each of the cyclic inputs, we have evaluated here the violation degree Dobs 100
  • 101.
    resulting from collecteddata. By measuring the coincidence pattern as a function of the path difference ∆x = c∆τ between the two photons, and thus by tuning their degree of distinguishability, we could address the transition from distinguishable to indistinguishable particles. The value of Dobs as a function of ∆x has been obtained as (i,j)forbidden PD i,j(N∆x i,j /ND i,j), where N∆x i,j and ND i,j are the number of measured coin- cidences for a given value of ∆x and for distinguishable particles respectively. Two different regions can be identified. For intermediate values of ∆x with respect to the coherence length of the photons, the measured data fall below the threshold DD, and hence the hypothesis of distinguishable particles can be ruled out. Then, for smaller values of the path difference up to ∆x → 0, true two-photon interference can be certified since both hypothesis of distinguishable particles and mean field state can be ruled out. The maximum violation occurring at ∆x = 0 delay can be evaluated using eq. (6.2). The experimental results retrieved from the protocol are shown in the table of figs. 6.10 and 6.11, in which we compare the values Dobs(0) with the expected values for distinguishable particles DD and for a Mean Field state DMF. As shown for our implementation, the robustness of the protocol is ensured by the high number of standard deviations separating the values in each comparison, thus unambiguously confirming the success of the certification protocol. In conclusion, the alternative hypotheses of distinguishable particles and of a mean field state can be ruled out for all experiments. 101
  • 102.
    Figure 6.10: Observedviolations Dobs as a function of the path difference |∆x| = c|∆τ| between the two photons, for the 4-mode interferometer. Blue shaded regions in the plots correspond to the cases where the hypothesis of distinguishable particles can be ruled out, and true two-particle interference is present. Blue points: input (1,3). Red points: input (2,4). Blue solid line: theoretical prediction for input (1,3). Red solid line: theoretical prediction for input (2,4). Black dashed line: theoretical prediction for a Fourier matrix. (b) Data for the 8-mode interferometer. Blue points: input (1,5). Red points: input (2,6). Green points: input (3,7). Magenta points: input (4,8). Colored solid lines: corresponding theoretical predictions for different inputs. Black dashed line: theoretical prediction for a Fourier matrix. Table: violations Dobs(0) at ∆x = 0 and discrepancies (in sigmas) with the expected values for distinguishable particles (DD) and Mean Field Sampler (DMF), for the cyclic inputs of the two interferometer. Dobs(0) are calculated following eq. (6.2), while expected values for the other two cases are DD = 0.5 and DMF = 0.25. Error bars in all experimental quantities are due to the Poissonian statistics of measured events. All theoretical predictions in solid lines are calculated from the reconstructed unitaries, obtained from different sets of experimental data to ensure statistical independence. Credits: [25]. 102
  • 103.
    Figure 6.11: Observedviolations Dobs as a function of the path difference |∆x| = c|∆τ| between the two photons, for the 8-mode interferometer. Blue shaded regions in the plots correspond to the cases where the hypothesis of distinguishable particles can be ruled out, and true two-particle interference is present. Blue points: input (1,5). Red points: input (2,6). Green points: input (3,7). Magenta points: input (4,8). Colored solid lines: corresponding theoretical predictions for different inputs. Black dashed line: theoretical prediction for a Fourier matrix. Table: violations Dobs(0) at ∆x = 0 and discrepancies (in sigmas) with the expected values for distinguishable particles (DD) and Mean Field Sampler (DMF), for the cyclic inputs of the two interferometer. Dobs(0) are calculated following eq. (6.2), while expected values for the other two cases are DD = 0.5 and DMF = 0.25. Error bars in all experimental quantities are due to the Poissonian statistics of measured events. All theoretical predictions in solid lines are calculated from the reconstructed unitaries, obtained from different sets of experimental data to ensure statistical independence. Credits: [25]. 103
  • 104.
    6.7 Discussion We havereported on the experimental observation of the suppression law on specific output combinations of a Fourier transformation due to quantum interference between photons. The observation of the suppression effect allowed us to rule out alternative hypotheses to the Fock state. The use of a novel implementation architecture, enabled by the 3-D capabilities of femtosecond laser micromachining, extends the scalability of this technique to larger systems with lower experimental effort with respect to other techniques. While the presented architecture is designed to implement a Fourier matrix for a number of modes equal to m = 2p , a generalization of the approach can be obtained by adopting a building block different from the beam splitter. For devices of odd dimension, for instance, such a tool can be provided by the tritter transformation [13]. At the same time, the universality of a generalized HOM effect with an arbitrary number of particles and modes is expected to make it a pivotal tool in the diagnostic and certification of quantum photonic platforms. BosonSampling represents a key example, since the scalability of the technique is expected to allow efficient certification of devices outperforming their classical counterparts. An interesting open problem is whether the computational hardness of BosonSampling is mantained if the randomness condition is relaxed, and thus which is the minimal interferometric architecture required for an evidence of post-classical computation. Fourier matrices can find application in different contexts. For instance, multiport beam splitters described by the Fourier matrix can be employed as building blocks for multiarm interferometers, which can be adopted for quantum-enhanced single and multiphase estimation protocols [88]. This would also allow the measurement of phase gradients with precision lower than the shot-noise limit [89]. Other fields where Fourier matrices are relevant include quantum communication scenarios [90], observation of two-photon correlations as a function of geometric phase [91], fundamental quantum information theory including mutually unbiased bases [92], as well as entanglement generation [93]. 104
  • 105.
    Conclusions In this workwe presented both theoretical and experimental advances in the context of validating boson sampling experiments, both with the originally proposed approach [8] and the recently proposed scattershot boson sampling scheme [19, 20]. From the theoretical point of view, the existing validation techniques have been analysed in light of the recently proposed, and experimentally implemented, scattershot boson sampling scheme [18, 19, 20]. We found that the existing methods to validate genuine many-boson interference cannot be efficiently applied to scattershot boson sampling experiments. Subsequently, we devised a new Zero-Transmission Law (ZTL) for Sylvester matrices which has dramatically higher predictive capabilities with respect to previously proposed ZTLs [23, 24]. We studied the fraction of suppressed input-output pairs predicted by this ZTL, and its scaling with the number of photons n and the number of modes m. We found that the number of predicted suppressed input-output pairs, while significantly higher than that presented by previous ZTLs, still scales badly with n, decreasing exponentially. However, it is less clear how the total number of suppressed pairs in Sylvester matrices actually scales. Brute-force computations are not possible beyond relatively small numbers of n, but the data gathered so far shows a rather complex behaviour, which leaves open the possibility that this ratio decreases only polynomially with n and m. Further work is needed in this direction to conclusively assess whether an efficient validation scheme based on Sylvester matrices is possible, but what is currently known leaves open this possibility. From the experimental point of view, we reported on the first implementation of a validation scheme [22] based on a ZTL for Fourier matrices [24]. This was realized using integrated photonics technology, with 2 photons injected into 4- and 8-mode in- terferometers realized on glass chips with femtosecond laser writing technique. A novel methodology to efficiently implement a 2p -dimensional Fourier matrix was employed, exploiting the 3-D capabilities of femtosecond laser writing to implement the Barak and Ben-Aryeh [26] quantum optics generalization of the Fast Fourier Transform algorithm [27]. A thorough analysis of the Hong-Ou-Mandel dips and peaks corresponding to the various possible configurations of 2-photon input-output pairs is reported, and the transition from distinguishable to indistinguishable photons studied, showing an excellent agreement with the predictions of quantum mechanics. An algorithm to effi- 105
  • 106.
    ciently reconstruct theunitary matrix implemented in the device is presented, showing a good agreement with the ideal Fourier matrix. Finally, the efficient validation scheme proposed in [22] is implemented, finding a perfect match with the expected pattern of suppressions, and ruling out alternative models such as distinguishable particles and Mean Field sampler [22]. 106
  • 107.
    Aknowledgments We aknowledge FabioSciarrino and Alessandro Cianchi, who supervised the presented work, for valuable teachings and useful discussions. We also wish to aknowledge Nicolò Spagnolo, Marco Bentivegna, Fulvio Flamini, and Niko Viggianiello, who closely followed every step of this work, for the many tips and discussions. 107
  • 108.
  • 109.
    Bibliography [1] Richard PFeynman. “Simulating physics with computers”. In: International jour- nal of theoretical physics 21.6/7 (1982), pp. 467–488. [2] John Preskill. Quantum computing and the entanglement frontier. 2012. arXiv: 1203.5813 [quant-ph]. [3] John Preskill. Supremacy Now? 2012. url: http://quantumfrontiers. com/2012/07/22/supremacy-now/ (visited on 10/27/2015). [4] John Preskill. Fault-tolerant quantum computation. 1998. arXiv: quant-ph/ 9712048. [5] Emanuel Knill, Raymond Laflamme, and Wojciech H Zurek. “Resilient quantum computation”. In: Science 279.5349 (1998), pp. 342–345. [6] Dorit Aharonov and Michael Ben-Or. “Fault-tolerant quantum computation with constant error rate”. In: SIAM Journal on Computing 38.4 (2008), pp. 1207–1282. arXiv: quant-ph/9906129. [7] Emanuel Knill, Raymond Laflamme, and Gerald J Milburn. “A scheme for efficient quantum computation with linear optics”. In: nature 409.6816 (2001), pp. 46–52. [8] Scott Aaronson and Alex Arkhipov. “The computational complexity of linear optics”. In: Proceedings of the forty-third annual ACM symposium on Theory of computing. ACM. 2011, pp. 333–342. arXiv: 1011.3245 [quant-ph]. [9] Matthew A Broome et al. “Photonic boson sampling in a tunable circuit”. In: Science 339.6121 (2013), pp. 794–798. arXiv: 1212.2234 [quant-ph]. [10] Andrea Crespi et al. “Integrated multimode interferometers with arbitrary designs for photonic boson sampling”. In: Nature Photonics 7.7 (2013), pp. 545–549. arXiv: 1212.2783 [quant-ph]. [11] Max Tillmann et al. “Experimental boson sampling”. In: Nature Photonics 7.7 (2013), pp. 540–544. arXiv: 1212.2240 [quant-ph]. [12] Justin B Spring et al. “Boson sampling on a photonic chip”. In: Science 339.6121 (2013), pp. 798–801. arXiv: 1212.2622 [quant-ph]. 109
  • 110.
    [13] Nicolò Spagnoloet al. “Three-photon bosonic coalescence in an integrated tritter”. In: Nature communications 4 (2013), p. 1606. arXiv: 1210.6935 [quant- ph]. [14] Nicolò Spagnolo et al. “Experimental validation of photonic boson sampling”. In: Nature Photonics 8.8 (2014), pp. 615–620. arXiv: 1311.1622 [quant-ph]. [15] Jacques Carolan et al. “On the experimental verification of quantum complexity in linear optics”. In: Nature Photonics 8.8 (2014), pp. 621–626. arXiv: 1311.2913 [quant-ph]. [16] Jacques Carolan et al. “Verifying quantum complexity in linear optical experi- ments”. In: CLEO: QELS_Fundamental Science. Optical Society of America. 2014, FM2A–7. [17] Jacques Carolan et al. “Universal linear optics”. In: Science 349.6249 (2015), pp. 711– 716. doi: 10.1126/science.aab3642. arXiv: 1311.2913 [quant- ph]. url: http://www.sciencemag.org/content/349/6249/ 711.abstract. [18] Marco Bentivegna et al. “Experimental scattershot boson sampling”. In: Science Advances 1.3 (2015), e1400255. arXiv: 1515.03708 [quant-ph]. [19] Scott Aaronson. Scattershot BosonSampling: A new approach to scalable Boson- Sampling experiments. 2013. url: http://www.scottaaronson.com/ blog/?p=1579 (visited on 12/22/2015). [20] AP Lund et al. “Boson sampling from a Gaussian state”. In: Physical review letters 113.10 (2014), p. 100502. [21] Scott Aaronson and Alex Arkhipov. “Bosonsampling is far from uniform”. In: Quantum Information & Computation 14.15-16 (2014), pp. 1383–1423. [22] Malte C Tichy et al. “Stringent and efficient assessment of boson-sampling de- vices”. In: Physical review letters 113.2 (2014), p. 020502. arXiv: 1312.3080 [quant-ph]. [23] Andrea Crespi. “Suppression laws for multiparticle interference in Sylvester interferometers”. In: Physical Review A 91.1 (2015), p. 013811. arXiv: 1502. 06372 [quant-ph]. [24] Malte Christopher Tichy et al. “Zero-transmission law for multiport beam split- ters”. In: Physical review letters 104.22 (2010), p. 220405. arXiv: 1002.5038 [quant-ph]. [25] Andrea Crespi et al. “Quantum suppression law in a 3-D photonic chip implement- ing the Fast Fourier Transform”. In: (2015). arXiv: 1508.00782 [quant- ph]. 110
  • 111.
    [26] Ronen Barakand Yacob Ben-Aryeh. “Quantum fast Fourier transform and quan- tum computation by linear optics”. In: JOSA B 24.2 (2007), pp. 231–240. [27] James W Cooley and John W Tukey. “An algorithm for the machine calculation of complex Fourier series”. In: Mathematics of computation 19.90 (1965), pp. 297–301. [28] Michael A Nielsen and Isaac L Chuang. Quantum computation and quantum information. Cambridge university press, 2010. [29] Robert Solovay and Volker Strassen. “A fast Monte-Carlo test for primality”. In: SIAM journal on Computing 6.1 (1977), pp. 84–85. [30] Peter W Shor. “Polynomial-time algorithms for prime factorization and discrete logarithms on a quantum computer”. In: SIAM journal on computing 26.5 (1997), pp. 1484–1509. arXiv: quant-ph/9508027. [31] Lane A Hemaspaandra. “Sigact news complexity theory column 36”. In: ACM SIGACT News 33.2 (2002), pp. 34–47. [32] Scott Aaronson. The Scientific Case for P != NP. 2014. url: http://www. scottaaronson.com/blog/?p=1720 (visited on 11/30/2015). [33] Scott Aaronson. “Guest column: NP-complete problems and physical reality”. In: ACM Sigact News 36.1 (2005), pp. 30–52. [34] Leslie G Valiant. “The complexity of computing the permanent”. In: Theoretical computer science 8.2 (1979), pp. 189–201. [35] Charles H Bennett and David P DiVincenzo. “Quantum information and compu- tation”. In: Nature 404.6775 (2000), pp. 247–255. [36] Ryszard Horodecki et al. “Quantum entanglement”. In: Reviews of Modern Physics 81.2 (2009), p. 865. [37] Otfried Gühne and Géza Tóth. “Entanglement detection”. In: Physics Reports 474.1 (2009), pp. 1–75. [38] Seth Lloyd. “Universal quantum simulators”. In: SCIENCE-NEW YORK THEN WASHINGTON- (1996), pp. 1073–1077. [39] Bryan T Gard et al. “An introduction to boson-sampling”. In: arXiv preprint arXiv:1406.6767 (2014). [40] Pieter Kok et al. “Linear optical quantum computing with photonic qubits”. In: Reviews of Modern Physics 79.1 (2007), p. 135. arXiv: quant-ph/0512071. [41] Henryk Minc. Permanents. Vol. 6. Cambridge University Press, 1984. [42] MD Eisaman et al. “Invited review article: Single-photon sources and detectors”. In: Review of Scientific Instruments 82.7 (2011), p. 071101. 111
  • 112.
    [43] Brahim Lounisand Michel Orrit. “Single-photon sources”. In: Reports on Progress in Physics 68.5 (2005), p. 1129. [44] Christopher Gerry and Peter Knight. Introductory quantum optics. Cambridge university press, 2005. [45] Romain Alléaume et al. “Experimental open-air quantum key distribution with a single-photon source”. In: New Journal of physics 6.1 (2004), p. 92. [46] T Gaebel et al. “Stable single-photon source in the near infrared”. In: New Journal of Physics 6.1 (2004), p. 98. [47] E Wu et al. “Room temperature triggered single-photon source in the near in- frared”. In: New Journal of Physics 9.12 (2007), p. 434. [48] Satoshi Kako et al. “A gallium nitride single-photon source operating at 200 K”. In: Nature materials 5.11 (2006), pp. 887–892. [49] Andrew J Shields. “Semiconductor quantum light sources”. In: Nature photonics 1.4 (2007), pp. 215–223. [50] Stefan Strauf et al. “High-frequency single-photon source with polarization control”. In: Nature photonics 1.12 (2007), pp. 704–708. [51] Markus Hennrich et al. “Photon statistics of a non-stationary periodically driven single-photon source”. In: New Journal of Physics 6.1 (2004), p. 86. [52] Christian Maurer et al. “A single-photon source based on a single Ca ion”. In: New journal of physics 6 (2004), pp. 94–94. [53] Mathias Steiner et al. “Highly efficient, tunable single photon source based on single molecules”. In: Applied physics letters 18 (2007). [54] Shuai Chen et al. “Deterministic and storable single-photon source based on a quantum memory”. In: Physical review letters 97.17 (2006), p. 173004. [55] S Cova et al. “Evolution and prospects for single-photon avalanche diodes and quenching circuits”. In: Journal of Modern Optics 51.9-10 (2004), pp. 1267–1288. [56] S Komiyama et al. “A single-photon detector in the far-infrared range”. In: Nature 403.6768 (2000), pp. 405–407. [57] Kristine M Rosfjord et al. “Nanowire single-photon detector with an integrated optical cavity and anti-reflection coating”. In: Optics Express 14.2 (2006), pp. 527– 534. [58] Marius A Albota and Franco NC Wong. “Efficient single-photon counting at 1.55 µm by means of frequency upconversion”. In: Optics letters 29.13 (2004), pp. 1449–1451. [59] Hiroki Takesue et al. “1.5-µm single photon counting using polarization-independent up-conversion detector”. In: Optics express 14.26 (2006), pp. 13067–13072. 112
  • 113.
    [60] Robert ThomasThew et al. “Low jitter up-conversion detectors for telecom wavelength GHz QKD”. In: New Journal of Physics 8.3 (2006), p. 32. [61] Kendall B Davis et al. “Bose-Einstein condensation in a gas of sodium atoms”. In: Physical review letters 75.22 (1995), p. 3969. [62] Mike H Anderson et al. “Observation of Bose-Einstein condensation in a dilute atomic vapor”. In: science 269.5221 (1995), pp. 198–201. [63] Jan Klaers et al. “Bose-Einstein condensation of photons in an optical microcavity”. In: Nature 468.7323 (2010), pp. 545–548. [64] CK Hong, ZY Ou, and Leonard Mandel. “Measurement of subpicosecond time intervals between two photons by interference”. In: Physical Review Letters 59.18 (1987), p. 2044. [65] Malte C Tichy. “Interference of identical particles from entanglement to boson- sampling”. In: Journal of Physics B: Atomic, Molecular and Optical Physics 47.10 (2014), p. 103001. arXiv: 1312.4266 [quant-ph]. [66] Malte Christopher Tichy. “Entanglement and interference of identical particles”. PhD thesis. Universitätsbibliothek Freiburg, 2011. [67] VS Shchesnovich. “Asymptotic evaluation of bosonic probability amplitudes in linear unitary networks in the case of large number of bosons”. In: International Journal of Quantum Information 11.05 (2013), p. 1350045. [68] David G Glynn. “The permanent of a square matrix”. In: European Journal of Combinatorics 31.7 (2010), pp. 1887–1891. [69] Lidror Troyansky and Naftali Tishby. “Permanent uncertainty: On the quantum evaluation of the determinant and the permanent of a matrix”. In: Proc. 4th Workshop on Physics and Computation (PhysComp’96). 1996. [70] Annalisa Marzuoli and Mario Rasetti. “Towards a quantum algorithm for the permanent”. In: International Journal of Quantum Information 5.01n02 (2007), pp. 223–228. [71] Nicolò Spagnolo et al. “General rules for bosonic bunching in multimode inter- ferometers”. In: Physical review letters 111.13 (2013), p. 130503. [72] VS Shchesnovich. “Sufficient condition for the mode mismatch of single photons for scalability of the boson-sampling computer”. In: Physical Review A 89.2 (2014), p. 022333. [73] Peter P Rohde et al. “Will boson-sampling ever disprove the Extended Church- Turing thesis?” In: arXiv preprint arXiv:1401.2199 (2014). [74] Peter P Rohde and Timothy C Ralph. “Error tolerance of the boson-sampling model for linear optics quantum computing”. In: Physical Review A 85.2 (2012), p. 022332. 113
  • 114.
    [75] Anthony ALeverrier and Raul Garcia-Patron Sanchez. “Analysis of circuit im- perfections in BosonSampling”. In: Quantum information & computation 15.5-6 (2015), pp. 489–512. [76] Alberto Peruzzo et al. “Quantum walks of correlated photons”. In: Science 329.5998 (2010), pp. 1500–1503. arXiv: 1006.4764 [quant-ph]. [77] C Gogolin et al. “Boson-sampling in the light of sample complexity”. In: arXiv (2013). arXiv: 1306.3995 [quant-ph]. [78] Anthony Laing and Jeremy L O’Brien. “Super-stable tomography of any linear optical device”. In: arXiv preprint arXiv:1208.2868 (2012). [79] Michael Reck et al. “Experimental realization of any discrete unitary operator”. In: Physical Review Letters 73.1 (1994), p. 58. [80] Alex Arkhipov and Greg Kuperberg. “The bosonic birthday paradox”. In: Geome- try & Topology Monographs 18 (2012), pp. 1–7. [81] Rodica Simion and Frank W Schmidt. “On (+ 1,- 1)-matrices with vanishing permanent”. In: Discrete Mathematics 46.1 (1983), pp. 107–108. [82] Roberto Osellame et al. “Femtosecond writing of active optical waveguides with astigmatically shaped beams”. In: JOSA B 20.7 (2003), pp. 1559–1567. [83] Rafael R Gattass and Eric Mazur. “Femtosecond laser micromachining in trans- parent materials”. In: Nature photonics 2.4 (2008), pp. 219–225. [84] Thomas Meany et al. “Non-classical interference in integrated 3D multiports”. In: Optics express 20.24 (2012), pp. 26895–26905. [85] Konstantinos Poulios et al. “Quantum walks of correlated photon pairs in two- dimensional waveguide arrays”. In: Physical review letters 112.14 (2014), p. 143604. [86] G Della Valle, R Osellame, and P Laporta. “Micromachining of photonic devices by femtosecond laser pulses”. In: Journal of Optics A: Pure and Applied Optics 11.1 (2009), p. 013001. [87] P Törmä, I Jex, and S Stenholm. “Beam splitter realizations of totally symmetric mode couplers”. In: journal of modern optics 43.2 (1996), pp. 245–251. [88] Nicolò Spagnolo et al. “Quantum interferometry with three-dimensional geome- try”. In: Scientific reports 2 (2012). [89] Keith R Motes et al. “Linear Optical Quantum Metrology with Single Photons: Exploiting Spontaneously Generated Entanglement to Beat the Shot-Noise Limit”. In: Physical review letters 114.17 (2015), p. 170802. [90] Saikat Guha. “Structured optical receivers to attain superadditive capacity and the Holevo limit”. In: Physical review letters 106.24 (2011), p. 240502. 114
  • 115.
    [91] Anthony Lainget al. “Observation of quantum interference as a function of Berry’s phase in a complex Hadamard optical network”. In: Physical review letters 108.26 (2012), p. 260505. [92] Ingemar Bengtsson et al. “Mutually unbiased bases and Hadamard matrices of order six”. In: Journal of mathematical physics 48.5 (2007), p. 052106. [93] Yuan Liang Lim and Almut Beige. “Multiphoton entanglement through a Bell- multiport beam splitter”. In: Physical Review A 71.6 (2005), p. 062311. 115