SlideShare a Scribd company logo
1 of 43
Download to read offline
Cosmic Microwave Background Radiation
Lecture 3 : Parameter Estimation
Jayanti Prasad
Inter-University Centre for Astronomy & Astrophysics (IUCAA)
Pune, India (411007)
Autumn School on Cosmology (5 - 15th Nov 2013)
BITS PILANI
1 / 35
CMB Data Analysis pipeline
In every step of CMB data analysis the aim is to reduce the volume
of data without losing information.
2 / 35
Plan of the Talk
Data Analysis Techniques
Inverse Problems
Chi-Square minimization
Maximum-Likelihood Estimation
3 / 35
Plan of the Talk
Data Analysis Techniques
Inverse Problems
Chi-Square minimization
Maximum-Likelihood Estimation
CMB data and Map making
3 / 35
Plan of the Talk
Data Analysis Techniques
Inverse Problems
Chi-Square minimization
Maximum-Likelihood Estimation
CMB data and Map making
Cosmological Parameter Estimation
CMB Likelihood
Markov Chain Monte Carlo Methods
3 / 35
Plan of the Talk
Data Analysis Techniques
Inverse Problems
Chi-Square minimization
Maximum-Likelihood Estimation
CMB data and Map making
Cosmological Parameter Estimation
CMB Likelihood
Markov Chain Monte Carlo Methods
WMAP and Planck
3 / 35
Plan of the Talk
Data Analysis Techniques
Inverse Problems
Chi-Square minimization
Maximum-Likelihood Estimation
CMB data and Map making
Cosmological Parameter Estimation
CMB Likelihood
Markov Chain Monte Carlo Methods
WMAP and Planck
Summary and Conclusions
3 / 35
Data Analysis Techniques
Scientific observations can be represented by a data vector d
which can be a time series {t} or temperature map of the sky
{T} or something else.
Data have information about some physical process for which
we have a theoretical model represented by a set of
parameters i.e., parameter vector Θ.
One of its example is a Gaussian process represented by two
parameters i.e., the mean µ and the variance σ2. The
probability of obtaining data d given a theoretical (Gaussian
model (µ, σ2)) is given by:
P(d|µ, σ2
) =
1
√
2πσ2
exp −
1
2
(d − µ)2
σ2
(1)
4 / 35
Data Analysis Techniques
Note that finding out which of the theoretical model is better
than others e.g., we should be fitting a parabola rather than a
line, is different from finding the parameters of a model.
In the present discussion we will not discuss model
comparison, we will always assume a model is true and will try
to find its parameters i.e., what are the values of the
parameters of ΛCDM cosmological model.
Values of parameters do not make sense unless we specify a
model.
Once we have a model (with its parameters) we can easily
create data by simulating that model and this is called a
forward problem.
Finding parameters Θ of a model from the data d is called the
inverse problem which is much harder to solve than the
forward problem.
5 / 35
Linear Problem
A problem is said to be a linear problem when the data d
depends on model parameters linearly, for example fitting a
polynomial is a linear problem:
di =
M
j=1
cj θj
for i = 1, N (2)
where ci are the model parameters.
By definition there is no linear relationship between the data
and model parameter for a non-linear problem, for example,
fitting a Gaussian is a non-linear problem.
Non-linear problems are harder to solve than linear problems.
Dependency of cosmological data i.e., CMB map, on
cosmological parameters is a non-linear problem.
6 / 35
Solvable problems
Not all the inverse problems are solvable and in some cases we
can easily find out why that is so.
On the basis of whether the size N of the data vector d is
larger, smaller or equal to the size M of the parameter vector
Θ, there are three possibilities.
N = M : Unique solution is possible
N > M : Over constrained problem, χ2
minimization, Unique
solution
N < M : Under-constrained problem, ill posed problem, priors,
regularization
Note that in the above consideration we have assumed that
all the data points are independent.
One of the common methods to solve an inverse is to
minimize a measure of misfit between the data and the
theoretical model.
7 / 35
Map Making in CMB
In CMB experiments like WMAP and Planck the time order
data or TOD d depends on the sky temperature T in the
following way:
Di = Aij Tj + Nj (3)
the index i and j are over time and pixels respectively.
In order to make a map from the TOD we have to estimate T
from D for which we can minimize the following function:
f (T) = (D − AT) (D − AT) (4)
which gives:
ˆT = (A A)−1
A D (5)
8 / 35
CMB Map making
In place of f (T) if we can also minimize χ2 :
χ2
(T) = (D − AT) C−1
N (D − AT) (6)
where CN =< NN > is the noise covariance matrix, then we
get:
ˆT = (A C−1
N A)−1
A C−1
N D (7)
which is called Maximum-Likelihood (ML) solution.
The covariance matrix of the maximum likelihood solution is
given by :
N = (A C−1
N A)−1
(8)
[Tegmark (1997)]
9 / 35
Chi-Square Distribution
If we have a variable y =
N
i=1 X2
i and X is drwan from Gaussian
random distribution then Y follows χ2
distribution of degree ν:
fχ2 (x) =
x
ν
2 −1
e− x
2
Γ(ν
2 )2
ν
2
(9)
10 / 35
Chi-Square Distribution
If we have a variable y =
N
i=1 X2
i and X is drwan from Gaussian
random distribution then Y follows χ2
distribution of degree ν:
fχ2 (x) =
x
ν
2 −1
e− x
2
Γ(ν
2 )2
ν
2
(9)
For large ν, χ2
distribution approaches Gaussian.
10 / 35
CMB anisotropies
CMB temperature anisotropies are expressed in terms of multipoles:
∆T(ˆn)
T
=
∞
l=0
m=l
m=−l
almYlm(ˆn) (10)
where
alm =
∆T(ˆn)
T
Ylm(ˆn)dˆn (11)
Where alm follow the Gaussian distribution with zero mean and
variance given by Cl :
< alm >= 0 (12)
and
alma∗
l m = δll δmm Cl (13)
An unbiased estimator of Cl is defined as:
ˆCl =
1
2l + 1
m=l
m=−l
alma∗
lm (14)
Note that the angular power spectrum Cl follows χ2
distribution.
11 / 35
Maximum Likelihood Estimation
The probability distribution P(Θ|d) (posterior) for model
parameters Θ given data d can be related to the probability
P(d|Θ) (likelihood) of an experiment giving data d for model
parameters Θ using the Bayes’ theorem:
P(Θ|d) =
P(d|Θ)P(Θ)
P(d)
(15)
where P(Θ) is called the prior and P(d) = P(d|Θ)P(Θ) is
used for the normalization purpose.
For the case of flat prior, posterior and likelihood are
proportional:
P(Θ|d) = P(d|Θ) = L(d|Θ) (16)
In Bayesian formalism we can easily incorporate new data in
analysis by considering the posterior of the old data as prior.
12 / 35
Bayesian Analysis
Note that when likelihood/posterior is not Gaussian then the
average value of the parameter < θ > may not coincide with the
value of θ0 at which the likelihood/posterior is maximum.
13 / 35
Errors in maximum Likelihood estimation
We are not only interested in finding the point Θ0 or < Θ >,
we are also also interested errors.
The spread of the likelihood function P(d|Θ) around the
maximum likelihood point Θ0 can be used to find the error
bars:
Expanding the likelihood function L = −2 log log P(d|Θ)
around the maximum likelihood point Θ0:
L(Θ) = L(Θ0)+(θi
−θi
0)
∂L(Θ)
∂θi
θi =θi
0
+
1
2
(θi
−θi
0)(θj
−θj
0)
∂2
L(Θ)
∂θi ∂θj
θi =θi
0
+....
(17)
Second derivative of the Likelihood function is called the
Hessian :
Hij =
∂2L(Θ)
∂θi ∂θj
(18)
14 / 35
There is an in-equality called the Cramer-Rao lower bound
which says that the error bars (variance) on any estimator
cannot be smaller than the inverse of information matrix I
Cij ≥
1
Iij
(19)
where I is defined as:
Iij =
∂2L(Θ)
∂θi ∂θj
(20)
How accurately a parameter can be estimated from the
likelihood depends upon how sensitive the likelihood function
is on that parameter, which is quantified by the Hessian
matrix.
15 / 35
Problems
1 Show that for a case when the noise is Gaussian, maximizing
likelihood is equivalent to minimization Chi-square.
2 Show that the maximum likelihood estimator is the minimum
variance estimator.
3 Show that finding the solution of the linear problem d = AΘ is
equivalent to finding of the maximum of the following function:
f (Θ) =
1
2
Θ AΘ − d Θ + c (21)
4 Show that the minimum chi-square solution of the model
y = c0 + c1x with data (xi , yi , σi ) is given by:
c0 =
S01S00 − S00S01
SS00 − S2
00
and c1 =
SS01 − S01S10
SS00 − S2
00
(22)
where S =
N
i=1 1/σ2
i and Sij =
N
k=1 xi
yj
/σ2
k . Also find the error.
16 / 35
Sometime likelihood is not much sensitive to individual
parameters (may not be Gaussian) but is highly sensitive to
some combinations of those.
For example, CMB Likelihood is almost Gaussian with respect
to the combination (Ωbh2, Ωch2, θ, A∗, tz) of cosmological
parameters [Chu et al. (2003)] where
A∗
=
A
76, 000
0.05Mpc−1
kpivot
1−ns
e−2τ
, (23)
and
t =
1
Ωbh2
2ns −1
(24)
17 / 35
Markov Chain Monte Carlo
Once we have the probability distribution P(Θ|d) for model
parameters Θ we can statistics of the parameters:
< Θ >= ΘdΘP(Θ|d) (25)
In practice, before carrying out the above integral we find out
the one dimensional probability distribution by marginalization
over other parameters:
P(θr ) = dθ1dθ2....dθr−1dθr+1...dθMP(θ1, θ2, ....sθM)
(26)
and
< θr >= θr dθr P(θr ) (27)
18 / 35
Carrying out multi-dimensional integration is very expansive
i.e., computational cost grows as O(nM) where n is the
number of grid points along one direction and M is the
dimensionality of the parameter space.
If we can replace the multi-dimensional integration by
summation over a finite number of points which represent the
probability distribution function then computational cost
becomes manageable.
< Θ >= ΘdΘP(Θ|d) =
1
N
N
i=1
Θi P(Θi |d) (28)
Markov-Chain Monte Carlo sampling samples the likelihood
function in such a way that there are more point in the region
where the likelihood function has the large values and less
where it has small values.
19 / 35
Markov-Chain Monte Carlo
When we toss a coin n times then the outcome of the nth toss
does not depend on the outcome of any of the previous
outcomes.
20 / 35
Markov-Chain Monte Carlo
When we toss a coin n times then the outcome of the nth toss
does not depend on the outcome of any of the previous
outcomes.
In a Markov-Chain the probability of a random variable Xn to
have value xn at step n depends on the probability of the
variable Xn−1 to have the value xn−1 at step n − 1.
P(XN) = P(Xn, Xn−1)P(Xn−1) (29)
where P(Xn, Xn−1) is called the transition probability,
transition kernel or proposal density.
20 / 35
Markov-Chain Monte Carlo
When we toss a coin n times then the outcome of the nth toss
does not depend on the outcome of any of the previous
outcomes.
In a Markov-Chain the probability of a random variable Xn to
have value xn at step n depends on the probability of the
variable Xn−1 to have the value xn−1 at step n − 1.
P(XN) = P(Xn, Xn−1)P(Xn−1) (29)
where P(Xn, Xn−1) is called the transition probability,
transition kernel or proposal density.
In most cases transition kernel is symmetric:
P(Xn, Xn−1) = P(Xn−1, Xn) (30)
20 / 35
Markov-Chain Monte Carlo
When we toss a coin n times then the outcome of the nth toss
does not depend on the outcome of any of the previous
outcomes.
In a Markov-Chain the probability of a random variable Xn to
have value xn at step n depends on the probability of the
variable Xn−1 to have the value xn−1 at step n − 1.
P(XN) = P(Xn, Xn−1)P(Xn−1) (29)
where P(Xn, Xn−1) is called the transition probability,
transition kernel or proposal density.
In most cases transition kernel is symmetric:
P(Xn, Xn−1) = P(Xn−1, Xn) (30)
The transition probability P(Xn, Xn−1) has the remarkable
property that after an initial burn-in period it generates a
sample which has the probability distribution P(X).
20 / 35
Example
We consider the following Gaussian transition probability:
P(xn|Xn−1) ∝ exp −
(Xn − Xn−1)2
2σ2
(31)
where σ is generally a fixed parameter called the step size. We can
go from (n − 1)th
step to nth step using the above transition
probability using the Metropolis-Hasting algorithm.
The choice of proposal density can affect the way the
sampling algorithm works so it is advisable to use a proposal
density which is of the similar shape of the distribution we are
aiming to sample.
[Lewis & Bridle (2002)]
21 / 35
Metropolis-Hesting Algorithm
The first step of the algorithm is to set the initial value of the
random variable i.e., X(n = 0) = X0 which should not be very
far from the best fit value (maximum likelihood value) of the
parameter.
Once the initial step is set, we can find a proposed value Y
for the (n + 1)th step using the proposal density P(Y |Xn).
In order to decide whether we should accept Y as Xn+1 we
compute the Metropolis ratio r
r =
P(Y )P(Xn|Y )
P(Xn)P(Y |Xn)
=
P(Y )
P(Xn)
for symmetric proposal density
If r ≥ 1 then we set Xn+1 = Y otherwise we accept Y with
probability r i.e., we draw a uniform random number U and
set Xn+1 = Y only when U > r.
[Gregory (2005)]
22 / 35
CMB Likelihood
CMB temperature and polarization observations can constrain
cosmological parameters if the likelihood function can be
computed exactly.
Computing the likelihood function exactly in a brute force way
is computationally challenging since it involves inversion of
the covariance matrix i.e., O(N3) computation.
In Cosmological parameter estimation a theoretical model is
represented by its angular power spectrum Cl .
For a set cosmological parameters we can compute the
angular power spectrum Cl using publicly available Boltzmann
codes like CMBFAST and CAMB and try to fit that with
observed Cl .
23 / 35
From Bayes theorem the posterior for the parameter Cl with data T
is given by:
P(Cl |T) =
P(T|Cl )P(Cl )
P(T)
(32)
where Cl is the theoretical Cl and T is the observed sky map of
CMB anisotropies.
T(ˆn) =
lm
almYlm(ˆn) (33)
Since computing the exact likelihood function is challenging,
approximations are generally made (Gaussian Likelihood, Gibbs
sampling etc).
L(T|Cl ) ∝
1
|S|
exp[−(TS−1
T)/2] (34)
where the covariance matrix S is related to angular power spectrum:
< T(ˆni )T(ˆnj ) >= Sij =
l
2l + 1
4π
Cl Pl (ˆni .ˆnj ) (35)
[Verde et al. (2003); Hamimeche & Lewis (2008)]
24 / 35
In terms of Cl the likelihood function can be written as:
L(T|Cl ) =
lm
1
√
Cl
exp[−|alm|2
/(2Cl )] (36)
Since we observe only one sky so we cannot measure the
power spectra directly, but instead form the rotationally
invariant estimators, Cl , for full-sky CMB maps given by
ˆCl =
1
2l + 1
m=l
m=−l
|alm|2
(37)
Problem 5
Show that ˆCl as given by equation (37) is an unbiased estimator
i.e., < ˆCl >= Cl (true power spectrum).
Note that the likelihood has a maximum when Cl = ˆCl so ˆCl
is the MLE.
25 / 35
Problem 6
From equation (36) show that:
χ2
= −2 log L(ˆCl |Cl ) =
l
(2l + 1) log
Cl
ˆCl
+
ˆCl
Cl
− 1 (38)
This expression for likelihood does not consider:
Finite resolution of the detector - window function
Detector noise
Cut-sky fsky to avoid foreground etc., which leads correlations
among different multi-poles.
When all these factors are taken into account computing the
likelihood function becomes challenging.
26 / 35
CosmoMC
In general, cosmological parameters from a CMB experiment
like WMAP or Planck are estimated using a publicly available
Markov Chain Monte Carlo code called CosmoMC [Lewis &
Bridle (2002)].
CosmoMC has been successfully used to estimate
cosmological parameters from WMAP data and a detail
discussion and working of the code is also discussed in a
WMAP first year paper [Verde et al. (2003)]
CosmoMC used publicly available code CAMB [Lewis et al.
(2000)] for computing theoretical Cl s.
WMAP team has provided a code for computing the
likelihood from the temperature and polarization data.
27 / 35
Cosmological Parameters
S. No Parameter Description
1 Ωbh2
physical baryon density
2 ΩDMh2
physical dark matter density (CDM+massive neutrinos)
3 θ 100 × ratio of the angular diameter distance to the LSS sound horizon
4 τ reionization optical depth
5 ΩK spatial curvature
6 fν neutrino energy density as fraction of ΩDMh2
7 w constant equation of state parameter for scalar field dark energy
8 ns spectral index for scalar power spectrum
9 nt spectral index for tensor power spectrum
10 nrun running for the index for scalar power spectrum
11 log[1010
As ] Amplitude of the scalar power spectrum
12 r ratio of tensor to scalar primordial amplitudes at pivot scale
13 ASZ SZ template amplitude, as in WMAP
14 ΩΛ energy density (parameter) Cosmological constant
15 Age/Gyr Age of the universe
16 Ωm dark matter density
17 σ8 Mass variance at 8 Mpc
18 zre Redshift of the reionization
19 r10 tensor-scalar Cl amplitude at l=10
20 H0 Hubble parameter is H0 km/s/Mpc
CosmoMC has 20 parameters which can be estimated from a CMB
data set. Note that not all the parameters are independent, in fact
we can change just seven parameters (shown in red) to fit a CMB
data set (WMAP).
28 / 35
How CosmoMC works?
Corresponding to every cosmological parameter θi we want to
estimate we give the (1) search range (2) an starting point
and guess (2) mean value and (3) standard deviation around
the mean.
At every step we compute the likelihood at current location Θ
in the parameter space and move to new location Θ1 in the
parameter space by taking a random step and compute:
r =
L(Θ1)
L(Θ1)
(39)
if r > 1 then we accept Θ1 as the new point in the chain
otherwise compare r with a uniform random number u and
accept Θ1 only when r > u.
We start multiple chains (random walks) in parallel and test
for a convergence criteria at every step.
Once we reach convergence, we compute the various
statistical measures from the chains.
29 / 35
CosmoMC
A typical Markov chain in CosmoMC.
30 / 35
CosmoMC
As a chain progresses χ2 = −2 log L decreases
31 / 35
CosmoMC
From the chains we can plot the probability distribution for
parameters. 32 / 35
CosmoMC
We can also plot scatter and contour plots from the chains.
33 / 35
Summary and Conclusions
34 / 35
References
Chu, M., Kaplinghat, M., & Knox, L. 2003, Astrophys. J. , 596, 725
Gregory, P. C. 2005, Bayesian Logical Data Analysis for the Physical
Sciences: A Comparative Approach with ‘Mathematica’ Support
(Cambridge University Press)
Hamimeche, S., & Lewis, A. 2008, Phys. Rev. D , 77, 103013
Lewis, A., & Bridle, S. 2002, Phys. Rev. D , 66, 103511
Lewis, A., Challinor, A., & Lasenby, A. 2000, Astrophys. J. , 538, 473
Tegmark, M. 1997, Astrophys. J. Lett. , 480, L87
Verde, L., et al. 2003, Astrophys. J. Suppl. Ser. , 148, 195
35 / 35

More Related Content

What's hot

Giáo trình Phân tích và thiết kế giải thuật - CHAP 8
Giáo trình Phân tích và thiết kế giải thuật - CHAP 8Giáo trình Phân tích và thiết kế giải thuật - CHAP 8
Giáo trình Phân tích và thiết kế giải thuật - CHAP 8Nguyễn Công Hoàng
 
Approximate Thin Plate Spline Mappings
Approximate Thin Plate Spline MappingsApproximate Thin Plate Spline Mappings
Approximate Thin Plate Spline MappingsArchzilon Eshun-Davies
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationAlexander Litvinenko
 
Modeling and quantification of uncertainties in numerical aerodynamics
Modeling and quantification of uncertainties in numerical aerodynamicsModeling and quantification of uncertainties in numerical aerodynamics
Modeling and quantification of uncertainties in numerical aerodynamicsAlexander Litvinenko
 
Lecture warshall floyd
Lecture warshall floydLecture warshall floyd
Lecture warshall floydDivya Ks
 
Concepts of Maxima And Minima
Concepts of Maxima And MinimaConcepts of Maxima And Minima
Concepts of Maxima And MinimaJitin Pillai
 
Computation of electromagnetic fields scattered from dielectric objects of un...
Computation of electromagnetic fields scattered from dielectric objects of un...Computation of electromagnetic fields scattered from dielectric objects of un...
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
 
Calculus and Numerical Method =_=
Calculus and Numerical Method =_=Calculus and Numerical Method =_=
Calculus and Numerical Method =_=Fazirah Zyra
 
A mid point ellipse drawing algorithm on a hexagonal grid
A mid  point ellipse drawing algorithm on a hexagonal gridA mid  point ellipse drawing algorithm on a hexagonal grid
A mid point ellipse drawing algorithm on a hexagonal gridS M K
 
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...theijes
 
Presentation of daa on approximation algorithm and vertex cover problem
Presentation of daa on approximation algorithm and vertex cover problem Presentation of daa on approximation algorithm and vertex cover problem
Presentation of daa on approximation algorithm and vertex cover problem sumit gyawali
 
Применение машинного обучения для навигации и управления роботами
Применение машинного обучения для навигации и управления роботамиПрименение машинного обучения для навигации и управления роботами
Применение машинного обучения для навигации и управления роботамиSkolkovo Robotics Center
 

What's hot (19)

Giáo trình Phân tích và thiết kế giải thuật - CHAP 8
Giáo trình Phân tích và thiết kế giải thuật - CHAP 8Giáo trình Phân tích và thiết kế giải thuật - CHAP 8
Giáo trình Phân tích và thiết kế giải thuật - CHAP 8
 
MUMS: Transition & SPUQ Workshop - Practical Bayesian Optimization for Urban ...
MUMS: Transition & SPUQ Workshop - Practical Bayesian Optimization for Urban ...MUMS: Transition & SPUQ Workshop - Practical Bayesian Optimization for Urban ...
MUMS: Transition & SPUQ Workshop - Practical Bayesian Optimization for Urban ...
 
Fourier Transform Assignment Help
Fourier Transform Assignment HelpFourier Transform Assignment Help
Fourier Transform Assignment Help
 
Approximate Thin Plate Spline Mappings
Approximate Thin Plate Spline MappingsApproximate Thin Plate Spline Mappings
Approximate Thin Plate Spline Mappings
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantification
 
Modeling and quantification of uncertainties in numerical aerodynamics
Modeling and quantification of uncertainties in numerical aerodynamicsModeling and quantification of uncertainties in numerical aerodynamics
Modeling and quantification of uncertainties in numerical aerodynamics
 
10.1.1.630.8055
10.1.1.630.805510.1.1.630.8055
10.1.1.630.8055
 
Lecture warshall floyd
Lecture warshall floydLecture warshall floyd
Lecture warshall floyd
 
Concepts of Maxima And Minima
Concepts of Maxima And MinimaConcepts of Maxima And Minima
Concepts of Maxima And Minima
 
Computation of electromagnetic fields scattered from dielectric objects of un...
Computation of electromagnetic fields scattered from dielectric objects of un...Computation of electromagnetic fields scattered from dielectric objects of un...
Computation of electromagnetic fields scattered from dielectric objects of un...
 
Calculus and Numerical Method =_=
Calculus and Numerical Method =_=Calculus and Numerical Method =_=
Calculus and Numerical Method =_=
 
Linkedin_PowerPoint
Linkedin_PowerPointLinkedin_PowerPoint
Linkedin_PowerPoint
 
Statistics Assignment Help
Statistics Assignment HelpStatistics Assignment Help
Statistics Assignment Help
 
Signal Processing Assignment Help
Signal Processing Assignment HelpSignal Processing Assignment Help
Signal Processing Assignment Help
 
A mid point ellipse drawing algorithm on a hexagonal grid
A mid  point ellipse drawing algorithm on a hexagonal gridA mid  point ellipse drawing algorithm on a hexagonal grid
A mid point ellipse drawing algorithm on a hexagonal grid
 
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...
An Efficient Elliptic Curve Cryptography Arithmetic Using Nikhilam Multiplica...
 
Presentation of daa on approximation algorithm and vertex cover problem
Presentation of daa on approximation algorithm and vertex cover problem Presentation of daa on approximation algorithm and vertex cover problem
Presentation of daa on approximation algorithm and vertex cover problem
 
Применение машинного обучения для навигации и управления роботами
Применение машинного обучения для навигации и управления роботамиПрименение машинного обучения для навигации и управления роботами
Применение машинного обучения для навигации и управления роботами
 
Cs 2008(1)
Cs 2008(1)Cs 2008(1)
Cs 2008(1)
 

Similar to CMB Lecture on Parameter Estimation Using MCMC Methods

An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...Alexander Decker
 
New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...Alexander Litvinenko
 
Generalized Nonlinear Models in R
Generalized Nonlinear Models in RGeneralized Nonlinear Models in R
Generalized Nonlinear Models in Rhtstatistics
 
Traveling Salesman Problem in Distributed Environment
Traveling Salesman Problem in Distributed EnvironmentTraveling Salesman Problem in Distributed Environment
Traveling Salesman Problem in Distributed Environmentcsandit
 
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENT
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENTTRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENT
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENTcscpconf
 
Numerical modeling-of-gas-turbine-engines
Numerical modeling-of-gas-turbine-enginesNumerical modeling-of-gas-turbine-engines
Numerical modeling-of-gas-turbine-enginesCemal Ardil
 
Chapter_09_ParameterEstimation.pptx
Chapter_09_ParameterEstimation.pptxChapter_09_ParameterEstimation.pptx
Chapter_09_ParameterEstimation.pptxVimalMehta19
 
DAA - UNIT 4 - Engineering.pptx
DAA - UNIT 4 - Engineering.pptxDAA - UNIT 4 - Engineering.pptx
DAA - UNIT 4 - Engineering.pptxvaishnavi339314
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big DataChristian Robert
 
Sparse data formats and efficient numerical methods for uncertainties in nume...
Sparse data formats and efficient numerical methods for uncertainties in nume...Sparse data formats and efficient numerical methods for uncertainties in nume...
Sparse data formats and efficient numerical methods for uncertainties in nume...Alexander Litvinenko
 
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...Varad Meru
 
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdfMohamedshabana38
 

Similar to CMB Lecture on Parameter Estimation Using MCMC Methods (20)

Presentation.pdf
Presentation.pdfPresentation.pdf
Presentation.pdf
 
An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...
 
New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...
 
likelihood_p1.pdf
likelihood_p1.pdflikelihood_p1.pdf
likelihood_p1.pdf
 
Generalized Nonlinear Models in R
Generalized Nonlinear Models in RGeneralized Nonlinear Models in R
Generalized Nonlinear Models in R
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
 
Network Security CS3-4
Network Security CS3-4 Network Security CS3-4
Network Security CS3-4
 
Traveling Salesman Problem in Distributed Environment
Traveling Salesman Problem in Distributed EnvironmentTraveling Salesman Problem in Distributed Environment
Traveling Salesman Problem in Distributed Environment
 
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENT
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENTTRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENT
TRAVELING SALESMAN PROBLEM IN DISTRIBUTED ENVIRONMENT
 
CMB Likelihood Part 1
CMB Likelihood Part 1CMB Likelihood Part 1
CMB Likelihood Part 1
 
Numerical modeling-of-gas-turbine-engines
Numerical modeling-of-gas-turbine-enginesNumerical modeling-of-gas-turbine-engines
Numerical modeling-of-gas-turbine-engines
 
Chapter_09_ParameterEstimation.pptx
Chapter_09_ParameterEstimation.pptxChapter_09_ParameterEstimation.pptx
Chapter_09_ParameterEstimation.pptx
 
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
 
DAA - UNIT 4 - Engineering.pptx
DAA - UNIT 4 - Engineering.pptxDAA - UNIT 4 - Engineering.pptx
DAA - UNIT 4 - Engineering.pptx
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
 
Sparse data formats and efficient numerical methods for uncertainties in nume...
Sparse data formats and efficient numerical methods for uncertainties in nume...Sparse data formats and efficient numerical methods for uncertainties in nume...
Sparse data formats and efficient numerical methods for uncertainties in nume...
 
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
 
Chang etal 2012a
Chang etal 2012aChang etal 2012a
Chang etal 2012a
 
15
1515
15
 
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf
27-Equivalent under modulo-27-Oct-2020Material_I_27-Oct-2020_Cryptography_.pdf
 

More from Jayanti Prasad Ph.D. (12)

parallel-computation.pdf
parallel-computation.pdfparallel-computation.pdf
parallel-computation.pdf
 
thesis_seminar.pdf
thesis_seminar.pdfthesis_seminar.pdf
thesis_seminar.pdf
 
pso2015.pdf
pso2015.pdfpso2015.pdf
pso2015.pdf
 
nbody_iitk.pdf
nbody_iitk.pdfnbody_iitk.pdf
nbody_iitk.pdf
 
likelihood_p2.pdf
likelihood_p2.pdflikelihood_p2.pdf
likelihood_p2.pdf
 
doing_parallel.pdf
doing_parallel.pdfdoing_parallel.pdf
doing_parallel.pdf
 
Artificial Intelligence - Anna Uni -v1.pdf
Artificial Intelligence - Anna Uni -v1.pdfArtificial Intelligence - Anna Uni -v1.pdf
Artificial Intelligence - Anna Uni -v1.pdf
 
Scikit-learn1
Scikit-learn1Scikit-learn1
Scikit-learn1
 
CMB Likelihood Part 2
CMB Likelihood Part 2CMB Likelihood Part 2
CMB Likelihood Part 2
 
Cmb part2
Cmb part2Cmb part2
Cmb part2
 
Cmb part1
Cmb part1Cmb part1
Cmb part1
 
Parallel computation
Parallel computationParallel computation
Parallel computation
 

Recently uploaded

Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentationtahreemzahra82
 
Forest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantForest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantadityabhardwaj282
 
Heredity: Inheritance and Variation of Traits
Heredity: Inheritance and Variation of TraitsHeredity: Inheritance and Variation of Traits
Heredity: Inheritance and Variation of TraitsCharlene Llagas
 
Temporomandibular joint Muscles of Mastication
Temporomandibular joint Muscles of MasticationTemporomandibular joint Muscles of Mastication
Temporomandibular joint Muscles of Masticationvidulajaib
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxSwapnil Therkar
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trssuser06f238
 
‏‏VIRUS - 123455555555555555555555555555555555555555
‏‏VIRUS -  123455555555555555555555555555555555555555‏‏VIRUS -  123455555555555555555555555555555555555555
‏‏VIRUS - 123455555555555555555555555555555555555555kikilily0909
 
Evidences of Evolution General Biology 2
Evidences of Evolution General Biology 2Evidences of Evolution General Biology 2
Evidences of Evolution General Biology 2John Carlo Rollon
 
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxLIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxmalonesandreagweneth
 
insect anatomy and insect body wall and their physiology
insect anatomy and insect body wall and their  physiologyinsect anatomy and insect body wall and their  physiology
insect anatomy and insect body wall and their physiologyDrAnita Sharma
 
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |aasikanpl
 
Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)DHURKADEVIBASKAR
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxyaramohamed343013
 
Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Nistarini College, Purulia (W.B) India
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.aasikanpl
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tanta
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tantaDashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tanta
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tantaPraksha3
 
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfAnalytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfSwapnil Therkar
 
zoogeography of pakistan.pptx fauna of Pakistan
zoogeography of pakistan.pptx fauna of Pakistanzoogeography of pakistan.pptx fauna of Pakistan
zoogeography of pakistan.pptx fauna of Pakistanzohaibmir069
 
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |aasikanpl
 

Recently uploaded (20)

Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentation
 
Forest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantForest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are important
 
Heredity: Inheritance and Variation of Traits
Heredity: Inheritance and Variation of TraitsHeredity: Inheritance and Variation of Traits
Heredity: Inheritance and Variation of Traits
 
Temporomandibular joint Muscles of Mastication
Temporomandibular joint Muscles of MasticationTemporomandibular joint Muscles of Mastication
Temporomandibular joint Muscles of Mastication
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 tr
 
‏‏VIRUS - 123455555555555555555555555555555555555555
‏‏VIRUS -  123455555555555555555555555555555555555555‏‏VIRUS -  123455555555555555555555555555555555555555
‏‏VIRUS - 123455555555555555555555555555555555555555
 
Evidences of Evolution General Biology 2
Evidences of Evolution General Biology 2Evidences of Evolution General Biology 2
Evidences of Evolution General Biology 2
 
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxLIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
 
insect anatomy and insect body wall and their physiology
insect anatomy and insect body wall and their  physiologyinsect anatomy and insect body wall and their  physiology
insect anatomy and insect body wall and their physiology
 
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
 
Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docx
 
Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
 
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tanta
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tantaDashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tanta
Dashanga agada a formulation of Agada tantra dealt in 3 Rd year bams agada tanta
 
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfAnalytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
 
zoogeography of pakistan.pptx fauna of Pakistan
zoogeography of pakistan.pptx fauna of Pakistanzoogeography of pakistan.pptx fauna of Pakistan
zoogeography of pakistan.pptx fauna of Pakistan
 
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Lajpat Nagar (Delhi) |
 

CMB Lecture on Parameter Estimation Using MCMC Methods

  • 1. Cosmic Microwave Background Radiation Lecture 3 : Parameter Estimation Jayanti Prasad Inter-University Centre for Astronomy & Astrophysics (IUCAA) Pune, India (411007) Autumn School on Cosmology (5 - 15th Nov 2013) BITS PILANI 1 / 35
  • 2. CMB Data Analysis pipeline In every step of CMB data analysis the aim is to reduce the volume of data without losing information. 2 / 35
  • 3. Plan of the Talk Data Analysis Techniques Inverse Problems Chi-Square minimization Maximum-Likelihood Estimation 3 / 35
  • 4. Plan of the Talk Data Analysis Techniques Inverse Problems Chi-Square minimization Maximum-Likelihood Estimation CMB data and Map making 3 / 35
  • 5. Plan of the Talk Data Analysis Techniques Inverse Problems Chi-Square minimization Maximum-Likelihood Estimation CMB data and Map making Cosmological Parameter Estimation CMB Likelihood Markov Chain Monte Carlo Methods 3 / 35
  • 6. Plan of the Talk Data Analysis Techniques Inverse Problems Chi-Square minimization Maximum-Likelihood Estimation CMB data and Map making Cosmological Parameter Estimation CMB Likelihood Markov Chain Monte Carlo Methods WMAP and Planck 3 / 35
  • 7. Plan of the Talk Data Analysis Techniques Inverse Problems Chi-Square minimization Maximum-Likelihood Estimation CMB data and Map making Cosmological Parameter Estimation CMB Likelihood Markov Chain Monte Carlo Methods WMAP and Planck Summary and Conclusions 3 / 35
  • 8. Data Analysis Techniques Scientific observations can be represented by a data vector d which can be a time series {t} or temperature map of the sky {T} or something else. Data have information about some physical process for which we have a theoretical model represented by a set of parameters i.e., parameter vector Θ. One of its example is a Gaussian process represented by two parameters i.e., the mean µ and the variance σ2. The probability of obtaining data d given a theoretical (Gaussian model (µ, σ2)) is given by: P(d|µ, σ2 ) = 1 √ 2πσ2 exp − 1 2 (d − µ)2 σ2 (1) 4 / 35
  • 9. Data Analysis Techniques Note that finding out which of the theoretical model is better than others e.g., we should be fitting a parabola rather than a line, is different from finding the parameters of a model. In the present discussion we will not discuss model comparison, we will always assume a model is true and will try to find its parameters i.e., what are the values of the parameters of ΛCDM cosmological model. Values of parameters do not make sense unless we specify a model. Once we have a model (with its parameters) we can easily create data by simulating that model and this is called a forward problem. Finding parameters Θ of a model from the data d is called the inverse problem which is much harder to solve than the forward problem. 5 / 35
  • 10. Linear Problem A problem is said to be a linear problem when the data d depends on model parameters linearly, for example fitting a polynomial is a linear problem: di = M j=1 cj θj for i = 1, N (2) where ci are the model parameters. By definition there is no linear relationship between the data and model parameter for a non-linear problem, for example, fitting a Gaussian is a non-linear problem. Non-linear problems are harder to solve than linear problems. Dependency of cosmological data i.e., CMB map, on cosmological parameters is a non-linear problem. 6 / 35
  • 11. Solvable problems Not all the inverse problems are solvable and in some cases we can easily find out why that is so. On the basis of whether the size N of the data vector d is larger, smaller or equal to the size M of the parameter vector Θ, there are three possibilities. N = M : Unique solution is possible N > M : Over constrained problem, χ2 minimization, Unique solution N < M : Under-constrained problem, ill posed problem, priors, regularization Note that in the above consideration we have assumed that all the data points are independent. One of the common methods to solve an inverse is to minimize a measure of misfit between the data and the theoretical model. 7 / 35
  • 12. Map Making in CMB In CMB experiments like WMAP and Planck the time order data or TOD d depends on the sky temperature T in the following way: Di = Aij Tj + Nj (3) the index i and j are over time and pixels respectively. In order to make a map from the TOD we have to estimate T from D for which we can minimize the following function: f (T) = (D − AT) (D − AT) (4) which gives: ˆT = (A A)−1 A D (5) 8 / 35
  • 13. CMB Map making In place of f (T) if we can also minimize χ2 : χ2 (T) = (D − AT) C−1 N (D − AT) (6) where CN =< NN > is the noise covariance matrix, then we get: ˆT = (A C−1 N A)−1 A C−1 N D (7) which is called Maximum-Likelihood (ML) solution. The covariance matrix of the maximum likelihood solution is given by : N = (A C−1 N A)−1 (8) [Tegmark (1997)] 9 / 35
  • 14. Chi-Square Distribution If we have a variable y = N i=1 X2 i and X is drwan from Gaussian random distribution then Y follows χ2 distribution of degree ν: fχ2 (x) = x ν 2 −1 e− x 2 Γ(ν 2 )2 ν 2 (9) 10 / 35
  • 15. Chi-Square Distribution If we have a variable y = N i=1 X2 i and X is drwan from Gaussian random distribution then Y follows χ2 distribution of degree ν: fχ2 (x) = x ν 2 −1 e− x 2 Γ(ν 2 )2 ν 2 (9) For large ν, χ2 distribution approaches Gaussian. 10 / 35
  • 16. CMB anisotropies CMB temperature anisotropies are expressed in terms of multipoles: ∆T(ˆn) T = ∞ l=0 m=l m=−l almYlm(ˆn) (10) where alm = ∆T(ˆn) T Ylm(ˆn)dˆn (11) Where alm follow the Gaussian distribution with zero mean and variance given by Cl : < alm >= 0 (12) and alma∗ l m = δll δmm Cl (13) An unbiased estimator of Cl is defined as: ˆCl = 1 2l + 1 m=l m=−l alma∗ lm (14) Note that the angular power spectrum Cl follows χ2 distribution. 11 / 35
  • 17. Maximum Likelihood Estimation The probability distribution P(Θ|d) (posterior) for model parameters Θ given data d can be related to the probability P(d|Θ) (likelihood) of an experiment giving data d for model parameters Θ using the Bayes’ theorem: P(Θ|d) = P(d|Θ)P(Θ) P(d) (15) where P(Θ) is called the prior and P(d) = P(d|Θ)P(Θ) is used for the normalization purpose. For the case of flat prior, posterior and likelihood are proportional: P(Θ|d) = P(d|Θ) = L(d|Θ) (16) In Bayesian formalism we can easily incorporate new data in analysis by considering the posterior of the old data as prior. 12 / 35
  • 18. Bayesian Analysis Note that when likelihood/posterior is not Gaussian then the average value of the parameter < θ > may not coincide with the value of θ0 at which the likelihood/posterior is maximum. 13 / 35
  • 19. Errors in maximum Likelihood estimation We are not only interested in finding the point Θ0 or < Θ >, we are also also interested errors. The spread of the likelihood function P(d|Θ) around the maximum likelihood point Θ0 can be used to find the error bars: Expanding the likelihood function L = −2 log log P(d|Θ) around the maximum likelihood point Θ0: L(Θ) = L(Θ0)+(θi −θi 0) ∂L(Θ) ∂θi θi =θi 0 + 1 2 (θi −θi 0)(θj −θj 0) ∂2 L(Θ) ∂θi ∂θj θi =θi 0 +.... (17) Second derivative of the Likelihood function is called the Hessian : Hij = ∂2L(Θ) ∂θi ∂θj (18) 14 / 35
  • 20. There is an in-equality called the Cramer-Rao lower bound which says that the error bars (variance) on any estimator cannot be smaller than the inverse of information matrix I Cij ≥ 1 Iij (19) where I is defined as: Iij = ∂2L(Θ) ∂θi ∂θj (20) How accurately a parameter can be estimated from the likelihood depends upon how sensitive the likelihood function is on that parameter, which is quantified by the Hessian matrix. 15 / 35
  • 21. Problems 1 Show that for a case when the noise is Gaussian, maximizing likelihood is equivalent to minimization Chi-square. 2 Show that the maximum likelihood estimator is the minimum variance estimator. 3 Show that finding the solution of the linear problem d = AΘ is equivalent to finding of the maximum of the following function: f (Θ) = 1 2 Θ AΘ − d Θ + c (21) 4 Show that the minimum chi-square solution of the model y = c0 + c1x with data (xi , yi , σi ) is given by: c0 = S01S00 − S00S01 SS00 − S2 00 and c1 = SS01 − S01S10 SS00 − S2 00 (22) where S = N i=1 1/σ2 i and Sij = N k=1 xi yj /σ2 k . Also find the error. 16 / 35
  • 22. Sometime likelihood is not much sensitive to individual parameters (may not be Gaussian) but is highly sensitive to some combinations of those. For example, CMB Likelihood is almost Gaussian with respect to the combination (Ωbh2, Ωch2, θ, A∗, tz) of cosmological parameters [Chu et al. (2003)] where A∗ = A 76, 000 0.05Mpc−1 kpivot 1−ns e−2τ , (23) and t = 1 Ωbh2 2ns −1 (24) 17 / 35
  • 23. Markov Chain Monte Carlo Once we have the probability distribution P(Θ|d) for model parameters Θ we can statistics of the parameters: < Θ >= ΘdΘP(Θ|d) (25) In practice, before carrying out the above integral we find out the one dimensional probability distribution by marginalization over other parameters: P(θr ) = dθ1dθ2....dθr−1dθr+1...dθMP(θ1, θ2, ....sθM) (26) and < θr >= θr dθr P(θr ) (27) 18 / 35
  • 24. Carrying out multi-dimensional integration is very expansive i.e., computational cost grows as O(nM) where n is the number of grid points along one direction and M is the dimensionality of the parameter space. If we can replace the multi-dimensional integration by summation over a finite number of points which represent the probability distribution function then computational cost becomes manageable. < Θ >= ΘdΘP(Θ|d) = 1 N N i=1 Θi P(Θi |d) (28) Markov-Chain Monte Carlo sampling samples the likelihood function in such a way that there are more point in the region where the likelihood function has the large values and less where it has small values. 19 / 35
  • 25. Markov-Chain Monte Carlo When we toss a coin n times then the outcome of the nth toss does not depend on the outcome of any of the previous outcomes. 20 / 35
  • 26. Markov-Chain Monte Carlo When we toss a coin n times then the outcome of the nth toss does not depend on the outcome of any of the previous outcomes. In a Markov-Chain the probability of a random variable Xn to have value xn at step n depends on the probability of the variable Xn−1 to have the value xn−1 at step n − 1. P(XN) = P(Xn, Xn−1)P(Xn−1) (29) where P(Xn, Xn−1) is called the transition probability, transition kernel or proposal density. 20 / 35
  • 27. Markov-Chain Monte Carlo When we toss a coin n times then the outcome of the nth toss does not depend on the outcome of any of the previous outcomes. In a Markov-Chain the probability of a random variable Xn to have value xn at step n depends on the probability of the variable Xn−1 to have the value xn−1 at step n − 1. P(XN) = P(Xn, Xn−1)P(Xn−1) (29) where P(Xn, Xn−1) is called the transition probability, transition kernel or proposal density. In most cases transition kernel is symmetric: P(Xn, Xn−1) = P(Xn−1, Xn) (30) 20 / 35
  • 28. Markov-Chain Monte Carlo When we toss a coin n times then the outcome of the nth toss does not depend on the outcome of any of the previous outcomes. In a Markov-Chain the probability of a random variable Xn to have value xn at step n depends on the probability of the variable Xn−1 to have the value xn−1 at step n − 1. P(XN) = P(Xn, Xn−1)P(Xn−1) (29) where P(Xn, Xn−1) is called the transition probability, transition kernel or proposal density. In most cases transition kernel is symmetric: P(Xn, Xn−1) = P(Xn−1, Xn) (30) The transition probability P(Xn, Xn−1) has the remarkable property that after an initial burn-in period it generates a sample which has the probability distribution P(X). 20 / 35
  • 29. Example We consider the following Gaussian transition probability: P(xn|Xn−1) ∝ exp − (Xn − Xn−1)2 2σ2 (31) where σ is generally a fixed parameter called the step size. We can go from (n − 1)th step to nth step using the above transition probability using the Metropolis-Hasting algorithm. The choice of proposal density can affect the way the sampling algorithm works so it is advisable to use a proposal density which is of the similar shape of the distribution we are aiming to sample. [Lewis & Bridle (2002)] 21 / 35
  • 30. Metropolis-Hesting Algorithm The first step of the algorithm is to set the initial value of the random variable i.e., X(n = 0) = X0 which should not be very far from the best fit value (maximum likelihood value) of the parameter. Once the initial step is set, we can find a proposed value Y for the (n + 1)th step using the proposal density P(Y |Xn). In order to decide whether we should accept Y as Xn+1 we compute the Metropolis ratio r r = P(Y )P(Xn|Y ) P(Xn)P(Y |Xn) = P(Y ) P(Xn) for symmetric proposal density If r ≥ 1 then we set Xn+1 = Y otherwise we accept Y with probability r i.e., we draw a uniform random number U and set Xn+1 = Y only when U > r. [Gregory (2005)] 22 / 35
  • 31. CMB Likelihood CMB temperature and polarization observations can constrain cosmological parameters if the likelihood function can be computed exactly. Computing the likelihood function exactly in a brute force way is computationally challenging since it involves inversion of the covariance matrix i.e., O(N3) computation. In Cosmological parameter estimation a theoretical model is represented by its angular power spectrum Cl . For a set cosmological parameters we can compute the angular power spectrum Cl using publicly available Boltzmann codes like CMBFAST and CAMB and try to fit that with observed Cl . 23 / 35
  • 32. From Bayes theorem the posterior for the parameter Cl with data T is given by: P(Cl |T) = P(T|Cl )P(Cl ) P(T) (32) where Cl is the theoretical Cl and T is the observed sky map of CMB anisotropies. T(ˆn) = lm almYlm(ˆn) (33) Since computing the exact likelihood function is challenging, approximations are generally made (Gaussian Likelihood, Gibbs sampling etc). L(T|Cl ) ∝ 1 |S| exp[−(TS−1 T)/2] (34) where the covariance matrix S is related to angular power spectrum: < T(ˆni )T(ˆnj ) >= Sij = l 2l + 1 4π Cl Pl (ˆni .ˆnj ) (35) [Verde et al. (2003); Hamimeche & Lewis (2008)] 24 / 35
  • 33. In terms of Cl the likelihood function can be written as: L(T|Cl ) = lm 1 √ Cl exp[−|alm|2 /(2Cl )] (36) Since we observe only one sky so we cannot measure the power spectra directly, but instead form the rotationally invariant estimators, Cl , for full-sky CMB maps given by ˆCl = 1 2l + 1 m=l m=−l |alm|2 (37) Problem 5 Show that ˆCl as given by equation (37) is an unbiased estimator i.e., < ˆCl >= Cl (true power spectrum). Note that the likelihood has a maximum when Cl = ˆCl so ˆCl is the MLE. 25 / 35
  • 34. Problem 6 From equation (36) show that: χ2 = −2 log L(ˆCl |Cl ) = l (2l + 1) log Cl ˆCl + ˆCl Cl − 1 (38) This expression for likelihood does not consider: Finite resolution of the detector - window function Detector noise Cut-sky fsky to avoid foreground etc., which leads correlations among different multi-poles. When all these factors are taken into account computing the likelihood function becomes challenging. 26 / 35
  • 35. CosmoMC In general, cosmological parameters from a CMB experiment like WMAP or Planck are estimated using a publicly available Markov Chain Monte Carlo code called CosmoMC [Lewis & Bridle (2002)]. CosmoMC has been successfully used to estimate cosmological parameters from WMAP data and a detail discussion and working of the code is also discussed in a WMAP first year paper [Verde et al. (2003)] CosmoMC used publicly available code CAMB [Lewis et al. (2000)] for computing theoretical Cl s. WMAP team has provided a code for computing the likelihood from the temperature and polarization data. 27 / 35
  • 36. Cosmological Parameters S. No Parameter Description 1 Ωbh2 physical baryon density 2 ΩDMh2 physical dark matter density (CDM+massive neutrinos) 3 θ 100 × ratio of the angular diameter distance to the LSS sound horizon 4 τ reionization optical depth 5 ΩK spatial curvature 6 fν neutrino energy density as fraction of ΩDMh2 7 w constant equation of state parameter for scalar field dark energy 8 ns spectral index for scalar power spectrum 9 nt spectral index for tensor power spectrum 10 nrun running for the index for scalar power spectrum 11 log[1010 As ] Amplitude of the scalar power spectrum 12 r ratio of tensor to scalar primordial amplitudes at pivot scale 13 ASZ SZ template amplitude, as in WMAP 14 ΩΛ energy density (parameter) Cosmological constant 15 Age/Gyr Age of the universe 16 Ωm dark matter density 17 σ8 Mass variance at 8 Mpc 18 zre Redshift of the reionization 19 r10 tensor-scalar Cl amplitude at l=10 20 H0 Hubble parameter is H0 km/s/Mpc CosmoMC has 20 parameters which can be estimated from a CMB data set. Note that not all the parameters are independent, in fact we can change just seven parameters (shown in red) to fit a CMB data set (WMAP). 28 / 35
  • 37. How CosmoMC works? Corresponding to every cosmological parameter θi we want to estimate we give the (1) search range (2) an starting point and guess (2) mean value and (3) standard deviation around the mean. At every step we compute the likelihood at current location Θ in the parameter space and move to new location Θ1 in the parameter space by taking a random step and compute: r = L(Θ1) L(Θ1) (39) if r > 1 then we accept Θ1 as the new point in the chain otherwise compare r with a uniform random number u and accept Θ1 only when r > u. We start multiple chains (random walks) in parallel and test for a convergence criteria at every step. Once we reach convergence, we compute the various statistical measures from the chains. 29 / 35
  • 38. CosmoMC A typical Markov chain in CosmoMC. 30 / 35
  • 39. CosmoMC As a chain progresses χ2 = −2 log L decreases 31 / 35
  • 40. CosmoMC From the chains we can plot the probability distribution for parameters. 32 / 35
  • 41. CosmoMC We can also plot scatter and contour plots from the chains. 33 / 35
  • 43. References Chu, M., Kaplinghat, M., & Knox, L. 2003, Astrophys. J. , 596, 725 Gregory, P. C. 2005, Bayesian Logical Data Analysis for the Physical Sciences: A Comparative Approach with ‘Mathematica’ Support (Cambridge University Press) Hamimeche, S., & Lewis, A. 2008, Phys. Rev. D , 77, 103013 Lewis, A., & Bridle, S. 2002, Phys. Rev. D , 66, 103511 Lewis, A., Challinor, A., & Lasenby, A. 2000, Astrophys. J. , 538, 473 Tegmark, M. 1997, Astrophys. J. Lett. , 480, L87 Verde, L., et al. 2003, Astrophys. J. Suppl. Ser. , 148, 195 35 / 35