SlideShare a Scribd company logo
1 of 31
Probability function P on subspace of S
1. P(S)=1
2. For every event A in S, P(A)≥0
3. If A1, A2, .. are mutually exclusive, then
P(∪Ai)=∑ P(Ai)
A random variable X is a function that assigns a
value to each outcome s in the sample space
S (realizations of the random variable).
Example: dart, bulls eye counts 50, i.e. X(s)=50,
s = bulls eye location
Example: measurement of a mass five times,
yielding the true value m + random
Probability Density Function (PDF), fX(x)
The relative probability of realizations for a
random variable
P(X≤a)=∫ fX(x)dx
∫ fX(x)dx = 1
PDF for common functions:
1. Uniform random variable on [a,b]:
fU(x)=1/(b-a), a≤x≤b; 0, x<a or x>b
a
-∞
-∞
-∞
PDF for common functions:
2. Normal (Gaussian) function:
fN(x)=1/σ√2π exp(-0.5(x-µ)2
/σ2
)
N(µ,σ2
) normal distribution
3. Exponential function:
fexp(x)=λ exp(-λx), x≥0; 0, x<0
4. Double-sided Exponential function:
fdexp(x)=[1/(23/2
σ)] exp(-√2|x-µ|/σ)
5. χ2
function:
f χ2
= 1/[2ν/2
Γ(ν/2)] x 0.5ν-1
exp(-x/2)
with Γ(x)=∫ ξx-1
e-ξ
dξ
n independent variable with standard normal
distributions, Z=∑Xi
2
is a χ2
random variable
with ν=n degrees of freedom.
∞
0
6 . Student's t distribution with ν degrees of
freedom:
ft(x)= Γ((ν+1)/2)]/Γ(ν/2) 1/√νπ (1+x2
/ν)-(ν+1)/2
with Γ(x)=∫ ξx-1
e-ξ
dξ
Approaches a standard normal distribution for
large number of degrees of freedom
Cumulative distribution functions (CDF):
FX(a)=P(X≤a)=∫ fX(x)dx
P(a≤X≤b) = ∫f(x)dx
a
-∞
b
a
Characterization of PDF’s:
Expected value of random variable x:
E[X] = ∫ xfX(x)dx, E[g(X)] = ∫ g(x)fX(x)dx
Peak of distribution: XML (maximum likelihood)
Variance:
Var(X)=σX
2
= E[(X-µx)2
] = E[X2
]-µX
2
=
∫ (x-µx)2
fX(x)dx, σx=√Var(X)
Variance measures the width of PDF’s. Wide
PDF’s indicate noisy data, narrow indicate
∞
-∞
∞
-∞
∞
-∞
Joint PDF’s:
The joint PDF can quantify the probability that a
set of random variables will take on a given
value.
f(X≤a,Y≤b)= ∫ ∫ f(x,y)dydx
Expected value for joint PDF:
E[g(X,Y)]=∫ ∫ g(x,y)f(x,y)dydx
If X and Y are independent,
f(x,y)=fX(x)fY(y)
b
-∞
a
-∞
∞
-∞
∞
-∞
Covariance of X and Y with Joint PDF:
Cov(X,Y)=E[(X-E[X])(Y-E[Y])]=E[XY]-E[X]E[Y]
X and Y independent, E[XY]=E[X]E[Y] and
Cov(X,Y)=0
Covariance of a variable with itself = variance
If Cov(X,Y)=0, X and Y are called uncorrelated
Correlation of X and Y
ρ(X,Y)=Cov(X,Y)/√Var(X) Var(Y)
(correlation is a scaled covariance)
Model covariance matrix cov(m):
The general model mest
=Md + v
has covariance matrix M[cov d]MT
(B.64)
The least-squares solution mest
=[GT
G]-1
GT
d
has covariance matrix
[cov m] = [GT
G]-1
GT
[cov d] ([GT
G]-1
GT
)T
If data is uncorrelated and has equal variance
σd
2
then
[cov m] = σd
2
[GT
G]-1
More on Gaussian (Normal) distributions:
Central limit theorem:
Let X1, X2, …, Xn be independent and identically
distributed random variables with finite
expected value µ and variance σ2
. Let
Zn=[X1+X2+..+Xn-nµ]/√nσ
In the limit as n approaches infinity, Zn
approaches the standard normal distribution
Thus many summed variables in nature are
normal distributions, thus LS solutions OK
Means and confidence intervals:
Given noisy measurements m1, m2, .., mn
Estimate the true value m and the uncertainty of
the estimate. Assume errors are independent
and normally distributed with expected value 0
and some unknown standard deviation σ
Compute average mave
=[m1+m2+..+mn]/n
s=[∑ (mi-mave
)2
]1/2
/n-1
n
i=1
Sampling Theorem:
Independent, normally distributed measurements,
with expected value m and standard deviated σ,
the random quantity
t = (m-mave
)/s√n
has a Student’s t distribution with n-1 degrees of
freedom. If the true standard deviation σ is known,
we are dealing with a standard normal distribution.
The t distribution converges toward the normal
distribution for large n
Confidence intervals: Probability that one realization
falls within a specified distance of the true mean
Let tn-1,0.975 = 97.5 percentile of t distribution
tn-1,0.025 = 2.5 percentile of t distribution
P(tn-1,0.975 ≤m-mave
/(s/√n) ≤tn-1,0.025)=0.95
P(tn-1,0.975 s/ √n ≤m-mave
≤tn-1,0.025 s/√n )=0.95
95% confidence interval
Due to symmetry:
mave
-tn-1,0.975 s/ √n to mave
+tn-1,0.975 s/ √n
Confidence intervals related to the σ - PDF’s with
large σ will have large CI, and vice versa
Gaussian PDF, the 68% CI is 1σ wide and the 95% CI
is 2σ wide
If a particular Gaussian random variable has σ=1, and
if a realization of that variable is 50, there is a 95%
chance that the mean of that random variable lies
between 48 and 52
Example B.12 illustrates the case where σ is
estimated
If σ is known (rarely the case), we have a normal
distribution. We can use the t-distribution with an
infinite # of observations.
E.g., 16 obs, m estimated to be 31.5. σ is known to be
5. Estimate 80% CI for m.
mave
-k ≤ m ≤ mave
+k
k=1.282 x 5/√16 = 1.6
31.5-1.6 ≤ m ≤ 31.5+1.6
Statistical Aspects of LS:
PDF for Normal Distribution:
fi(di|m)=1/σι√2π exp(-(di-(Gm)i)2
/2σι
2
)
Maximum likelihood function L(m|d) is the product of all
individual probability functions:
L(m|d)=f1(d1|m)*f2(d2|m)* … * fm(dm|m)
Idea: Maximize L(m|d)
Maximize log{L(m|d)}
Minimize - log{L(m|d)}
Minimize -2 log{L(m|d)} min ∑ [di-(Gm)I]2
/σι
2
Aside from the 1/σι
2
factor, we have the LS solution
min ∑ [di-(Gm)I]2
/σι
2
W=diag(1/σ1, 1/σ2, …,1/σm)
Gw=WG
dw=Wd
Gwm=dw
mL2=[Gw
T
Gw)-1
Gw
T
dw
||dw-Gwmw||2
2
= ∑ [di-(Gm)I]2
/σι
2
χobs
2
= ∑ [di-(Gm)I]2
/σι
2
χobs
2
has a χ2
distribution with m-n degrees of freedom
The probability of finding a χ2
value as large or larger than the
observed value is
p=∫ fχ
2
(x)dx
p-value test
With correct model and independent error, the p-values will be
uniformly distributed between 0 and 1. Near-0 or near-1 p-
values indicates problems (incorrect model,
underestimation of data errors, unlikely realization)
∞
χobs
2
Multivariate normal distribution
Random variables X1, …, Xn have a multivariate normal
distribution, then the joint PDF is (B.61)
f(x)=(2π)-n/2
(det[C])-1/2
exp[-(x-µ)T
C-1
(x-µ)/2]
Ci,j=Cov(Xi,Xj)
Eigenvalues and eigenvectors
Ax= λ x
(A-λI)x=0
det (A-λI) = 0
The roots λ are the eigenvalues
SVD
G=USVT
U (m x m) orthogonal spanning data space
V (n x n) orthogonal spanning model space
S (m x n) eigenvalues along diagonal
Let rank(G)=p
G=[Up U0] [Sp 0;0 0] [Vp V0]T
= UpSpVp
T
G+
=VpSp
-1
Up
T =
Generalized Inverse (pseudoinverse)
m+
=G+
d=VpSp
-1
Up
T
d = pseudoinverse solution
Since Vp
-1
=Vp
T
and Up
-1
=Up
T
(A.6)
SVD
G=USVT
U (m x m) orthogonal spanning data space
V (n x n) orthogonal spanning model space
S (m x n) eigenvalues along diagonal
Rank(G)=p
Theorem A.5: N(GT
)+R(G)=Rm
, i.e.
p columns of Up form an orthonormal basis for R(G)
columns of U0 form an orthonormal basis for N(GT
)
p columns of Vp form an orthonormal basis for R(GT
)
columns of V0 form an orthonormal basis for N(G)
Properties of SVD
G=USVT
1. N(G) and N(GT
) are trivial (only null vector):
Up=U, Vp=V square orthogonal matrices,
and Up
T
=Up
-1
and Vp
T
=Vp
-1
G+
=VpSp
-1
Up
T
= (UpSpVp
T
)-1
= G-1
(inverse for full rank matrix, m=n=p). Unique solution,
data are fit exactly.
Properties of SVD
G=USVT
2. N(G) is nontrivial (model, V); N(GT
) is trivial (data, U)
Up
T
=Up
-1
and Vp
T
Vp=Ip
Gm+
=GG+
d=UpSpVp
T
VpSp
-1
Up
T
d=UpSpIpSp
-1
Up
T
d=d
i.e. the data are fit exactly an LS solution, but nonuniquely
due to the nontrivial model null space
m=m+
+m0 =m+
+∑ αiV.,I
||m||2
2
=||m+
||2
2
+∑ α2
i ≥||m+
||2
2
-> minimum length solution
n
i=p+1
n
i=p+1
Properties of SVD
G=USVT
3. N(G) is trivial (model, V); N(GT
) is nontrivial (data, U)
Gm+
=GG+
d=UpSpVp
T
VpSp
-1
Up
T
d=UpUp
T
d
= projection of d onto R(G), I.e. the point in R(G)
that is closest to d, m+
is LS solution to Gm=d.
If d is in R(G), m+
will be solution to Gm=d.
Solution is unique but does not fit data exactly
i.e. the data are fit exactly an LS solution, but nonuniquely
due to the nontrivial model null space
n
i=p+1
n
i=p+1
Properties of SVD
G=USVT
4. N(G) is nontrivial (model, V); N(GT
) is nontrivial (data, U)
p < (m,n)
Gm+
=GG+
d=UpSpVp
T
VpSp
-1
Up
T
d=UpUp
T
d
= projection of d onto R(G), I.e. the point in R(G)
that is closest to d, m+
is LS solution to Gm=d.
i.e. LS solution to minimum norm, as case 2)
Properties of SVD
G=USVT
- Always exists
- LS or minimum length
- Properly accommodates the rank and dimensions of G
- Nontrivial model null space m0 the is heart of the problem ->
- Infinite # of solutions will fit the data equally well, since
components of N(G) have no effect on data fit, I.e.,
selection of a particular solution requires a priori constraints
(smoothing, minimum length)
- Nontrivial data space are vectors d0 that have no influence
on m+
. If p<n, there are an infinite # of data sets that will
produce the same model
Properties of SVD - covariance/resolution
Least squares solution is unbiased:
min ∑ [di-(Gm)I]2
/σι
2
W=diag(1/σ1, 1/σ2, …,1/σm)
Gw=WG, dw=Wd, Gwm=dw
mL2=[Gw
T
Gw]-1
Gw
T
dw
||dw-Gwmw||2
2
= ∑ [di-(Gm)I]2
/σι
2
E[mL2]=E[(Gw
T
Gw)-1
Gw
T
dw] = (Gw
T
Gw)-1
Gw
T
E[dw]=
(Gw
T
Gw)-1
Gw
T
dw
true
= (Gw
T
Gw)-1
Gw
T
Gwmtrue
= mtrue
Properties of SVD - covariance/resolution
Generalized inverse not necessarily unbiased:
E[m+
]=E[G+
d] = G+
E[d] = G+
Gmtrue
= Rmmtrue
Bias= E[m+
]-mtrue =
Rmmtrue
-mtrue
= (Rm-I)mtrue
= VpVp
T
-VVT
mtrue
=-V0V0
T
mtrue
I.e., as p increasees Rm->I
Cov(mL2)=σ2
(GT
G)-1
Cov(m+
)=G+
[Cov(d)]G+T
= σ2
G+
G+T
= σ2
VpSp
-2
Vp
T
= σ2
∑ V.,i V.,i
T
/σi
2
I.e., as p increases, Rm->I:
P
I=1
Model resolution
Rm -> I: increasing resolution
Resolution test: multiply Rm onto a particular model, fx a spike
model, with one element 1 and the rest 0, picks out the
corresponding column of Rm
Data resolution
D+
=Gm+
= GG+
d = Rdd
Rd=UpSpVp
T
VpSp
-1
Up
T
=UpUp
T
p=m -> Rd=I, d+
=d
p<m -> Rd<>I, m+
doesn’t fit data exactly
Instabilitites of SVD
Small eigenvalues -> m+ sensitive to small amounts of noise
Small eigenvalues maybe indistinguishable from 0
Possible to remove small eigenvalues to stabilize solution ->
Truncated SVD, TSVD
Condition number cond(G)=s1/sk

More Related Content

What's hot

Can we estimate a constant?
Can we estimate a constant?Can we estimate a constant?
Can we estimate a constant?Christian Robert
 
Murphy: Machine learning A probabilistic perspective: Ch.9
Murphy: Machine learning A probabilistic perspective: Ch.9Murphy: Machine learning A probabilistic perspective: Ch.9
Murphy: Machine learning A probabilistic perspective: Ch.9Daisuke Yoneoka
 
Mesh Processing Course : Active Contours
Mesh Processing Course : Active ContoursMesh Processing Course : Active Contours
Mesh Processing Course : Active ContoursGabriel Peyré
 
On the Jensen-Shannon symmetrization of distances relying on abstract means
On the Jensen-Shannon symmetrization of distances relying on abstract meansOn the Jensen-Shannon symmetrization of distances relying on abstract means
On the Jensen-Shannon symmetrization of distances relying on abstract meansFrank Nielsen
 
Reinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsReinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsSean Meyn
 
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...BRNSS Publication Hub
 
MAPE regression, seminar @ QUT (Brisbane)
MAPE regression, seminar @ QUT (Brisbane)MAPE regression, seminar @ QUT (Brisbane)
MAPE regression, seminar @ QUT (Brisbane)Arnaud de Myttenaere
 
Introducing Zap Q-Learning
Introducing Zap Q-Learning   Introducing Zap Q-Learning
Introducing Zap Q-Learning Sean Meyn
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt msFaeco Bot
 
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...Gabriel Peyré
 
Roots equations
Roots equationsRoots equations
Roots equationsoscar
 
Runtime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary AlgorithmsRuntime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary AlgorithmsPK Lehre
 
Predicates and Quantifiers
Predicates and Quantifiers Predicates and Quantifiers
Predicates and Quantifiers Istiak Ahmed
 

What's hot (18)

QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Can we estimate a constant?
Can we estimate a constant?Can we estimate a constant?
Can we estimate a constant?
 
Murphy: Machine learning A probabilistic perspective: Ch.9
Murphy: Machine learning A probabilistic perspective: Ch.9Murphy: Machine learning A probabilistic perspective: Ch.9
Murphy: Machine learning A probabilistic perspective: Ch.9
 
Mesh Processing Course : Active Contours
Mesh Processing Course : Active ContoursMesh Processing Course : Active Contours
Mesh Processing Course : Active Contours
 
On the Jensen-Shannon symmetrization of distances relying on abstract means
On the Jensen-Shannon symmetrization of distances relying on abstract meansOn the Jensen-Shannon symmetrization of distances relying on abstract means
On the Jensen-Shannon symmetrization of distances relying on abstract means
 
CLIM Fall 2017 Course: Statistics for Climate Research, Statistics of Climate...
CLIM Fall 2017 Course: Statistics for Climate Research, Statistics of Climate...CLIM Fall 2017 Course: Statistics for Climate Research, Statistics of Climate...
CLIM Fall 2017 Course: Statistics for Climate Research, Statistics of Climate...
 
Reinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsReinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast Algorithms
 
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
 
MAPE regression, seminar @ QUT (Brisbane)
MAPE regression, seminar @ QUT (Brisbane)MAPE regression, seminar @ QUT (Brisbane)
MAPE regression, seminar @ QUT (Brisbane)
 
Introducing Zap Q-Learning
Introducing Zap Q-Learning   Introducing Zap Q-Learning
Introducing Zap Q-Learning
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
 
Roots equations
Roots equationsRoots equations
Roots equations
 
ABC in Venezia
ABC in VeneziaABC in Venezia
ABC in Venezia
 
Runtime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary AlgorithmsRuntime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary Algorithms
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Predicates and Quantifiers
Predicates and Quantifiers Predicates and Quantifiers
Predicates and Quantifiers
 
Statistical Method In Economics
Statistical Method In EconomicsStatistical Method In Economics
Statistical Method In Economics
 

Viewers also liked

Animales en peligro de extincion:YAGUARETE
Animales en peligro de extincion:YAGUARETEAnimales en peligro de extincion:YAGUARETE
Animales en peligro de extincion:YAGUARETEEugeeTyler
 
Visual Communication Presentation
Visual Communication PresentationVisual Communication Presentation
Visual Communication PresentationMoses Ifamose
 
Publicidad subliminal
Publicidad subliminalPublicidad subliminal
Publicidad subliminalpaulajblvt
 
Неуловимые Мстители
Неуловимые МстителиНеуловимые Мстители
Неуловимые Мстители093224445
 
Manual robotronic
Manual robotronicManual robotronic
Manual robotronicpablo9408
 
redes sociales
redes socialesredes sociales
redes socialesJean71321
 
Modelo pedagógico vanessa fda
Modelo pedagógico  vanessa fdaModelo pedagógico  vanessa fda
Modelo pedagógico vanessa fdaAlvarez Vanessa
 
Universidad estatal a distancia.pps
Universidad estatal a distancia.ppsUniversidad estatal a distancia.pps
Universidad estatal a distancia.ppsKarol1989
 
1606 probabilistic risk assessment in environmental toxicology
1606 probabilistic risk assessment in environmental toxicology1606 probabilistic risk assessment in environmental toxicology
1606 probabilistic risk assessment in environmental toxicologyDr Fereidoun Dejahang
 
A batea Julián, María, Andrés.
A batea Julián, María, Andrés.A batea Julián, María, Andrés.
A batea Julián, María, Andrés.vinhamedios
 
Limbajul corpului la summitul G20-2011
Limbajul corpului la summitul G20-2011Limbajul corpului la summitul G20-2011
Limbajul corpului la summitul G20-2011ion marin
 
Realidad virtual
Realidad virtualRealidad virtual
Realidad virtualhdzjavo4
 
1616 probability-the foundation of probability theory
1616 probability-the foundation of probability theory1616 probability-the foundation of probability theory
1616 probability-the foundation of probability theoryDr Fereidoun Dejahang
 
Proyecto educativo institucional vanesa
Proyecto educativo institucional vanesaProyecto educativo institucional vanesa
Proyecto educativo institucional vanesaAlvarez Vanessa
 

Viewers also liked (20)

Animales en peligro de extincion:YAGUARETE
Animales en peligro de extincion:YAGUARETEAnimales en peligro de extincion:YAGUARETE
Animales en peligro de extincion:YAGUARETE
 
1537 graphing lines
1537 graphing lines1537 graphing lines
1537 graphing lines
 
1633 the inverse z-transform
1633 the inverse z-transform1633 the inverse z-transform
1633 the inverse z-transform
 
Visual Communication Presentation
Visual Communication PresentationVisual Communication Presentation
Visual Communication Presentation
 
Publicidad subliminal
Publicidad subliminalPublicidad subliminal
Publicidad subliminal
 
Неуловимые Мстители
Неуловимые МстителиНеуловимые Мстители
Неуловимые Мстители
 
Manual robotronic
Manual robotronicManual robotronic
Manual robotronic
 
redes sociales
redes socialesredes sociales
redes sociales
 
Modelo pedagógico vanessa fda
Modelo pedagógico  vanessa fdaModelo pedagógico  vanessa fda
Modelo pedagógico vanessa fda
 
1571 mean
1571 mean1571 mean
1571 mean
 
Prezi - Power Point
Prezi - Power PointPrezi - Power Point
Prezi - Power Point
 
1533 game mathematics
1533 game mathematics1533 game mathematics
1533 game mathematics
 
Universidad estatal a distancia.pps
Universidad estatal a distancia.ppsUniversidad estatal a distancia.pps
Universidad estatal a distancia.pps
 
Redes
RedesRedes
Redes
 
1606 probabilistic risk assessment in environmental toxicology
1606 probabilistic risk assessment in environmental toxicology1606 probabilistic risk assessment in environmental toxicology
1606 probabilistic risk assessment in environmental toxicology
 
A batea Julián, María, Andrés.
A batea Julián, María, Andrés.A batea Julián, María, Andrés.
A batea Julián, María, Andrés.
 
Limbajul corpului la summitul G20-2011
Limbajul corpului la summitul G20-2011Limbajul corpului la summitul G20-2011
Limbajul corpului la summitul G20-2011
 
Realidad virtual
Realidad virtualRealidad virtual
Realidad virtual
 
1616 probability-the foundation of probability theory
1616 probability-the foundation of probability theory1616 probability-the foundation of probability theory
1616 probability-the foundation of probability theory
 
Proyecto educativo institucional vanesa
Proyecto educativo institucional vanesaProyecto educativo institucional vanesa
Proyecto educativo institucional vanesa
 

Similar to 1609 probability function p on subspace of s

Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheetSuvrat Mishra
 
Basic math including gradient
Basic math including gradientBasic math including gradient
Basic math including gradientRamesh Kesavan
 
Problem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodProblem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodPeter Herbert
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distancesChristian Robert
 
Cs229 notes9
Cs229 notes9Cs229 notes9
Cs229 notes9VuTran231
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distancesChristian Robert
 
Number Theory for Security
Number Theory for SecurityNumber Theory for Security
Number Theory for SecurityAbhijit Mondal
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionDexlab Analytics
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheetJoachim Gwoke
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualStéphane Canu
 
Lecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the DualLecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the DualStéphane Canu
 
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming ProblemHigher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Probleminventionjournals
 
Complex analysis notes
Complex analysis notesComplex analysis notes
Complex analysis notesPrakash Dabhi
 
Threshold network models
Threshold network modelsThreshold network models
Threshold network modelsNaoki Masuda
 

Similar to 1609 probability function p on subspace of s (20)

Probability Cheatsheet.pdf
Probability Cheatsheet.pdfProbability Cheatsheet.pdf
Probability Cheatsheet.pdf
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheet
 
Basic math including gradient
Basic math including gradientBasic math including gradient
Basic math including gradient
 
Problem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodProblem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element Method
 
kactl.pdf
kactl.pdfkactl.pdf
kactl.pdf
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distances
 
Cs229 notes9
Cs229 notes9Cs229 notes9
Cs229 notes9
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distances
 
cswiercz-general-presentation
cswiercz-general-presentationcswiercz-general-presentation
cswiercz-general-presentation
 
Number Theory for Security
Number Theory for SecurityNumber Theory for Security
Number Theory for Security
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling Distribution
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheet
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dual
 
Lecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the DualLecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the Dual
 
NODDEA2012_VANKOVA
NODDEA2012_VANKOVANODDEA2012_VANKOVA
NODDEA2012_VANKOVA
 
lecture6.ppt
lecture6.pptlecture6.ppt
lecture6.ppt
 
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming ProblemHigher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
 
Complex analysis notes
Complex analysis notesComplex analysis notes
Complex analysis notes
 
Threshold network models
Threshold network modelsThreshold network models
Threshold network models
 

More from Dr Fereidoun Dejahang

27 j20 my news punch -dr f dejahang 27-01-2020
27 j20 my news punch -dr f dejahang  27-01-202027 j20 my news punch -dr f dejahang  27-01-2020
27 j20 my news punch -dr f dejahang 27-01-2020Dr Fereidoun Dejahang
 
028 fast-tracking projects &amp; cost overrun
028 fast-tracking projects &amp; cost overrun028 fast-tracking projects &amp; cost overrun
028 fast-tracking projects &amp; cost overrunDr Fereidoun Dejahang
 
026 fast react-productivity improvement
026 fast react-productivity improvement026 fast react-productivity improvement
026 fast react-productivity improvementDr Fereidoun Dejahang
 
022 b construction productivity-write
022 b construction productivity-write022 b construction productivity-write
022 b construction productivity-writeDr Fereidoun Dejahang
 
016 communication in construction sector
016 communication in construction sector016 communication in construction sector
016 communication in construction sectorDr Fereidoun Dejahang
 
014 changes-cost overrun measurement
014 changes-cost overrun measurement014 changes-cost overrun measurement
014 changes-cost overrun measurementDr Fereidoun Dejahang
 
013 changes in construction projects
013 changes in construction projects013 changes in construction projects
013 changes in construction projectsDr Fereidoun Dejahang
 

More from Dr Fereidoun Dejahang (20)

27 j20 my news punch -dr f dejahang 27-01-2020
27 j20 my news punch -dr f dejahang  27-01-202027 j20 my news punch -dr f dejahang  27-01-2020
27 j20 my news punch -dr f dejahang 27-01-2020
 
28 dej my news punch rev 28-12-2019
28 dej my news punch rev 28-12-201928 dej my news punch rev 28-12-2019
28 dej my news punch rev 28-12-2019
 
16 fd my news punch rev 16-12-2019
16 fd my news punch rev 16-12-201916 fd my news punch rev 16-12-2019
16 fd my news punch rev 16-12-2019
 
029 fast-tracking projects
029 fast-tracking projects029 fast-tracking projects
029 fast-tracking projects
 
028 fast-tracking projects &amp; cost overrun
028 fast-tracking projects &amp; cost overrun028 fast-tracking projects &amp; cost overrun
028 fast-tracking projects &amp; cost overrun
 
027 fast-tracked projects-materials
027 fast-tracked projects-materials027 fast-tracked projects-materials
027 fast-tracked projects-materials
 
026 fast react-productivity improvement
026 fast react-productivity improvement026 fast react-productivity improvement
026 fast react-productivity improvement
 
025 enterprise resources management
025 enterprise resources management025 enterprise resources management
025 enterprise resources management
 
022 b construction productivity-write
022 b construction productivity-write022 b construction productivity-write
022 b construction productivity-write
 
022 a construction productivity (2)
022 a construction productivity (2)022 a construction productivity (2)
022 a construction productivity (2)
 
021 construction productivity (1)
021 construction productivity (1)021 construction productivity (1)
021 construction productivity (1)
 
019 competencies-managers
019 competencies-managers019 competencies-managers
019 competencies-managers
 
018 company productivity
018 company productivity018 company productivity
018 company productivity
 
017 communication
017 communication017 communication
017 communication
 
016 communication in construction sector
016 communication in construction sector016 communication in construction sector
016 communication in construction sector
 
015 changes-process model
015 changes-process model015 changes-process model
015 changes-process model
 
014 changes-cost overrun measurement
014 changes-cost overrun measurement014 changes-cost overrun measurement
014 changes-cost overrun measurement
 
013 changes in construction projects
013 changes in construction projects013 changes in construction projects
013 changes in construction projects
 
012 bussiness planning process
012 bussiness planning process012 bussiness planning process
012 bussiness planning process
 
011 business performance management
011 business performance management011 business performance management
011 business performance management
 

Recently uploaded

Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Celine George
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Educationpboyjonauth
 
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991RKavithamani
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesFatimaKhan178732
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingTechSoup
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 

Recently uploaded (20)

Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Education
 
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and Actinides
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 

1609 probability function p on subspace of s

  • 1. Probability function P on subspace of S 1. P(S)=1 2. For every event A in S, P(A)≥0 3. If A1, A2, .. are mutually exclusive, then P(∪Ai)=∑ P(Ai) A random variable X is a function that assigns a value to each outcome s in the sample space S (realizations of the random variable). Example: dart, bulls eye counts 50, i.e. X(s)=50, s = bulls eye location Example: measurement of a mass five times, yielding the true value m + random
  • 2. Probability Density Function (PDF), fX(x) The relative probability of realizations for a random variable P(X≤a)=∫ fX(x)dx ∫ fX(x)dx = 1 PDF for common functions: 1. Uniform random variable on [a,b]: fU(x)=1/(b-a), a≤x≤b; 0, x<a or x>b a -∞ -∞ -∞
  • 3. PDF for common functions: 2. Normal (Gaussian) function: fN(x)=1/σ√2π exp(-0.5(x-µ)2 /σ2 ) N(µ,σ2 ) normal distribution 3. Exponential function: fexp(x)=λ exp(-λx), x≥0; 0, x<0
  • 4. 4. Double-sided Exponential function: fdexp(x)=[1/(23/2 σ)] exp(-√2|x-µ|/σ) 5. χ2 function: f χ2 = 1/[2ν/2 Γ(ν/2)] x 0.5ν-1 exp(-x/2) with Γ(x)=∫ ξx-1 e-ξ dξ n independent variable with standard normal distributions, Z=∑Xi 2 is a χ2 random variable with ν=n degrees of freedom. ∞ 0
  • 5. 6 . Student's t distribution with ν degrees of freedom: ft(x)= Γ((ν+1)/2)]/Γ(ν/2) 1/√νπ (1+x2 /ν)-(ν+1)/2 with Γ(x)=∫ ξx-1 e-ξ dξ Approaches a standard normal distribution for large number of degrees of freedom Cumulative distribution functions (CDF): FX(a)=P(X≤a)=∫ fX(x)dx P(a≤X≤b) = ∫f(x)dx a -∞ b a
  • 6. Characterization of PDF’s: Expected value of random variable x: E[X] = ∫ xfX(x)dx, E[g(X)] = ∫ g(x)fX(x)dx Peak of distribution: XML (maximum likelihood) Variance: Var(X)=σX 2 = E[(X-µx)2 ] = E[X2 ]-µX 2 = ∫ (x-µx)2 fX(x)dx, σx=√Var(X) Variance measures the width of PDF’s. Wide PDF’s indicate noisy data, narrow indicate ∞ -∞ ∞ -∞ ∞ -∞
  • 7. Joint PDF’s: The joint PDF can quantify the probability that a set of random variables will take on a given value. f(X≤a,Y≤b)= ∫ ∫ f(x,y)dydx Expected value for joint PDF: E[g(X,Y)]=∫ ∫ g(x,y)f(x,y)dydx If X and Y are independent, f(x,y)=fX(x)fY(y) b -∞ a -∞ ∞ -∞ ∞ -∞
  • 8. Covariance of X and Y with Joint PDF: Cov(X,Y)=E[(X-E[X])(Y-E[Y])]=E[XY]-E[X]E[Y] X and Y independent, E[XY]=E[X]E[Y] and Cov(X,Y)=0 Covariance of a variable with itself = variance If Cov(X,Y)=0, X and Y are called uncorrelated Correlation of X and Y ρ(X,Y)=Cov(X,Y)/√Var(X) Var(Y) (correlation is a scaled covariance)
  • 9. Model covariance matrix cov(m): The general model mest =Md + v has covariance matrix M[cov d]MT (B.64) The least-squares solution mest =[GT G]-1 GT d has covariance matrix [cov m] = [GT G]-1 GT [cov d] ([GT G]-1 GT )T If data is uncorrelated and has equal variance σd 2 then [cov m] = σd 2 [GT G]-1
  • 10. More on Gaussian (Normal) distributions: Central limit theorem: Let X1, X2, …, Xn be independent and identically distributed random variables with finite expected value µ and variance σ2 . Let Zn=[X1+X2+..+Xn-nµ]/√nσ In the limit as n approaches infinity, Zn approaches the standard normal distribution Thus many summed variables in nature are normal distributions, thus LS solutions OK
  • 11. Means and confidence intervals: Given noisy measurements m1, m2, .., mn Estimate the true value m and the uncertainty of the estimate. Assume errors are independent and normally distributed with expected value 0 and some unknown standard deviation σ Compute average mave =[m1+m2+..+mn]/n s=[∑ (mi-mave )2 ]1/2 /n-1 n i=1
  • 12. Sampling Theorem: Independent, normally distributed measurements, with expected value m and standard deviated σ, the random quantity t = (m-mave )/s√n has a Student’s t distribution with n-1 degrees of freedom. If the true standard deviation σ is known, we are dealing with a standard normal distribution. The t distribution converges toward the normal distribution for large n
  • 13. Confidence intervals: Probability that one realization falls within a specified distance of the true mean Let tn-1,0.975 = 97.5 percentile of t distribution tn-1,0.025 = 2.5 percentile of t distribution P(tn-1,0.975 ≤m-mave /(s/√n) ≤tn-1,0.025)=0.95 P(tn-1,0.975 s/ √n ≤m-mave ≤tn-1,0.025 s/√n )=0.95 95% confidence interval Due to symmetry: mave -tn-1,0.975 s/ √n to mave +tn-1,0.975 s/ √n
  • 14. Confidence intervals related to the σ - PDF’s with large σ will have large CI, and vice versa Gaussian PDF, the 68% CI is 1σ wide and the 95% CI is 2σ wide If a particular Gaussian random variable has σ=1, and if a realization of that variable is 50, there is a 95% chance that the mean of that random variable lies between 48 and 52
  • 15. Example B.12 illustrates the case where σ is estimated If σ is known (rarely the case), we have a normal distribution. We can use the t-distribution with an infinite # of observations. E.g., 16 obs, m estimated to be 31.5. σ is known to be 5. Estimate 80% CI for m. mave -k ≤ m ≤ mave +k k=1.282 x 5/√16 = 1.6 31.5-1.6 ≤ m ≤ 31.5+1.6
  • 16. Statistical Aspects of LS: PDF for Normal Distribution: fi(di|m)=1/σι√2π exp(-(di-(Gm)i)2 /2σι 2 ) Maximum likelihood function L(m|d) is the product of all individual probability functions: L(m|d)=f1(d1|m)*f2(d2|m)* … * fm(dm|m) Idea: Maximize L(m|d) Maximize log{L(m|d)} Minimize - log{L(m|d)} Minimize -2 log{L(m|d)} min ∑ [di-(Gm)I]2 /σι 2 Aside from the 1/σι 2 factor, we have the LS solution
  • 17. min ∑ [di-(Gm)I]2 /σι 2 W=diag(1/σ1, 1/σ2, …,1/σm) Gw=WG dw=Wd Gwm=dw mL2=[Gw T Gw)-1 Gw T dw ||dw-Gwmw||2 2 = ∑ [di-(Gm)I]2 /σι 2 χobs 2 = ∑ [di-(Gm)I]2 /σι 2 χobs 2 has a χ2 distribution with m-n degrees of freedom
  • 18. The probability of finding a χ2 value as large or larger than the observed value is p=∫ fχ 2 (x)dx p-value test With correct model and independent error, the p-values will be uniformly distributed between 0 and 1. Near-0 or near-1 p- values indicates problems (incorrect model, underestimation of data errors, unlikely realization) ∞ χobs 2
  • 19. Multivariate normal distribution Random variables X1, …, Xn have a multivariate normal distribution, then the joint PDF is (B.61) f(x)=(2π)-n/2 (det[C])-1/2 exp[-(x-µ)T C-1 (x-µ)/2] Ci,j=Cov(Xi,Xj)
  • 20. Eigenvalues and eigenvectors Ax= λ x (A-λI)x=0 det (A-λI) = 0 The roots λ are the eigenvalues
  • 21. SVD G=USVT U (m x m) orthogonal spanning data space V (n x n) orthogonal spanning model space S (m x n) eigenvalues along diagonal Let rank(G)=p G=[Up U0] [Sp 0;0 0] [Vp V0]T = UpSpVp T G+ =VpSp -1 Up T = Generalized Inverse (pseudoinverse) m+ =G+ d=VpSp -1 Up T d = pseudoinverse solution Since Vp -1 =Vp T and Up -1 =Up T (A.6)
  • 22. SVD G=USVT U (m x m) orthogonal spanning data space V (n x n) orthogonal spanning model space S (m x n) eigenvalues along diagonal Rank(G)=p Theorem A.5: N(GT )+R(G)=Rm , i.e. p columns of Up form an orthonormal basis for R(G) columns of U0 form an orthonormal basis for N(GT ) p columns of Vp form an orthonormal basis for R(GT ) columns of V0 form an orthonormal basis for N(G)
  • 23. Properties of SVD G=USVT 1. N(G) and N(GT ) are trivial (only null vector): Up=U, Vp=V square orthogonal matrices, and Up T =Up -1 and Vp T =Vp -1 G+ =VpSp -1 Up T = (UpSpVp T )-1 = G-1 (inverse for full rank matrix, m=n=p). Unique solution, data are fit exactly.
  • 24. Properties of SVD G=USVT 2. N(G) is nontrivial (model, V); N(GT ) is trivial (data, U) Up T =Up -1 and Vp T Vp=Ip Gm+ =GG+ d=UpSpVp T VpSp -1 Up T d=UpSpIpSp -1 Up T d=d i.e. the data are fit exactly an LS solution, but nonuniquely due to the nontrivial model null space m=m+ +m0 =m+ +∑ αiV.,I ||m||2 2 =||m+ ||2 2 +∑ α2 i ≥||m+ ||2 2 -> minimum length solution n i=p+1 n i=p+1
  • 25. Properties of SVD G=USVT 3. N(G) is trivial (model, V); N(GT ) is nontrivial (data, U) Gm+ =GG+ d=UpSpVp T VpSp -1 Up T d=UpUp T d = projection of d onto R(G), I.e. the point in R(G) that is closest to d, m+ is LS solution to Gm=d. If d is in R(G), m+ will be solution to Gm=d. Solution is unique but does not fit data exactly i.e. the data are fit exactly an LS solution, but nonuniquely due to the nontrivial model null space n i=p+1 n i=p+1
  • 26. Properties of SVD G=USVT 4. N(G) is nontrivial (model, V); N(GT ) is nontrivial (data, U) p < (m,n) Gm+ =GG+ d=UpSpVp T VpSp -1 Up T d=UpUp T d = projection of d onto R(G), I.e. the point in R(G) that is closest to d, m+ is LS solution to Gm=d. i.e. LS solution to minimum norm, as case 2)
  • 27. Properties of SVD G=USVT - Always exists - LS or minimum length - Properly accommodates the rank and dimensions of G - Nontrivial model null space m0 the is heart of the problem -> - Infinite # of solutions will fit the data equally well, since components of N(G) have no effect on data fit, I.e., selection of a particular solution requires a priori constraints (smoothing, minimum length) - Nontrivial data space are vectors d0 that have no influence on m+ . If p<n, there are an infinite # of data sets that will produce the same model
  • 28. Properties of SVD - covariance/resolution Least squares solution is unbiased: min ∑ [di-(Gm)I]2 /σι 2 W=diag(1/σ1, 1/σ2, …,1/σm) Gw=WG, dw=Wd, Gwm=dw mL2=[Gw T Gw]-1 Gw T dw ||dw-Gwmw||2 2 = ∑ [di-(Gm)I]2 /σι 2 E[mL2]=E[(Gw T Gw)-1 Gw T dw] = (Gw T Gw)-1 Gw T E[dw]= (Gw T Gw)-1 Gw T dw true = (Gw T Gw)-1 Gw T Gwmtrue = mtrue
  • 29. Properties of SVD - covariance/resolution Generalized inverse not necessarily unbiased: E[m+ ]=E[G+ d] = G+ E[d] = G+ Gmtrue = Rmmtrue Bias= E[m+ ]-mtrue = Rmmtrue -mtrue = (Rm-I)mtrue = VpVp T -VVT mtrue =-V0V0 T mtrue I.e., as p increasees Rm->I Cov(mL2)=σ2 (GT G)-1 Cov(m+ )=G+ [Cov(d)]G+T = σ2 G+ G+T = σ2 VpSp -2 Vp T = σ2 ∑ V.,i V.,i T /σi 2 I.e., as p increases, Rm->I: P I=1
  • 30. Model resolution Rm -> I: increasing resolution Resolution test: multiply Rm onto a particular model, fx a spike model, with one element 1 and the rest 0, picks out the corresponding column of Rm Data resolution D+ =Gm+ = GG+ d = Rdd Rd=UpSpVp T VpSp -1 Up T =UpUp T p=m -> Rd=I, d+ =d p<m -> Rd<>I, m+ doesn’t fit data exactly
  • 31. Instabilitites of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small eigenvalues to stabilize solution -> Truncated SVD, TSVD Condition number cond(G)=s1/sk