SlideShare a Scribd company logo
1 of 23
Newton’s Method applied to a scalar function
Newton’s method for minimizing f(x):
Twice differentiable function f(x), initial solution x0. Generate a
sequence of solutions x1, x2, …and stop if the sequence
converges to a solution with f(x)=0.
1. Solve -f(xk) ≈ 2f(xk)x
2. Let xk+1=xk+x. 3. let k=k+1
Newton’s Method applied to LS
Not directly applicable to most nonlinear regression and
inverse problems (not equal # of model parameters and
data points, no exact solution to G(m)=d). Instead we will
use N.M. to minimize a nonlinear LS problem, e.g. fit a
vector of n parameters to a data vector d.
f(m)=∑ [(G(m)i-di)/i]2
Let fi(m)=(G(m)i-di)/i i=1,2,…,m, F(m)=[f1(m) … fm(m)]T
So that f(m)= ∑ fi(m)2 f(m)=∑ fi(m)2]
m
i=1
m
i=1
m
i=1
NM: Solve -f(mk) ≈ 2f(mk)m
LHS:f(mk)j = -∑ 2fi(mk)jF(m)j = -2 J(mk)F(mk)
RHS: 2f(mk)m = [2J(m)TJ(m)+Q(m)]m, where
Q(m)= 2 ∑ fi(m) fi(m)
-2 J(mk)F(mk) = 2 H(m) m
m = -H-1 J(mk)F(mk) =
-H-1 f(mk) (eq. 9.19)
H(m)= 2J(m)TJ(m)+Q(m)
fi(m)=(G(m)i-di)/i i=1,2,…,m,
F(m)=[f1(m) … fm(m)]T
Gauss-Newton (GN) method
2f(mk)m = H(m) m = [2J(mk)TJ(mk)+Q(m)]m
ignores Q(m)=2∑ fi(m) fi(m) :
f(m)≈2J(m)TJ(m), assuming fi(m) will be reasonably small
as we approach m*. That is,
Solve -f(xk) ≈ 2f(xk)x
f(m)j=∑ 2fi(m)jF(m)j, i.e.
J(mk)TJ(mk)m=-J(mk)TF(mk)
fi(m)=(G(m)i-di)/i i=1,2,…,m,
F(m)=[f1(m) … fm(m)]T
Newton’s Method applied to LS
Levenberg-Marquardt (LM) method uses
[J(mk)TJ(mk)+I]m=-J(mk)TF(mk)
->0 : GN
->large, steepest descent (SD) (down-gradient most rapidly).
SD provides slow but certain convergence.
Which value of  to use? Small values when GN is working
well, switch to larger values in problem areas. Start with
small value of , then adjust.
Statistics of iterative methods
Cov(Ad)=A Cov(d) AT (d has multivariate N.D.)
Cov(mL2)=(GTG)-1GT Cov(d) G(GTG)-1
Cov(d)=2I: Cov(mL2)=2(GTG)-1
However, we don’t have a linear relationship between data
and estimated model parameters for the nonlinear
regression, so cannot use these formulas. Instead:
F(m*+m)≈F(m*)+J(m*)m
Cov(m*)≈(J(m*)TJ(m*))-1 not exact due to linearization,
confidence intervals may not be accurate :)
ri=G(m*)i-di , i=1
s=[∑ ri
2/(m-n)]
Cov(m*)=s2(J(m*)TJ(m*))-1 establish confidence intervals, 2
Implementation Issues
1. Explicit (analytical) expressions for derivatives
2. Finite difference approximation for derivatives
3. When to stop iterating?
f(m)~0
||mk+1-mk||~0, |f(m)k+1-f(m)k|~0
eqs 9.47-9.49
4. Multistart method to optimize globally
Iterative Methods
SVD impractical when matrix has 1000’s of rows and columns
e.g., 256 x 256 cell tomography model, 100,000 ray paths, < 1% ray hits
50 Gb storage of system matrix, U ~ 80 Gb, V ~ 35 Gb
Waste of storage when system matrix is sparse
Iterative methods do not store the system matrix
Provide approximate solution, rather than exact using Gaussian elimination
Definition iterative method:
Starting point x0, do steps x1, x2, …
Hopefully converge toward right solution x
Iterative Methods
Kaczmarz’s algorithm:
Each of m rows of Gi.m = di define an n-dimensional hyperplane in Rm
1) Project initial m(0) solution onto hyperplane defined by first row of G
2) Project m(1) solution onto hyperplane defined by second row of G
3) … until projections have been done onto all hyperplanes
4) Start new cycle of projections until converge
If Gm=d has a unique solution, Kaczmarz’s algorithm will converge towards this
If several solutions, it will converge toward the solution closest to m(0)
If m(0)=0, we obtain the minimum length solution
If no exact solution, converge fails, bouncing around near approximate solution
If hyperplanes are nearly orthogonal, convergence is fast
If hyperplanes are nearly parallel, convergence is slow
Conjugate Gradients Method
Symmetric, positive definite system of equations Ax=b
min (x) = min(1/2 xTAx - bTx)
 (x) = Ax - b = 0 or Ax = b
CG method: construct basis p0, p1, … , pn-1 such that pi
TApj=0
when i≠j. Such basis is mutually conjugate w/r to A.
Only walk once in each direction and minimize
x = ∑ ipi - maximum of n steps required!
() = 1/2 [ ∑ ipi]T A∑ ipi - bT [∑ ipi ]
= 1/2 ∑ ∑ ijpi
TApj - bT [∑ ipi ] (summations 0->n-1)
n-1
i=0
Conjugate Gradients Method
pi
TApj=0 when i≠j. Such basis is mutually conjugate w/r to A.
(pi are said to be ‘A orthogonal’)
() = 1/2 ∑ ∑ ijpi
TApj - bT [∑ ipi ]
= 1/2 ∑ i
pi
TApi - bT [∑ ipi ]
= 1/2 (∑ i
pi
TApi - 2 ibT pi) - n independent terms
Thus min () by minimizing ith term i
pi
TApi - 2 ibT pi
->diff w/r i and set derivative to zero: i = bTpi / pi
TApi
i.e., IF we have mutually conjugate basis, it is easy to
minimize ()
Conjugate Gradients Method
CG constructs sequence of xi, ri=b-Axi, pi
Start: x0, r0=b, p0=r0, 0=r0
Tr0/p0
TAp0
Assume at kth iteration, we have x0, x1, …, xk; r0, r1, …, rk;
p0, p1, …, pk; 0, 1, …, k
Assume first k+1 basis vectors pi are mutually conjugate to A,
first k+1 ri are mutually orthogonal, and ri
Tpj=0 for i ≠j
Let xk+1= xk+kpk
and rk+1= rk-kApk which updates correctly, since:
rk+1 = b - Axk+1 = b - A(xk+kpk) = (b-Axk) - kApk = rk-kApk
Conjugate Gradients Method
Let xk+1= xk+kpk
and rk+1= rk-kApk
k+1 = rk+1
Trk+1/rk
Trk
pk+1= rk+1+k+1pk
bTpk=rk
Trk (eq 6.33-6.38)
Now we need proof of the assumptions
1) rk+1 is orthogonal to ri for i≤k (eq 6.39-6.43)
2) rk+1
Trk=0 (eq 6.44-6.48)
3) rk+1 is orthogonal to pi for i≤k (eq 6.49-6.54)
4) pk+1
TApi = 0 for i≤k (eq 6.55-6.60)
5) i=k: pk+1
TApk = 0 ie CG generates mutually conjugate basis
Conjugate Gradients Method
Thus shown that CG generates a sequence of mutually
conjugate basis vectors. In theory, the method will find an
exact solution in n iterations.
Given positive definite, symmetric system of eqs Ax=b, initial
solution x0, let 0=0, p-1=0,r0=b-Ax0, k=0
1. If k>0, let k = rk
Trk/rk-1
Trk-1
2. Let pk= rk+kpk-1
3. Let k = rk
Trk / pk
TApk
4. Let xk+1= xk+kpk
5. Let rk+1= rk-kApk
6. Let k=k+1
Conjugate Gradients Least Squares Method
CG can only be applied to positive definite systems of
equations, thus not applicable to general LS problems.
Instead, we can apply the CGLS method to
min ||Gm-d||2
GTGm=GTd
Conjugate Gradients Least Squares Method
GTGm=GTd
rk = GTd-GTGmk = GT(d-Gmk) = GTsk
sk+1=d-Gmk+1=d-G(mk+kpk)=(d-Gmk)-kGpk=sk-kGpk
Given a system of eqs Gm=d, k=0, m0=0, p-1=0, 0=0, r0=GTs0.
1. If k>0, let k = rk
Trk/[rk-1
Trk-1]
2. Let pk= rk+kpk-1
3. Let k = rk
Trk / [Gpk]T[G pk]
4. Let mk+1= mk+kpk
5. Let sk+1=sk-kGpk
6. Let rk+1= rk-kGpk
7. Let k=k+1; never computing GTG, only Gpk, GTsk+1
L1 Regression
LS (L2) is strongly affected by outliers
If outliers are due to incorrect measurements, the inversion
should minimize their effect on the estimated model.
Effects of outliers in LS is shown by rapid fall-off of the tails of
the Normal Distribution
In contrast the Exponential Distribution has a longer tail,
implying that the probability of realizing data far from the
mean is higher. A few data points several  from <d> is much
more probable if drawn from an exponential rather than from
a normal distribution. Therefore methods based on
exponential distributions are able to handle outliers better
than methods based on normal distributions. Such methods
are said to be robust.
L1 Regression
min ∑ [di -(Gm)i]/i = min ||dw-Gwm||1
thus more robust to outliers because error is not squared
Example: repeating measurement m times:
[1 1 … 1]T m =[d1 d2 … dm]T
mL2 = (GTG)-1GTd = m-1 ∑ di
f(m) = ||d-Gm||1 = ∑ |di-m|
Non-differentiable if m=di
Convex, so local minima=global minima
f’(m) = ∑ sgn(di-m), sgn(x)=+1 if x>0, =-1 if x<0, =0 if x=0
=0 if half is +, half is -
<d>est = median, where 1/2 of data is < <d>est, 1/2 > <d>est
L1 Regression
Finding min ||dw-Gwm||1 is not trivial. Several methods
available, such as IRLS, solving a series of LS problems
converging to a 1-norm:
r=d-Gm
f(m) = ||d-Gm||1 = ||r||1 = ∑ |ri|
non-differentiable if ri=0. At other points:
f’(m) = ∂f(m)/∂mk = - ∑ Gi,k sgn(ri) = -∑ Gi,k ri/|ri|
f(m) = -GTRr = -GTR(d-Gm)
Ri,i=1/|ri|
f(m) = -GTR(d-Gm) = 0
GTRGm = GTRd R depends on m, nonlinear system :(
IRLS!
convolution
S(t)=h(t)*f(t) = ∫h(t-k) f(k) dk = ∑ ht-k fk
h1 0 0 s1 assuming h(t) and f(t) are of length
h2 h1 0 s2 5 and 3, respectively
h3 h2 h1 f1 s3
h4 h3 h2 f2 = s4
h5 h4 h3 f3 s5
0 h5 h4 s6
0 0 h5 s7 Here, recursive solution is
easy
convolution
‘Shaping’ filtering: A*x=D, D ‘desired’ response, A,D known
a1 a2 a3
a0 a1 a2
a-1 a0 a1
.
The matrix ij is formed by the auto-correlation of at with zero-
lag values along the diagonal and auto-correlations of
successively higher lags off the diagonal. ij is symmetric of
a1 a0 a-1
a2 a1 a0
a3 a2 a1
.




convolution
ATD becomes
.
a-1 a-2 a-3
a0 a-1 a-2
a1 a0 a-1
.
The matrix cij is formed by the cross-correlation of the elements
of A and D. Solution: (ATA)-1ATD = -1c
.
d-1
d0
d1
.
.
c1
= c0
c-1
.
Example
Find a filter, 3 elements long, that convolved with (2,1)
produced (1,0,0,0): (2,1)*(f1,f2,f3)=(1,0,0,0)
a-1 a-2 a-3
a0 a-1 a-2
a1 a0 a-1
.
The matrix cij is formed by the cross-correlation of the elements
of A and D. Solution: (ATA)-1ATD = -1c
d-1
d0
d1
.
c1
= c0
c-1
.

More Related Content

Similar to lecture6.ppt

SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionLilyana Vankova
 
Ivan Dimitrijević "Nonlocal cosmology"
Ivan Dimitrijević "Nonlocal cosmology"Ivan Dimitrijević "Nonlocal cosmology"
Ivan Dimitrijević "Nonlocal cosmology"SEENET-MTP
 
Numerical Methods
Numerical MethodsNumerical Methods
Numerical MethodsTeja Ande
 
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法Computational Materials Science Initiative
 
Problem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodProblem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodPeter Herbert
 
Quantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averagesQuantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averagesVjekoslavKovac1
 
New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...Alexander Litvinenko
 
Natural and Clamped Cubic Splines
Natural and Clamped Cubic SplinesNatural and Clamped Cubic Splines
Natural and Clamped Cubic SplinesMark Brandao
 
Conjugate Gradient Methods
Conjugate Gradient MethodsConjugate Gradient Methods
Conjugate Gradient MethodsMTiti1
 
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...Transverse magnetic plane-wave scattering equations for infinite and semi-inf...
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...Yong Heui Cho
 
Trilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operatorsTrilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operatorsVjekoslavKovac1
 
Notes.on.popularity.versus.similarity.model
Notes.on.popularity.versus.similarity.modelNotes.on.popularity.versus.similarity.model
Notes.on.popularity.versus.similarity.modelsun peiyuan
 
Complex analysis notes
Complex analysis notesComplex analysis notes
Complex analysis notesPrakash Dabhi
 

Similar to lecture6.ppt (20)

SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last version
 
Ivan Dimitrijević "Nonlocal cosmology"
Ivan Dimitrijević "Nonlocal cosmology"Ivan Dimitrijević "Nonlocal cosmology"
Ivan Dimitrijević "Nonlocal cosmology"
 
Numerical Methods
Numerical MethodsNumerical Methods
Numerical Methods
 
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
 
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法
第5回CCMSハンズオン(ソフトウェア講習会): AkaiKKRチュートリアル 1. KKR法
 
Problem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element MethodProblem Solving by Computer Finite Element Method
Problem Solving by Computer Finite Element Method
 
Quantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averagesQuantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averages
 
stochastic processes assignment help
stochastic processes assignment helpstochastic processes assignment help
stochastic processes assignment help
 
Lecture5
Lecture5Lecture5
Lecture5
 
E33018021
E33018021E33018021
E33018021
 
The Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal FunctionThe Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal Function
 
New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...New data structures and algorithms for \\post-processing large data sets and ...
New data structures and algorithms for \\post-processing large data sets and ...
 
Natural and Clamped Cubic Splines
Natural and Clamped Cubic SplinesNatural and Clamped Cubic Splines
Natural and Clamped Cubic Splines
 
Conjugate Gradient Methods
Conjugate Gradient MethodsConjugate Gradient Methods
Conjugate Gradient Methods
 
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...Transverse magnetic plane-wave scattering equations for infinite and semi-inf...
Transverse magnetic plane-wave scattering equations for infinite and semi-inf...
 
Trilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operatorsTrilinear embedding for divergence-form operators
Trilinear embedding for divergence-form operators
 
Notes.on.popularity.versus.similarity.model
Notes.on.popularity.versus.similarity.modelNotes.on.popularity.versus.similarity.model
Notes.on.popularity.versus.similarity.model
 
Igv2008
Igv2008Igv2008
Igv2008
 
Ps02 cmth03 unit 1
Ps02 cmth03 unit 1Ps02 cmth03 unit 1
Ps02 cmth03 unit 1
 
Complex analysis notes
Complex analysis notesComplex analysis notes
Complex analysis notes
 

Recently uploaded

B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdf
B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdfB.tech Civil Engineering Major Project by Deepak Kumar ppt.pdf
B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdfDeepak15CivilEngg
 
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...Specialize in a MSc within Biomanufacturing, and work part-time as Process En...
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...Juli Boned
 
Prest Reed Portfolio revamp Full Sail Presentation 2
Prest Reed Portfolio revamp Full Sail Presentation 2Prest Reed Portfolio revamp Full Sail Presentation 2
Prest Reed Portfolio revamp Full Sail Presentation 25203records
 
Biography of Doctor Arif Patel Preston UK
Biography of Doctor Arif Patel Preston UKBiography of Doctor Arif Patel Preston UK
Biography of Doctor Arif Patel Preston UKArifPatel42
 
一比一原版赫尔大学毕业证如何办理
一比一原版赫尔大学毕业证如何办理一比一原版赫尔大学毕业证如何办理
一比一原版赫尔大学毕业证如何办理F
 
We’re looking for a junior patent engineer to join our Team!
We’re looking for a junior patent engineer to join our Team!We’re looking for a junior patent engineer to join our Team!
We’re looking for a junior patent engineer to join our Team!Juli Boned
 
Launch Your Research Career: A Beginner's Guide
Launch Your Research Career: A Beginner's GuideLaunch Your Research Career: A Beginner's Guide
Launch Your Research Career: A Beginner's GuideKaziFaisalAlam
 
B.tech civil major project by Deepak Kumar
B.tech civil major project by Deepak KumarB.tech civil major project by Deepak Kumar
B.tech civil major project by Deepak KumarDeepak15CivilEngg
 
Fracture design PowerPoint presentations
Fracture design PowerPoint presentationsFracture design PowerPoint presentations
Fracture design PowerPoint presentationsjitiniift
 
UIowa Application Instructions - 2024 Update
UIowa Application Instructions - 2024 UpdateUIowa Application Instructions - 2024 Update
UIowa Application Instructions - 2024 UpdateUniversity of Iowa
 
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证epyhpep
 
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdf
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdfOchsen Screenplay Coverage - JACOB - 10.16.23.pdf
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdfRachel Ochsenschlager
 
Crafting an effective CV for AYUSH Doctors.pdf
Crafting an effective CV for AYUSH Doctors.pdfCrafting an effective CV for AYUSH Doctors.pdf
Crafting an effective CV for AYUSH Doctors.pdfShri Dr Arul Selvan
 
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...manju garg
 
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证eqaqen
 
Chennai (Chennai) Independent Escorts - 9632533318 100% genuine
Chennai (Chennai) Independent Escorts - 9632533318 100% genuineChennai (Chennai) Independent Escorts - 9632533318 100% genuine
Chennai (Chennai) Independent Escorts - 9632533318 100% genuineruksarkahn825
 
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证gakamzu
 
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证eqaqen
 
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...yulianti213969
 

Recently uploaded (20)

B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdf
B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdfB.tech Civil Engineering Major Project by Deepak Kumar ppt.pdf
B.tech Civil Engineering Major Project by Deepak Kumar ppt.pdf
 
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...Specialize in a MSc within Biomanufacturing, and work part-time as Process En...
Specialize in a MSc within Biomanufacturing, and work part-time as Process En...
 
Prest Reed Portfolio revamp Full Sail Presentation 2
Prest Reed Portfolio revamp Full Sail Presentation 2Prest Reed Portfolio revamp Full Sail Presentation 2
Prest Reed Portfolio revamp Full Sail Presentation 2
 
Biography of Doctor Arif Patel Preston UK
Biography of Doctor Arif Patel Preston UKBiography of Doctor Arif Patel Preston UK
Biography of Doctor Arif Patel Preston UK
 
一比一原版赫尔大学毕业证如何办理
一比一原版赫尔大学毕业证如何办理一比一原版赫尔大学毕业证如何办理
一比一原版赫尔大学毕业证如何办理
 
We’re looking for a junior patent engineer to join our Team!
We’re looking for a junior patent engineer to join our Team!We’re looking for a junior patent engineer to join our Team!
We’re looking for a junior patent engineer to join our Team!
 
Launch Your Research Career: A Beginner's Guide
Launch Your Research Career: A Beginner's GuideLaunch Your Research Career: A Beginner's Guide
Launch Your Research Career: A Beginner's Guide
 
B.tech civil major project by Deepak Kumar
B.tech civil major project by Deepak KumarB.tech civil major project by Deepak Kumar
B.tech civil major project by Deepak Kumar
 
Fracture design PowerPoint presentations
Fracture design PowerPoint presentationsFracture design PowerPoint presentations
Fracture design PowerPoint presentations
 
UIowa Application Instructions - 2024 Update
UIowa Application Instructions - 2024 UpdateUIowa Application Instructions - 2024 Update
UIowa Application Instructions - 2024 Update
 
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(Columbia毕业证书)哥伦比亚大学毕业证成绩单本科硕士学位证留信学历认证
 
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdf
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdfOchsen Screenplay Coverage - JACOB - 10.16.23.pdf
Ochsen Screenplay Coverage - JACOB - 10.16.23.pdf
 
Crafting an effective CV for AYUSH Doctors.pdf
Crafting an effective CV for AYUSH Doctors.pdfCrafting an effective CV for AYUSH Doctors.pdf
Crafting an effective CV for AYUSH Doctors.pdf
 
Transferable Skills - Roadmaps & Resources - Dirk Spencer
Transferable Skills - Roadmaps & Resources - Dirk SpencerTransferable Skills - Roadmaps & Resources - Dirk Spencer
Transferable Skills - Roadmaps & Resources - Dirk Spencer
 
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...
❤️Mangalore Call Girls Service ❤️🍑 6378878445 👄🫦Independent Escort Service Ch...
 
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证
一比一定(购)南昆士兰大学毕业证(USQ毕业证)成绩单学位证
 
Chennai (Chennai) Independent Escorts - 9632533318 100% genuine
Chennai (Chennai) Independent Escorts - 9632533318 100% genuineChennai (Chennai) Independent Escorts - 9632533318 100% genuine
Chennai (Chennai) Independent Escorts - 9632533318 100% genuine
 
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证
如何办理(NEU毕业证书)东北大学毕业证成绩单本科硕士学位证留信学历认证
 
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证
一比一定(购)堪培拉大学毕业证(UC毕业证)成绩单学位证
 
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...
obat aborsi pacitan wa 081336238223 jual obat aborsi cytotec asli di pacitan0...
 

lecture6.ppt

  • 1. Newton’s Method applied to a scalar function Newton’s method for minimizing f(x): Twice differentiable function f(x), initial solution x0. Generate a sequence of solutions x1, x2, …and stop if the sequence converges to a solution with f(x)=0. 1. Solve -f(xk) ≈ 2f(xk)x 2. Let xk+1=xk+x. 3. let k=k+1
  • 2. Newton’s Method applied to LS Not directly applicable to most nonlinear regression and inverse problems (not equal # of model parameters and data points, no exact solution to G(m)=d). Instead we will use N.M. to minimize a nonlinear LS problem, e.g. fit a vector of n parameters to a data vector d. f(m)=∑ [(G(m)i-di)/i]2 Let fi(m)=(G(m)i-di)/i i=1,2,…,m, F(m)=[f1(m) … fm(m)]T So that f(m)= ∑ fi(m)2 f(m)=∑ fi(m)2] m i=1 m i=1 m i=1
  • 3. NM: Solve -f(mk) ≈ 2f(mk)m LHS:f(mk)j = -∑ 2fi(mk)jF(m)j = -2 J(mk)F(mk) RHS: 2f(mk)m = [2J(m)TJ(m)+Q(m)]m, where Q(m)= 2 ∑ fi(m) fi(m) -2 J(mk)F(mk) = 2 H(m) m m = -H-1 J(mk)F(mk) = -H-1 f(mk) (eq. 9.19) H(m)= 2J(m)TJ(m)+Q(m) fi(m)=(G(m)i-di)/i i=1,2,…,m, F(m)=[f1(m) … fm(m)]T
  • 4. Gauss-Newton (GN) method 2f(mk)m = H(m) m = [2J(mk)TJ(mk)+Q(m)]m ignores Q(m)=2∑ fi(m) fi(m) : f(m)≈2J(m)TJ(m), assuming fi(m) will be reasonably small as we approach m*. That is, Solve -f(xk) ≈ 2f(xk)x f(m)j=∑ 2fi(m)jF(m)j, i.e. J(mk)TJ(mk)m=-J(mk)TF(mk) fi(m)=(G(m)i-di)/i i=1,2,…,m, F(m)=[f1(m) … fm(m)]T
  • 5. Newton’s Method applied to LS Levenberg-Marquardt (LM) method uses [J(mk)TJ(mk)+I]m=-J(mk)TF(mk) ->0 : GN ->large, steepest descent (SD) (down-gradient most rapidly). SD provides slow but certain convergence. Which value of  to use? Small values when GN is working well, switch to larger values in problem areas. Start with small value of , then adjust.
  • 6. Statistics of iterative methods Cov(Ad)=A Cov(d) AT (d has multivariate N.D.) Cov(mL2)=(GTG)-1GT Cov(d) G(GTG)-1 Cov(d)=2I: Cov(mL2)=2(GTG)-1 However, we don’t have a linear relationship between data and estimated model parameters for the nonlinear regression, so cannot use these formulas. Instead: F(m*+m)≈F(m*)+J(m*)m Cov(m*)≈(J(m*)TJ(m*))-1 not exact due to linearization, confidence intervals may not be accurate :) ri=G(m*)i-di , i=1 s=[∑ ri 2/(m-n)] Cov(m*)=s2(J(m*)TJ(m*))-1 establish confidence intervals, 2
  • 7. Implementation Issues 1. Explicit (analytical) expressions for derivatives 2. Finite difference approximation for derivatives 3. When to stop iterating? f(m)~0 ||mk+1-mk||~0, |f(m)k+1-f(m)k|~0 eqs 9.47-9.49 4. Multistart method to optimize globally
  • 8. Iterative Methods SVD impractical when matrix has 1000’s of rows and columns e.g., 256 x 256 cell tomography model, 100,000 ray paths, < 1% ray hits 50 Gb storage of system matrix, U ~ 80 Gb, V ~ 35 Gb Waste of storage when system matrix is sparse Iterative methods do not store the system matrix Provide approximate solution, rather than exact using Gaussian elimination Definition iterative method: Starting point x0, do steps x1, x2, … Hopefully converge toward right solution x
  • 9. Iterative Methods Kaczmarz’s algorithm: Each of m rows of Gi.m = di define an n-dimensional hyperplane in Rm 1) Project initial m(0) solution onto hyperplane defined by first row of G 2) Project m(1) solution onto hyperplane defined by second row of G 3) … until projections have been done onto all hyperplanes 4) Start new cycle of projections until converge If Gm=d has a unique solution, Kaczmarz’s algorithm will converge towards this If several solutions, it will converge toward the solution closest to m(0) If m(0)=0, we obtain the minimum length solution If no exact solution, converge fails, bouncing around near approximate solution If hyperplanes are nearly orthogonal, convergence is fast If hyperplanes are nearly parallel, convergence is slow
  • 10. Conjugate Gradients Method Symmetric, positive definite system of equations Ax=b min (x) = min(1/2 xTAx - bTx)  (x) = Ax - b = 0 or Ax = b CG method: construct basis p0, p1, … , pn-1 such that pi TApj=0 when i≠j. Such basis is mutually conjugate w/r to A. Only walk once in each direction and minimize x = ∑ ipi - maximum of n steps required! () = 1/2 [ ∑ ipi]T A∑ ipi - bT [∑ ipi ] = 1/2 ∑ ∑ ijpi TApj - bT [∑ ipi ] (summations 0->n-1) n-1 i=0
  • 11. Conjugate Gradients Method pi TApj=0 when i≠j. Such basis is mutually conjugate w/r to A. (pi are said to be ‘A orthogonal’) () = 1/2 ∑ ∑ ijpi TApj - bT [∑ ipi ] = 1/2 ∑ i pi TApi - bT [∑ ipi ] = 1/2 (∑ i pi TApi - 2 ibT pi) - n independent terms Thus min () by minimizing ith term i pi TApi - 2 ibT pi ->diff w/r i and set derivative to zero: i = bTpi / pi TApi i.e., IF we have mutually conjugate basis, it is easy to minimize ()
  • 12. Conjugate Gradients Method CG constructs sequence of xi, ri=b-Axi, pi Start: x0, r0=b, p0=r0, 0=r0 Tr0/p0 TAp0 Assume at kth iteration, we have x0, x1, …, xk; r0, r1, …, rk; p0, p1, …, pk; 0, 1, …, k Assume first k+1 basis vectors pi are mutually conjugate to A, first k+1 ri are mutually orthogonal, and ri Tpj=0 for i ≠j Let xk+1= xk+kpk and rk+1= rk-kApk which updates correctly, since: rk+1 = b - Axk+1 = b - A(xk+kpk) = (b-Axk) - kApk = rk-kApk
  • 13. Conjugate Gradients Method Let xk+1= xk+kpk and rk+1= rk-kApk k+1 = rk+1 Trk+1/rk Trk pk+1= rk+1+k+1pk bTpk=rk Trk (eq 6.33-6.38) Now we need proof of the assumptions 1) rk+1 is orthogonal to ri for i≤k (eq 6.39-6.43) 2) rk+1 Trk=0 (eq 6.44-6.48) 3) rk+1 is orthogonal to pi for i≤k (eq 6.49-6.54) 4) pk+1 TApi = 0 for i≤k (eq 6.55-6.60) 5) i=k: pk+1 TApk = 0 ie CG generates mutually conjugate basis
  • 14. Conjugate Gradients Method Thus shown that CG generates a sequence of mutually conjugate basis vectors. In theory, the method will find an exact solution in n iterations. Given positive definite, symmetric system of eqs Ax=b, initial solution x0, let 0=0, p-1=0,r0=b-Ax0, k=0 1. If k>0, let k = rk Trk/rk-1 Trk-1 2. Let pk= rk+kpk-1 3. Let k = rk Trk / pk TApk 4. Let xk+1= xk+kpk 5. Let rk+1= rk-kApk 6. Let k=k+1
  • 15. Conjugate Gradients Least Squares Method CG can only be applied to positive definite systems of equations, thus not applicable to general LS problems. Instead, we can apply the CGLS method to min ||Gm-d||2 GTGm=GTd
  • 16. Conjugate Gradients Least Squares Method GTGm=GTd rk = GTd-GTGmk = GT(d-Gmk) = GTsk sk+1=d-Gmk+1=d-G(mk+kpk)=(d-Gmk)-kGpk=sk-kGpk Given a system of eqs Gm=d, k=0, m0=0, p-1=0, 0=0, r0=GTs0. 1. If k>0, let k = rk Trk/[rk-1 Trk-1] 2. Let pk= rk+kpk-1 3. Let k = rk Trk / [Gpk]T[G pk] 4. Let mk+1= mk+kpk 5. Let sk+1=sk-kGpk 6. Let rk+1= rk-kGpk 7. Let k=k+1; never computing GTG, only Gpk, GTsk+1
  • 17. L1 Regression LS (L2) is strongly affected by outliers If outliers are due to incorrect measurements, the inversion should minimize their effect on the estimated model. Effects of outliers in LS is shown by rapid fall-off of the tails of the Normal Distribution In contrast the Exponential Distribution has a longer tail, implying that the probability of realizing data far from the mean is higher. A few data points several  from <d> is much more probable if drawn from an exponential rather than from a normal distribution. Therefore methods based on exponential distributions are able to handle outliers better than methods based on normal distributions. Such methods are said to be robust.
  • 18. L1 Regression min ∑ [di -(Gm)i]/i = min ||dw-Gwm||1 thus more robust to outliers because error is not squared Example: repeating measurement m times: [1 1 … 1]T m =[d1 d2 … dm]T mL2 = (GTG)-1GTd = m-1 ∑ di f(m) = ||d-Gm||1 = ∑ |di-m| Non-differentiable if m=di Convex, so local minima=global minima f’(m) = ∑ sgn(di-m), sgn(x)=+1 if x>0, =-1 if x<0, =0 if x=0 =0 if half is +, half is - <d>est = median, where 1/2 of data is < <d>est, 1/2 > <d>est
  • 19. L1 Regression Finding min ||dw-Gwm||1 is not trivial. Several methods available, such as IRLS, solving a series of LS problems converging to a 1-norm: r=d-Gm f(m) = ||d-Gm||1 = ||r||1 = ∑ |ri| non-differentiable if ri=0. At other points: f’(m) = ∂f(m)/∂mk = - ∑ Gi,k sgn(ri) = -∑ Gi,k ri/|ri| f(m) = -GTRr = -GTR(d-Gm) Ri,i=1/|ri| f(m) = -GTR(d-Gm) = 0 GTRGm = GTRd R depends on m, nonlinear system :( IRLS!
  • 20. convolution S(t)=h(t)*f(t) = ∫h(t-k) f(k) dk = ∑ ht-k fk h1 0 0 s1 assuming h(t) and f(t) are of length h2 h1 0 s2 5 and 3, respectively h3 h2 h1 f1 s3 h4 h3 h2 f2 = s4 h5 h4 h3 f3 s5 0 h5 h4 s6 0 0 h5 s7 Here, recursive solution is easy
  • 21. convolution ‘Shaping’ filtering: A*x=D, D ‘desired’ response, A,D known a1 a2 a3 a0 a1 a2 a-1 a0 a1 . The matrix ij is formed by the auto-correlation of at with zero- lag values along the diagonal and auto-correlations of successively higher lags off the diagonal. ij is symmetric of a1 a0 a-1 a2 a1 a0 a3 a2 a1 .    
  • 22. convolution ATD becomes . a-1 a-2 a-3 a0 a-1 a-2 a1 a0 a-1 . The matrix cij is formed by the cross-correlation of the elements of A and D. Solution: (ATA)-1ATD = -1c . d-1 d0 d1 . . c1 = c0 c-1 .
  • 23. Example Find a filter, 3 elements long, that convolved with (2,1) produced (1,0,0,0): (2,1)*(f1,f2,f3)=(1,0,0,0) a-1 a-2 a-3 a0 a-1 a-2 a1 a0 a-1 . The matrix cij is formed by the cross-correlation of the elements of A and D. Solution: (ATA)-1ATD = -1c d-1 d0 d1 . c1 = c0 c-1 .