SlideShare a Scribd company logo
1 of 50
Download to read offline
Day 4: Stochastic Dynamic Programming
Day 4 NotesHowitt and Msangi 1
 Understand Bellman’s Principle of Optimality
and the basic Stochastic Dynamic
programming problem
 Solve the SDP with value function iteration
 Apply the concepts of models to agro-
forestry and livestock herd dynamics
 Make changes to the SDP and simulate the
corresponding change in optimal solution
Day 4 NotesHowitt and Msangi 2
 Re-cap on rangeland stocking model….
 Introduction to Stochastic Dynamic Programming
◦ Extend DP framework to include stochastic state variables
and apply to herd and agro forestry management
 Stochastic Cake Eating
 Multi-State Models
◦ Function Approximation
 Agro-Forestry Application
◦ Input Data and State Space
◦ Simulation
 Herd Dynamics Application
◦ Input Data
◦ Simulation
Day 4 NotesHowitt and Msangi 3
An Application to Reservoir
Management
Day 3 NotesHowitt and Msangi 4
 “Estimating Intertemporal Preferences for
Resource Allocation” AJAE, 87(4): 969-983.
(Howitt RE, S Msangi, A Reynaud, KC Knapp)
 What started out as a calibration exercise –
ended up as a research project (with some
interesting research discoveries)
Day 3 NotesHowitt and Msangi 5
 Many of the Important Policy Questions in
Natural Resource Management Revolve
Around How to Deal with Uncertainty over
Time (Global Climate Change, Extreme
Weather Events, Invasive Species
Encroachment, Disease Outbreak, etc. )
 Policy Makers look to Economic Models to
Provide them with Guidance on Best
Management Practices
 Economic Policy Models Have Typically
Downplayed the Role of Risk in the
Preferences of the Decision-maker
 Few Studies Have Ever Tried to Measure the
Degree to Which Risk Aversion Matters in
Resource Management Problems
 Time-Additive Separability in Dynamic
Models Imposes Severe Constraints on
Intertemporal Preferences
 In order to Address this Gap in the Natural
Resources literature….
 We Applied Dynamic Estimation Methods to
an Example of Reservoir Management
 We Relaxed the Assumption of Time-Additive
Separability of the Decision-Maker’s Utility
 We Tested with Alternative Utility Forms to
Determine the Importance of Risk Aversion
 Koopmans (1960) laid the foundation for
eliminating deficiencies of TAS with
recursive preferences.
 Recursive Utility is a class of functionals
designed to offer a generality to time
preferences while still maintaining time
consistency in behavior.
 Allows for the potential smoothing of
consumption by allowing complementarity
between time periods.
( )W 
( )1( ) ( ), ( )U W u c U S=c c
States the weak separability of the future from present
where
is an aggregator function
For TAS, the aggregator is simply
( )( ), ( )W u c x u c xβ= +
( ) ( )
1
( ), 1 ( )W u c x u c x
ρρ ρ
β β = − ⋅ + ⋅ 
1
( )
1
EIS σ
ρ
=
−
So we choose our aggregator to be
and the implied elasticity (“resistance”) to inter-temporal substitution
is given by
where ( ),0 (0,1]ρ ∈ −∞ ∪
Time Additive Separable Utility
Using Bellman’s recursive relationship:
{ }
{ }
{ }
{ }
1
2
1 2
1
1 1 2
2 2 3
2 3
1 2 3
, ,
1
1, 2
( ) max ( ) ( )
( ) max ( ) ( )
( ) max ( ) ( )
:
( ) max ( ) ( ) ( ) ( )
( )
( )
t
t
t
t t t
t t t
c
t t t
c
t t t
c
t t t t t
c c c
t
t t
V x U c V x
V x U c V x
V x U c V x
Substituting and simplifying
V x U c U c U c V x
u c
Note that MRS c
β
β
β
β β β
β
+
+
+ +
+
+ + +
+ + +
+ + +
+
+ +
+
+
+
= + + +
′
=
2( )tu c +
′
Iso-Elastic Recursive Utility
A utility function with a CES across time periods.
1
2
1
1
1
1
1 1 2
1
2 2 3
( 1) 1
,
( ) max (1 ) ( ) ( )
( ) max (1 ) ( ) ( )
( ) max (1 ) ( ) ( )
:
( ) max (1 ) ( ) (1 ) (
t
t
t
t t
t t t
c
t t t
c
t t t
c
t At t t t
c c
V x U c V x
V x U c V x
V x U c V x
Substituting and simplifying
V x U c U c
ρ ρ ρ
ρ ρ ρ
ρ ρ ρ
ρ
β β
β β
β β
β β β
+
+
+
+
+ + +
+ + +
+ +
 = − + 
 = − + 
 = − + 
= − + −
1
1 2
1 1
2
2
1
2 3
( 2) 1 2 3
, ,
1 2
1 1 2
1
2
2
) ( )
( ) max (1 ) ( ) (1 ) ( ) (1 ) ( ) ( )
( )
(1 ) ( ) (1 ) ( ) (1 ) ( ) ( )
( )
(1 )
t t t
t
t At t t t t t
c c c
t
t t t t
t
t
t
V x
V x U c U c U c V x
V x
u c U c U c V x
c
V x
c
ρ
ρ
ρ ρ
ρ ρ ρ ρ ρ
ρ ρ ρ ρ
β
β β β β β β
β β β β β β
β β
+ +
−
+
+ + + +
−
+ + +
+
+
 + 
 = − + − + − + 
∂
′  = − − + − + ∂
∂
′= −
∂
1 1
1 2 3
2 1 2 3( ) (1 ) ( ) (1 ) ( ) (1 ) ( ) ( )t t t t tu c U c U c U c V x
ρ
ρ ρ ρ ρ ρ
β β β β β β
−
−
+ + + +
 − + − + − + 
With Recursive Utility All Periods Enter into
MRS
1
1 1
2
1 1 2
1, 2 2 3
2 1 2 3
( ) (1 ) ( ) (1 ) ( ) ( )1
( ) (1 ) ( ) (1 ) ( ) (1 ) ( ) ( )
t t t t
t t
t t t t t
u c U c U c V x
MRS
u c U c U c U c V x
ρ ρ ρ ρ ρ
ρ ρ ρ ρ
β β β β
β β β β β β β
− −
+ + +
+ +
+ + + +
   ′ − + − +
=    ′ − + − + − +   
In micro-economics we have an appreciation of the difference
between linear and CES utility in static consumer theory
The same intuition applies here in a dynamic context….
 The previous equations show that the marginal rate
of substitution across time is path dependent.
 Timing is now an explicit economic control variable
 We no longer assume that “The marginal rate of
substitution between lunch and dinner is
independent of the amount of breakfast” (Henry
Wan).
 A smaller elasticity of intertemporal substitution
flattens out the optimal time path of resource use-
yielding a time consistent sustainable result.
 Stochastic Equations of Motion link Stocks
and Flows
 Randomness in the equations of motion or
exogenous random shocks change the
system evolution
 The current state and future distributions
are usually known to decision makers
 Management decisions inherently optimize
a stochastic dynamic path of resource use
and consequently maximize dynamic
stochastic utility
A Simple Resource Network with a
Single State Variable
te1
~
te2
~
Demand
tS
tw
( ) ( )2 1
1
1Max (1 ) E ( ) Et e t t e t
w
U W q U
ρ ρ
ρ α α
β β +
 
 = − ⋅ +  
 
.








≥
≤
≥
+=
−+=
+
+
+
0
~
~
1
1
2
11
t
t
t
ttt
tttt
w
SS
SS
ewq
weSS
The Optimization
Problem
( )
1
2 2 1 1
0
, Max (1 ) ( )d ( , )dt /
w
V S e W w e Φ V S e Φ
ρ ρ
ρ α α
β β
≥
 
   = − ⋅ + +   
  
 
∫ ∫ 
Which can be re-stated in
terms of Bellman’s Recurrence
Relationship…
..and which we solve by numerically with
Continuous-valued State and Control
Variables
Solving for the Expected Value Function
Initialize with
Intermediate Value Function W(Xt , ut )
Nodes for: State Evaluation and Stochastic Inflow values
Probabilities of Inflow over k Stochastic node values
Define the State Variable on
[ -1, 1] Interval for each polynomial node j
Value Iteration Loop (n steps)
n = n+1
Error =
If Error > 0.1 e-7
Stop
Value Function Chebychev Polynomial Coefficients
jXaVpuXWPVNB
k i
kj
ti
n
iktt
n
j ∀




















+= ∑ ∑ +
−
)(),(max ,
1
1

φβ
( )∑ −
−
i
n
i
n
i aa
21
( )
( ) ( )∑
∑
++
+
=
j
nj
ti
nj
ti
j
nj
ti
n
j
n
i
XX
XPVNB
a ,
1
,
1
,
1


φφ
φ
( )j
x

( ) 32
0067.045.0150 qqqqW ⋅+⋅−⋅=
ttt
ttttt
capee
eecapesp
⋅⋅⋅+
⋅+⋅=
1
3
1
2
111
~0.02305-~0.000993
~0.005024~0.095382),~(
Current Profit Function
Spill Function
Net Benefit Function for Water
0
1
2
3
4
5
6
0 4 8 12 16 20 24 28 32 36 40 44 48 52 56
q, (MAF)
W(q),1000M$US
 We employ a nested procedure to solve the
SDP problem with value iteration, while we
systematically change the parameter values
of the objective function to maximize a
likelihood function.
 We employ a derivative-free ( Nelder Meade)
search algorithm to implement the ‘hill-
climbing’ procedure that searches for the
likelihood-maximizing values of preference
parameters
EIS value
( )1 1 ρ−
Coeff. of Risk
Aversion
1 α−
ρ
α
These parameters were calculated with a fixed discount rate of β
Parameter
Estimated
Value
Standard Error
-9.000 4.60 0.100
-0.440 0.23 1.440
Log
Likelihood
-10.257
=0.95.
Standard errors are based on 500 bootstrap repetitions
1,set estimateα ρ=
For Risk-Neutral Recursive model (RNR)
For Risk-Neutral (non-Recursive) model (RN)
For Non-Recursive model (with Risk) use CRRA
1set ρ α= =
( )0.95fix β =
( )0.95fix β =
)(
1 1
)1(
+
−
+
−
= t
t
t UE
W
U βα
α
,estimate α β
0
0.5
1
1.5
2
2.5
3
3.5
1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
MillionAcrefeet
RN
ACTUAL
0
0.5
1
1.5
2
2.5
3
3.5
1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
MillionAcrefeet
CRRA
ACTUAL
0
0.5
1
1.5
2
2.5
3
3.5
1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
MillionAcrefeet
REC
ACTUAL
0
1
2
3
4
5
6
7
8
9
1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
MillionAcrefeet
RN
ACTUAL
0
1
2
3
4
5
6
7
8
9
1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
MillionAcrefeet
REC
CRRA
ACTUAL
 Clearly a non-recursive model that ignores
risk fares the worst, when compared to actual
storage and releases
 Adding risk, but not recursivity of
preferences, gets you closer to actual
values…but not quite….
 A Recursive Specification outperforms both of
these, with or without risk aversion
 Estimation of the Fully-Recursive model is
robust to Discount Values and the
Parameter Estimates appear to be Stationary
over the Study Period
 Once we allow Intertemporal Preferences to
be recursive, the role of Risk in explaining
Resource Management Behavior is Reduced
 Imposing Time-Additive Separability on
Dynamic Models may have more severe
implications for behavior than most
researchers realize…..
Day 3 NotesHowitt and Msangi 35
 Extend DP framework to include stochastic
state variables in the model
 Apply the new framework to herd dynamics
and agro-forestry management
 Return to cake eating example
Day 4 NotesHowitt and Msangi 36
 Stochastic Cake Eating
◦ What if I want cake today, but not tomorrow?
 Cake Eating Example:
 CakeEatingDP_ChebyAprox_Stochastic_Day4.gms
 Consider a taste shock , so that utility from
cake consumption is now:
◦ Knows the value of stochastic shock today, but
unknown for future periods.
◦ Agent should factor in the potential future shocks
in today’s consumption decision
Day 4 NotesHowitt and Msangi 37
ε
( , )u c ε
 Step 1: Define nature of stochastic shock
◦ First-order Markov process: probability of future
shocks is described by current period
◦ Two states: , described by and
◦ The transition between states follows a first-order
Markov process, described by matrix :
◦ An element in the matrix yields the probability of
moving from state i to j in the next period:
Day 4 NotesHowitt and Msangi 38
andl h hε lε
Π
ll lh
hl hh
π π
π π
 
Π = 
 
( )1Pr |ij t j t iπ ε ε ε ε+≡ = =
 Agent’s choice of how much cake to eat
depends on:
◦ Size of cake
◦ Realization of the taste shock
 With current shock knowledge and expected
transition to future periods, the stochastic
cake-eating problem can be written as:
Day 4 NotesHowitt and Msangi 39
( ) ( ){ }1| 1 1 1( , ) max , , , ~t t
t
t t t t t t t t t
c
V x u c E V x x x c Markovε εε ε β ε ε+ + + +=+ =−
 Markov process for evolution of taste shock
states that today’s preferences yields the
probability of tomorrow’s preferences
◦ This may not hold if we believe that tomorrow does not
depend on the value today
 We can specify any type of random variable in the
SDP problem.
◦ Consider specifying the taste shock as a random variable
◦ Define e points, with known probability, , of a shock
with magnitude , we define the probabilities such
that:
◦
Day 4 NotesHowitt and Msangi 40
epr
eshk
1e
e
pr =∑
 After defining the known probability and shock of
magnitude, we can re-write the stochastic cake-eating
problem as:
 Assume the stochastic shock affects utility multiplicatively:
 Simple stochastic process where the distribution of e in
future periods is independent of the current period and
independent of other states and the control.
 The contraction mapping theorem holds: there exists a
fixed point of the function equation (Bellman)
◦ Solve for this point using same methods for the deterministic DP
Day 4 NotesHowitt and Msangi 41
( ) ( ){ }1| 1 1 1( , ) max , , , ~t t
t
t t t t e e t t t t t
c
V x e u c e E V x e x x c e RVβ + + + +=+ =−
( )1 1 1 1( , ) max ( ) ( ) ( ) ( ) ,
t
t t t t t t t t t t
c
e
V x e shk e u c pr e shk e V x e x x cβ + + + +
 
=+ =− 
 
∑
 SDP and DP framework both extend naturally to
models with several state variables.
◦ Will generally involve multiple states that we need to
simultaneously model
 For example: Herd stocking (prices, disease, rainfall, herd
size and population dynamics
◦ In general, we can write for any number of states m:
 Computational costs of extending the dynamic
framework to many states
◦ As the number of states increases, so does the number
of points we must evaluate and solve the DP.
◦ “Curse of dimensionality”
Day 4 NotesHowitt and Msangi 42
( ) ( ) ( ){ }1
1
1 1( ) max , ,..., ,..., ,t
t
m m m m m
t t t t t t t t
c
V x f c x x V x x x g x cβ ++ += + =
 Function Approximation
◦ Extend naturally to multi-state applications
 Chebychev approximation approach
◦ Extension to m states
 Define the state variables upper and lower bounds:
 Map to the [-1,1] interval using the same formula:
 Transformation back to the interval can be
calculated as:
Day 4 NotesHowitt and Msangi 43
,m m
L U  
2 1
ˆ cos , for 1,...,
2j
m j
x j n
n
π
− 
= = 
 
,m m
L U  
( )( )ˆ
2
j
j
m m m m m
m
x L U U L
x
+ −
=
◦ Given the mapping back to the interval, we can
now define the Chebychev interpolation matrix using the
recursive formula:
◦ Defined the state space and Chebychev nodes and basis
functions for each state variable m.
◦ We can write the Chebychev approximation to the value
function as:
◦ The value function approximation with multiple state
simply extends the Chebychev polynomials to additional
dimensions to approximate the solution over each state.
Day 4 NotesHowitt and Msangi 44
,m m
L U  
1
2
1 2
1
ˆ
ˆ2 3
m
m
m
j j j
x
x j
φ
φ
φ φ φ− −
=
=
= − ∀ ≥
1..1
.... m jj jm
m
mj j
V a φ= ∑ ∑ ∏
 Agro-Forestry Example:
 AgroForestryModel_DP_Day4.gms
◦ Varying degree of age, expected yield and
profitability—how do I manage a fixed amount of
land with new plantings and removals?
 Input Data and State Space
◦ 20 year time horizon
◦ Early, mature and old trees
◦ 60% of early tree plantings transition to mature
trees and 30% of mature trees transition to old
trees
Day 4 NotesHowitt and Msangi 45
 The transition between age profiles are as follows:
 Model Data
◦ 100 hectares
◦ Cost to uproot is 20/ha
◦ Cost to replant is 100/ha
◦ 5% discount rate
 Key Model Parameters
Day 4 NotesHowitt and Msangi 46
Transition Matrix Early Mature Old
Early 0.4 0.6 0
Mature 0 0.7 0.3
Old 0 0 1
Model Data Early Mature Old
Price per kg 10 10 10
Yield (kg/ha) 0 10 5
Initial profile (plantings) 10 5 4
 Simulation
◦ Three state variables: early, mature, old
◦ Approximate the solution of the infinite horizon
problem by Chebychev approximation of the value
function
 Define m=3, and:
Day 4 NotesHowitt and Msangi 47
1 2 3, ,1 2 3
jj j j
m
mj j j
V a φ= ∑∑∑ ∏
 Herd Dynamics Example:
 HerdDynamics_DP_Day4.gms
◦ Varying degree of age and productivity
◦ Three state variables: juvenile, female adult and
male adult
 Productive output: milk and meat
 Grazing land: fixed amount and known productivity
 Minimum number of livestock for breeding purposes
◦ When do we add to the herd, or sell from the herd,
given market conditions and resource constraints?
Day 4 NotesHowitt and Msangi 48
 Input Data
◦ 40 year time horizon
◦ 5% discount rate
◦ Other key input assumptions:
◦ Females birth rate = 1.5 juveniles per year
 30% juveniles, 30% transition to males, and 40% transition to females
◦ Herd can be fed by grazing on a fixed amount of land, or from off-farm
purchased feed
 Different nutrient content and ultimately different animal productivity
Day 4 NotesHowitt and Msangi 49
Input Data
Juven
ile
Adult
Male
Adult
Female
Animal weight 40 300 275
Milk yield (kg/yr/animal) 0 0 50
Initial animals 60 20 30
Birth rate per female
(animal/yr) 1.5 0 0
Transition Matrix Juvenile
Adult
Male
Adult
Female
Juvenile 0.3 0.3 0.4
Adult Male 0 1 0
Adult Female 0 0 1
 Simulation
◦ Over a 100 year time horizon
◦ Approximates the value function at 3 Chebychev
nodes
◦ Agent to maximize present value of profits by
determining optimal rates of:
 Animals sold and purchased
 Milk sold
◦ Agent may purchase off-farm feed, and responds to
fixed and known market demand and supply for
inputs and outputs
◦ Herd age evolves endogenously by defined
parameters
Day 4 NotesHowitt and Msangi 50

More Related Content

What's hot

Operators n dirac in qm
Operators n dirac in qmOperators n dirac in qm
Operators n dirac in qm
Anda Tywabi
 
EENG519FinalProjectReport
EENG519FinalProjectReportEENG519FinalProjectReport
EENG519FinalProjectReport
Daniel K
 

What's hot (20)

FJMS
FJMSFJMS
FJMS
 
Operators n dirac in qm
Operators n dirac in qmOperators n dirac in qm
Operators n dirac in qm
 
Ladder operator
Ladder operatorLadder operator
Ladder operator
 
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithmno U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
 
Wigner Quasi-probability Distribution of the Cosecant-squared Potential Well
Wigner Quasi-probability Distribution of the Cosecant-squared Potential WellWigner Quasi-probability Distribution of the Cosecant-squared Potential Well
Wigner Quasi-probability Distribution of the Cosecant-squared Potential Well
 
Response spectra
Response spectraResponse spectra
Response spectra
 
CLIM Fall 2017 Course: Statistics for Climate Research, Detection & Attributi...
CLIM Fall 2017 Course: Statistics for Climate Research, Detection & Attributi...CLIM Fall 2017 Course: Statistics for Climate Research, Detection & Attributi...
CLIM Fall 2017 Course: Statistics for Climate Research, Detection & Attributi...
 
PosterDinius
PosterDiniusPosterDinius
PosterDinius
 
Supervisory control of discrete event systems for linear temporal logic speci...
Supervisory control of discrete event systems for linear temporal logic speci...Supervisory control of discrete event systems for linear temporal logic speci...
Supervisory control of discrete event systems for linear temporal logic speci...
 
Proje kt report
Proje kt reportProje kt report
Proje kt report
 
Response spectrum method
Response spectrum methodResponse spectrum method
Response spectrum method
 
Cyclo-stationary processes
Cyclo-stationary processesCyclo-stationary processes
Cyclo-stationary processes
 
LieGroup
LieGroupLieGroup
LieGroup
 
Introduction to the Keldysh non-equlibrium Green's function technique
Introduction to the Keldysh non-equlibrium Green's function techniqueIntroduction to the Keldysh non-equlibrium Green's function technique
Introduction to the Keldysh non-equlibrium Green's function technique
 
Stochastic Approximation and Simulated Annealing
Stochastic Approximation and Simulated AnnealingStochastic Approximation and Simulated Annealing
Stochastic Approximation and Simulated Annealing
 
Response spectrum
Response spectrumResponse spectrum
Response spectrum
 
Multivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributionsMultivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributions
 
Gravitational Waves and Binary Systems (2) - Thibault Damour
Gravitational Waves and Binary Systems (2) - Thibault DamourGravitational Waves and Binary Systems (2) - Thibault Damour
Gravitational Waves and Binary Systems (2) - Thibault Damour
 
Proje kt
Proje ktProje kt
Proje kt
 
EENG519FinalProjectReport
EENG519FinalProjectReportEENG519FinalProjectReport
EENG519FinalProjectReport
 

Similar to Biosight: Quantitative Methods for Policy Analysis: Stochastic Dynamic Programming

fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
Karl Rudeen
 
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language ModelsAnomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Cynthia Freeman
 
SLAM of Multi-Robot System Considering Its Network Topology
SLAM of Multi-Robot System Considering Its Network TopologySLAM of Multi-Robot System Considering Its Network Topology
SLAM of Multi-Robot System Considering Its Network Topology
toukaigi
 
IFAC2008art
IFAC2008artIFAC2008art
IFAC2008art
Yuri Kim
 

Similar to Biosight: Quantitative Methods for Policy Analysis: Stochastic Dynamic Programming (20)

pRO
pROpRO
pRO
 
fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
fb69b412-97cb-4e8d-8a28-574c09557d35-160618025920
 
Project Paper
Project PaperProject Paper
Project Paper
 
Universal approximators for Direct Policy Search in multi-purpose water reser...
Universal approximators for Direct Policy Search in multi-purpose water reser...Universal approximators for Direct Policy Search in multi-purpose water reser...
Universal approximators for Direct Policy Search in multi-purpose water reser...
 
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language ModelsAnomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
 
Families of Triangular Norm Based Kernel Function and Its Application to Kern...
Families of Triangular Norm Based Kernel Function and Its Application to Kern...Families of Triangular Norm Based Kernel Function and Its Application to Kern...
Families of Triangular Norm Based Kernel Function and Its Application to Kern...
 
Mx/G(a,b)/1 With Modified Vacation, Variant Arrival Rate With Restricted Admi...
Mx/G(a,b)/1 With Modified Vacation, Variant Arrival Rate With Restricted Admi...Mx/G(a,b)/1 With Modified Vacation, Variant Arrival Rate With Restricted Admi...
Mx/G(a,b)/1 With Modified Vacation, Variant Arrival Rate With Restricted Admi...
 
MVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priorsMVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priors
 
SLAM of Multi-Robot System Considering Its Network Topology
SLAM of Multi-Robot System Considering Its Network TopologySLAM of Multi-Robot System Considering Its Network Topology
SLAM of Multi-Robot System Considering Its Network Topology
 
Actuator Constrained Optimal Control of Formations Near the Libration Points
Actuator Constrained Optimal Control of Formations Near the Libration PointsActuator Constrained Optimal Control of Formations Near the Libration Points
Actuator Constrained Optimal Control of Formations Near the Libration Points
 
An Exponential Observer Design for a Class of Chaotic Systems with Exponentia...
An Exponential Observer Design for a Class of Chaotic Systems with Exponentia...An Exponential Observer Design for a Class of Chaotic Systems with Exponentia...
An Exponential Observer Design for a Class of Chaotic Systems with Exponentia...
 
Chaos Suppression and Stabilization of Generalized Liu Chaotic Control System
Chaos Suppression and Stabilization of Generalized Liu Chaotic Control SystemChaos Suppression and Stabilization of Generalized Liu Chaotic Control System
Chaos Suppression and Stabilization of Generalized Liu Chaotic Control System
 
Av 738- Adaptive Filtering - Background Material
Av 738- Adaptive Filtering - Background MaterialAv 738- Adaptive Filtering - Background Material
Av 738- Adaptive Filtering - Background Material
 
Slides ensae-2016-9
Slides ensae-2016-9Slides ensae-2016-9
Slides ensae-2016-9
 
Scalable trust-region method for deep reinforcement learning using Kronecker-...
Scalable trust-region method for deep reinforcement learning using Kronecker-...Scalable trust-region method for deep reinforcement learning using Kronecker-...
Scalable trust-region method for deep reinforcement learning using Kronecker-...
 
On the principle of optimality for linear stochastic dynamic system
On the principle of optimality for linear stochastic dynamic systemOn the principle of optimality for linear stochastic dynamic system
On the principle of optimality for linear stochastic dynamic system
 
GDRR Opening Workshop - Variance Reduction for Reliability Assessment with St...
GDRR Opening Workshop - Variance Reduction for Reliability Assessment with St...GDRR Opening Workshop - Variance Reduction for Reliability Assessment with St...
GDRR Opening Workshop - Variance Reduction for Reliability Assessment with St...
 
1789 1800
1789 18001789 1800
1789 1800
 
1789 1800
1789 18001789 1800
1789 1800
 
IFAC2008art
IFAC2008artIFAC2008art
IFAC2008art
 

More from IFPRI-EPTD

FAO_PRESS_RELEASE_PWC_Buffalo
FAO_PRESS_RELEASE_PWC_BuffaloFAO_PRESS_RELEASE_PWC_Buffalo
FAO_PRESS_RELEASE_PWC_Buffalo
IFPRI-EPTD
 

More from IFPRI-EPTD (12)

FAO_PRESS_RELEASE_PWC_Buffalo
FAO_PRESS_RELEASE_PWC_BuffaloFAO_PRESS_RELEASE_PWC_Buffalo
FAO_PRESS_RELEASE_PWC_Buffalo
 
Proyecciónde la Emisión, reservaCarbono, y economía Baja emisióny desarrollo
Proyecciónde la Emisión, reservaCarbono, y economía Baja emisióny desarrolloProyecciónde la Emisión, reservaCarbono, y economía Baja emisióny desarrollo
Proyecciónde la Emisión, reservaCarbono, y economía Baja emisióny desarrollo
 
IFPRI Low Emissions Development Strategies (LEDS) Colombia
IFPRI Low Emissions Development Strategies (LEDS) ColombiaIFPRI Low Emissions Development Strategies (LEDS) Colombia
IFPRI Low Emissions Development Strategies (LEDS) Colombia
 
Biosight: Quantitative Methods for Policy Analysis: Multi Market Models
Biosight: Quantitative Methods for Policy Analysis: Multi Market ModelsBiosight: Quantitative Methods for Policy Analysis: Multi Market Models
Biosight: Quantitative Methods for Policy Analysis: Multi Market Models
 
Biosight: Quantitative Methods for Policy Analysis : Dynamic Models
Biosight: Quantitative Methods for Policy Analysis : Dynamic ModelsBiosight: Quantitative Methods for Policy Analysis : Dynamic Models
Biosight: Quantitative Methods for Policy Analysis : Dynamic Models
 
Biosight: Quantitative Methods for Policy Analysis: CES Production Function a...
Biosight: Quantitative Methods for Policy Analysis: CES Production Function a...Biosight: Quantitative Methods for Policy Analysis: CES Production Function a...
Biosight: Quantitative Methods for Policy Analysis: CES Production Function a...
 
Biosight: Quantitative Methods for Policy Analysis - Introduction to GAMS, Li...
Biosight: Quantitative Methods for Policy Analysis - Introduction to GAMS, Li...Biosight: Quantitative Methods for Policy Analysis - Introduction to GAMS, Li...
Biosight: Quantitative Methods for Policy Analysis - Introduction to GAMS, Li...
 
Biosight: Quantitative Methods for Policy Analysis using GAMS
Biosight: Quantitative Methods for Policy Analysis using GAMSBiosight: Quantitative Methods for Policy Analysis using GAMS
Biosight: Quantitative Methods for Policy Analysis using GAMS
 
Asti @ caadp pp
Asti @ caadp ppAsti @ caadp pp
Asti @ caadp pp
 
Low Emissions Development Strategies (Colombia Feb 20, 2014)
Low Emissions Development Strategies (Colombia Feb 20, 2014)Low Emissions Development Strategies (Colombia Feb 20, 2014)
Low Emissions Development Strategies (Colombia Feb 20, 2014)
 
Low Emissions Development Strategies (LEDS) Training Sept 9, 2013
Low Emissions Development Strategies (LEDS) Training Sept 9, 2013Low Emissions Development Strategies (LEDS) Training Sept 9, 2013
Low Emissions Development Strategies (LEDS) Training Sept 9, 2013
 
Future African Competitiveness: Foresight for better agricultural futures
Future African Competitiveness: Foresight for better agricultural futuresFuture African Competitiveness: Foresight for better agricultural futures
Future African Competitiveness: Foresight for better agricultural futures
 

Recently uploaded

Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
negromaestrong
 
Gardella_Mateo_IntellectualProperty.pdf.
Gardella_Mateo_IntellectualProperty.pdf.Gardella_Mateo_IntellectualProperty.pdf.
Gardella_Mateo_IntellectualProperty.pdf.
MateoGardella
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
QucHHunhnh
 

Recently uploaded (20)

Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
 
Gardella_Mateo_IntellectualProperty.pdf.
Gardella_Mateo_IntellectualProperty.pdf.Gardella_Mateo_IntellectualProperty.pdf.
Gardella_Mateo_IntellectualProperty.pdf.
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Advance Mobile Application Development class 07
Advance Mobile Application Development class 07Advance Mobile Application Development class 07
Advance Mobile Application Development class 07
 

Biosight: Quantitative Methods for Policy Analysis: Stochastic Dynamic Programming

  • 1. Day 4: Stochastic Dynamic Programming Day 4 NotesHowitt and Msangi 1
  • 2.  Understand Bellman’s Principle of Optimality and the basic Stochastic Dynamic programming problem  Solve the SDP with value function iteration  Apply the concepts of models to agro- forestry and livestock herd dynamics  Make changes to the SDP and simulate the corresponding change in optimal solution Day 4 NotesHowitt and Msangi 2
  • 3.  Re-cap on rangeland stocking model….  Introduction to Stochastic Dynamic Programming ◦ Extend DP framework to include stochastic state variables and apply to herd and agro forestry management  Stochastic Cake Eating  Multi-State Models ◦ Function Approximation  Agro-Forestry Application ◦ Input Data and State Space ◦ Simulation  Herd Dynamics Application ◦ Input Data ◦ Simulation Day 4 NotesHowitt and Msangi 3
  • 4. An Application to Reservoir Management Day 3 NotesHowitt and Msangi 4
  • 5.  “Estimating Intertemporal Preferences for Resource Allocation” AJAE, 87(4): 969-983. (Howitt RE, S Msangi, A Reynaud, KC Knapp)  What started out as a calibration exercise – ended up as a research project (with some interesting research discoveries) Day 3 NotesHowitt and Msangi 5
  • 6.  Many of the Important Policy Questions in Natural Resource Management Revolve Around How to Deal with Uncertainty over Time (Global Climate Change, Extreme Weather Events, Invasive Species Encroachment, Disease Outbreak, etc. )  Policy Makers look to Economic Models to Provide them with Guidance on Best Management Practices
  • 7.  Economic Policy Models Have Typically Downplayed the Role of Risk in the Preferences of the Decision-maker  Few Studies Have Ever Tried to Measure the Degree to Which Risk Aversion Matters in Resource Management Problems  Time-Additive Separability in Dynamic Models Imposes Severe Constraints on Intertemporal Preferences
  • 8.  In order to Address this Gap in the Natural Resources literature….  We Applied Dynamic Estimation Methods to an Example of Reservoir Management  We Relaxed the Assumption of Time-Additive Separability of the Decision-Maker’s Utility  We Tested with Alternative Utility Forms to Determine the Importance of Risk Aversion
  • 9.  Koopmans (1960) laid the foundation for eliminating deficiencies of TAS with recursive preferences.  Recursive Utility is a class of functionals designed to offer a generality to time preferences while still maintaining time consistency in behavior.  Allows for the potential smoothing of consumption by allowing complementarity between time periods.
  • 10. ( )W  ( )1( ) ( ), ( )U W u c U S=c c States the weak separability of the future from present where is an aggregator function For TAS, the aggregator is simply ( )( ), ( )W u c x u c xβ= +
  • 11. ( ) ( ) 1 ( ), 1 ( )W u c x u c x ρρ ρ β β = − ⋅ + ⋅  1 ( ) 1 EIS σ ρ = − So we choose our aggregator to be and the implied elasticity (“resistance”) to inter-temporal substitution is given by where ( ),0 (0,1]ρ ∈ −∞ ∪
  • 12. Time Additive Separable Utility Using Bellman’s recursive relationship: { } { } { } { } 1 2 1 2 1 1 1 2 2 2 3 2 3 1 2 3 , , 1 1, 2 ( ) max ( ) ( ) ( ) max ( ) ( ) ( ) max ( ) ( ) : ( ) max ( ) ( ) ( ) ( ) ( ) ( ) t t t t t t t t t c t t t c t t t c t t t t t c c c t t t V x U c V x V x U c V x V x U c V x Substituting and simplifying V x U c U c U c V x u c Note that MRS c β β β β β β β + + + + + + + + + + + + + + + + + + + + = + + + ′ = 2( )tu c + ′
  • 13. Iso-Elastic Recursive Utility A utility function with a CES across time periods. 1 2 1 1 1 1 1 1 2 1 2 2 3 ( 1) 1 , ( ) max (1 ) ( ) ( ) ( ) max (1 ) ( ) ( ) ( ) max (1 ) ( ) ( ) : ( ) max (1 ) ( ) (1 ) ( t t t t t t t t c t t t c t t t c t At t t t c c V x U c V x V x U c V x V x U c V x Substituting and simplifying V x U c U c ρ ρ ρ ρ ρ ρ ρ ρ ρ ρ β β β β β β β β β + + + + + + + + + + + +  = − +   = − +   = − +  = − + − 1 1 2 1 1 2 2 1 2 3 ( 2) 1 2 3 , , 1 2 1 1 2 1 2 2 ) ( ) ( ) max (1 ) ( ) (1 ) ( ) (1 ) ( ) ( ) ( ) (1 ) ( ) (1 ) ( ) (1 ) ( ) ( ) ( ) (1 ) t t t t t At t t t t t c c c t t t t t t t t V x V x U c U c U c V x V x u c U c U c V x c V x c ρ ρ ρ ρ ρ ρ ρ ρ ρ ρ ρ ρ ρ β β β β β β β β β β β β β β β + + − + + + + + − + + + + +  +   = − + − + − +  ∂ ′  = − − + − + ∂ ∂ ′= − ∂ 1 1 1 2 3 2 1 2 3( ) (1 ) ( ) (1 ) ( ) (1 ) ( ) ( )t t t t tu c U c U c U c V x ρ ρ ρ ρ ρ ρ β β β β β β − − + + + +  − + − + − + 
  • 14. With Recursive Utility All Periods Enter into MRS 1 1 1 2 1 1 2 1, 2 2 3 2 1 2 3 ( ) (1 ) ( ) (1 ) ( ) ( )1 ( ) (1 ) ( ) (1 ) ( ) (1 ) ( ) ( ) t t t t t t t t t t t u c U c U c V x MRS u c U c U c U c V x ρ ρ ρ ρ ρ ρ ρ ρ ρ β β β β β β β β β β β − − + + + + + + + + +    ′ − + − + =    ′ − + − + − +    In micro-economics we have an appreciation of the difference between linear and CES utility in static consumer theory The same intuition applies here in a dynamic context….
  • 15.  The previous equations show that the marginal rate of substitution across time is path dependent.  Timing is now an explicit economic control variable  We no longer assume that “The marginal rate of substitution between lunch and dinner is independent of the amount of breakfast” (Henry Wan).  A smaller elasticity of intertemporal substitution flattens out the optimal time path of resource use- yielding a time consistent sustainable result.
  • 16.  Stochastic Equations of Motion link Stocks and Flows  Randomness in the equations of motion or exogenous random shocks change the system evolution  The current state and future distributions are usually known to decision makers  Management decisions inherently optimize a stochastic dynamic path of resource use and consequently maximize dynamic stochastic utility
  • 17. A Simple Resource Network with a Single State Variable te1 ~ te2 ~ Demand tS tw
  • 18. ( ) ( )2 1 1 1Max (1 ) E ( ) Et e t t e t w U W q U ρ ρ ρ α α β β +    = − ⋅ +     .         ≥ ≤ ≥ += −+= + + + 0 ~ ~ 1 1 2 11 t t t ttt tttt w SS SS ewq weSS The Optimization Problem
  • 19. ( ) 1 2 2 1 1 0 , Max (1 ) ( )d ( , )dt / w V S e W w e Φ V S e Φ ρ ρ ρ α α β β ≥      = − ⋅ + +         ∫ ∫  Which can be re-stated in terms of Bellman’s Recurrence Relationship… ..and which we solve by numerically with Continuous-valued State and Control Variables
  • 20. Solving for the Expected Value Function Initialize with Intermediate Value Function W(Xt , ut ) Nodes for: State Evaluation and Stochastic Inflow values Probabilities of Inflow over k Stochastic node values Define the State Variable on [ -1, 1] Interval for each polynomial node j Value Iteration Loop (n steps) n = n+1 Error = If Error > 0.1 e-7 Stop Value Function Chebychev Polynomial Coefficients jXaVpuXWPVNB k i kj ti n iktt n j ∀                     += ∑ ∑ + − )(),(max , 1 1  φβ ( )∑ − − i n i n i aa 21 ( ) ( ) ( )∑ ∑ ++ + = j nj ti nj ti j nj ti n j n i XX XPVNB a , 1 , 1 , 1   φφ φ ( )j x 
  • 21.
  • 22.
  • 23. ( ) 32 0067.045.0150 qqqqW ⋅+⋅−⋅= ttt ttttt capee eecapesp ⋅⋅⋅+ ⋅+⋅= 1 3 1 2 111 ~0.02305-~0.000993 ~0.005024~0.095382),~( Current Profit Function Spill Function
  • 24. Net Benefit Function for Water 0 1 2 3 4 5 6 0 4 8 12 16 20 24 28 32 36 40 44 48 52 56 q, (MAF) W(q),1000M$US
  • 25.  We employ a nested procedure to solve the SDP problem with value iteration, while we systematically change the parameter values of the objective function to maximize a likelihood function.  We employ a derivative-free ( Nelder Meade) search algorithm to implement the ‘hill- climbing’ procedure that searches for the likelihood-maximizing values of preference parameters
  • 26. EIS value ( )1 1 ρ− Coeff. of Risk Aversion 1 α− ρ α These parameters were calculated with a fixed discount rate of β Parameter Estimated Value Standard Error -9.000 4.60 0.100 -0.440 0.23 1.440 Log Likelihood -10.257 =0.95. Standard errors are based on 500 bootstrap repetitions
  • 27. 1,set estimateα ρ= For Risk-Neutral Recursive model (RNR) For Risk-Neutral (non-Recursive) model (RN) For Non-Recursive model (with Risk) use CRRA 1set ρ α= = ( )0.95fix β = ( )0.95fix β = )( 1 1 )1( + − + − = t t t UE W U βα α ,estimate α β
  • 28. 0 0.5 1 1.5 2 2.5 3 3.5 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 MillionAcrefeet RN ACTUAL
  • 29. 0 0.5 1 1.5 2 2.5 3 3.5 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 MillionAcrefeet CRRA ACTUAL
  • 30. 0 0.5 1 1.5 2 2.5 3 3.5 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 MillionAcrefeet REC ACTUAL
  • 31. 0 1 2 3 4 5 6 7 8 9 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 MillionAcrefeet RN ACTUAL
  • 32. 0 1 2 3 4 5 6 7 8 9 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 MillionAcrefeet REC CRRA ACTUAL
  • 33.  Clearly a non-recursive model that ignores risk fares the worst, when compared to actual storage and releases  Adding risk, but not recursivity of preferences, gets you closer to actual values…but not quite….  A Recursive Specification outperforms both of these, with or without risk aversion
  • 34.  Estimation of the Fully-Recursive model is robust to Discount Values and the Parameter Estimates appear to be Stationary over the Study Period  Once we allow Intertemporal Preferences to be recursive, the role of Risk in explaining Resource Management Behavior is Reduced  Imposing Time-Additive Separability on Dynamic Models may have more severe implications for behavior than most researchers realize…..
  • 35. Day 3 NotesHowitt and Msangi 35
  • 36.  Extend DP framework to include stochastic state variables in the model  Apply the new framework to herd dynamics and agro-forestry management  Return to cake eating example Day 4 NotesHowitt and Msangi 36
  • 37.  Stochastic Cake Eating ◦ What if I want cake today, but not tomorrow?  Cake Eating Example:  CakeEatingDP_ChebyAprox_Stochastic_Day4.gms  Consider a taste shock , so that utility from cake consumption is now: ◦ Knows the value of stochastic shock today, but unknown for future periods. ◦ Agent should factor in the potential future shocks in today’s consumption decision Day 4 NotesHowitt and Msangi 37 ε ( , )u c ε
  • 38.  Step 1: Define nature of stochastic shock ◦ First-order Markov process: probability of future shocks is described by current period ◦ Two states: , described by and ◦ The transition between states follows a first-order Markov process, described by matrix : ◦ An element in the matrix yields the probability of moving from state i to j in the next period: Day 4 NotesHowitt and Msangi 38 andl h hε lε Π ll lh hl hh π π π π   Π =    ( )1Pr |ij t j t iπ ε ε ε ε+≡ = =
  • 39.  Agent’s choice of how much cake to eat depends on: ◦ Size of cake ◦ Realization of the taste shock  With current shock knowledge and expected transition to future periods, the stochastic cake-eating problem can be written as: Day 4 NotesHowitt and Msangi 39 ( ) ( ){ }1| 1 1 1( , ) max , , , ~t t t t t t t t t t t t c V x u c E V x x x c Markovε εε ε β ε ε+ + + +=+ =−
  • 40.  Markov process for evolution of taste shock states that today’s preferences yields the probability of tomorrow’s preferences ◦ This may not hold if we believe that tomorrow does not depend on the value today  We can specify any type of random variable in the SDP problem. ◦ Consider specifying the taste shock as a random variable ◦ Define e points, with known probability, , of a shock with magnitude , we define the probabilities such that: ◦ Day 4 NotesHowitt and Msangi 40 epr eshk 1e e pr =∑
  • 41.  After defining the known probability and shock of magnitude, we can re-write the stochastic cake-eating problem as:  Assume the stochastic shock affects utility multiplicatively:  Simple stochastic process where the distribution of e in future periods is independent of the current period and independent of other states and the control.  The contraction mapping theorem holds: there exists a fixed point of the function equation (Bellman) ◦ Solve for this point using same methods for the deterministic DP Day 4 NotesHowitt and Msangi 41 ( ) ( ){ }1| 1 1 1( , ) max , , , ~t t t t t t t e e t t t t t c V x e u c e E V x e x x c e RVβ + + + +=+ =− ( )1 1 1 1( , ) max ( ) ( ) ( ) ( ) , t t t t t t t t t t t c e V x e shk e u c pr e shk e V x e x x cβ + + + +   =+ =−    ∑
  • 42.  SDP and DP framework both extend naturally to models with several state variables. ◦ Will generally involve multiple states that we need to simultaneously model  For example: Herd stocking (prices, disease, rainfall, herd size and population dynamics ◦ In general, we can write for any number of states m:  Computational costs of extending the dynamic framework to many states ◦ As the number of states increases, so does the number of points we must evaluate and solve the DP. ◦ “Curse of dimensionality” Day 4 NotesHowitt and Msangi 42 ( ) ( ) ( ){ }1 1 1 1( ) max , ,..., ,..., ,t t m m m m m t t t t t t t t c V x f c x x V x x x g x cβ ++ += + =
  • 43.  Function Approximation ◦ Extend naturally to multi-state applications  Chebychev approximation approach ◦ Extension to m states  Define the state variables upper and lower bounds:  Map to the [-1,1] interval using the same formula:  Transformation back to the interval can be calculated as: Day 4 NotesHowitt and Msangi 43 ,m m L U   2 1 ˆ cos , for 1,..., 2j m j x j n n π −  = =    ,m m L U   ( )( )ˆ 2 j j m m m m m m x L U U L x + − =
  • 44. ◦ Given the mapping back to the interval, we can now define the Chebychev interpolation matrix using the recursive formula: ◦ Defined the state space and Chebychev nodes and basis functions for each state variable m. ◦ We can write the Chebychev approximation to the value function as: ◦ The value function approximation with multiple state simply extends the Chebychev polynomials to additional dimensions to approximate the solution over each state. Day 4 NotesHowitt and Msangi 44 ,m m L U   1 2 1 2 1 ˆ ˆ2 3 m m m j j j x x j φ φ φ φ φ− − = = = − ∀ ≥ 1..1 .... m jj jm m mj j V a φ= ∑ ∑ ∏
  • 45.  Agro-Forestry Example:  AgroForestryModel_DP_Day4.gms ◦ Varying degree of age, expected yield and profitability—how do I manage a fixed amount of land with new plantings and removals?  Input Data and State Space ◦ 20 year time horizon ◦ Early, mature and old trees ◦ 60% of early tree plantings transition to mature trees and 30% of mature trees transition to old trees Day 4 NotesHowitt and Msangi 45
  • 46.  The transition between age profiles are as follows:  Model Data ◦ 100 hectares ◦ Cost to uproot is 20/ha ◦ Cost to replant is 100/ha ◦ 5% discount rate  Key Model Parameters Day 4 NotesHowitt and Msangi 46 Transition Matrix Early Mature Old Early 0.4 0.6 0 Mature 0 0.7 0.3 Old 0 0 1 Model Data Early Mature Old Price per kg 10 10 10 Yield (kg/ha) 0 10 5 Initial profile (plantings) 10 5 4
  • 47.  Simulation ◦ Three state variables: early, mature, old ◦ Approximate the solution of the infinite horizon problem by Chebychev approximation of the value function  Define m=3, and: Day 4 NotesHowitt and Msangi 47 1 2 3, ,1 2 3 jj j j m mj j j V a φ= ∑∑∑ ∏
  • 48.  Herd Dynamics Example:  HerdDynamics_DP_Day4.gms ◦ Varying degree of age and productivity ◦ Three state variables: juvenile, female adult and male adult  Productive output: milk and meat  Grazing land: fixed amount and known productivity  Minimum number of livestock for breeding purposes ◦ When do we add to the herd, or sell from the herd, given market conditions and resource constraints? Day 4 NotesHowitt and Msangi 48
  • 49.  Input Data ◦ 40 year time horizon ◦ 5% discount rate ◦ Other key input assumptions: ◦ Females birth rate = 1.5 juveniles per year  30% juveniles, 30% transition to males, and 40% transition to females ◦ Herd can be fed by grazing on a fixed amount of land, or from off-farm purchased feed  Different nutrient content and ultimately different animal productivity Day 4 NotesHowitt and Msangi 49 Input Data Juven ile Adult Male Adult Female Animal weight 40 300 275 Milk yield (kg/yr/animal) 0 0 50 Initial animals 60 20 30 Birth rate per female (animal/yr) 1.5 0 0 Transition Matrix Juvenile Adult Male Adult Female Juvenile 0.3 0.3 0.4 Adult Male 0 1 0 Adult Female 0 0 1
  • 50.  Simulation ◦ Over a 100 year time horizon ◦ Approximates the value function at 3 Chebychev nodes ◦ Agent to maximize present value of profits by determining optimal rates of:  Animals sold and purchased  Milk sold ◦ Agent may purchase off-farm feed, and responds to fixed and known market demand and supply for inputs and outputs ◦ Herd age evolves endogenously by defined parameters Day 4 NotesHowitt and Msangi 50