You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: lectures/robustness.md
+18-18Lines changed: 18 additions & 18 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -85,7 +85,7 @@ import quantecon as qe
85
85
```
86
86
87
87
(rb_vec)=
88
-
### Sets of Models Imply Sets Of Values
88
+
### Sets of models imply sets of values
89
89
90
90
Our "robust" decision-maker wants to know how well a given rule will work when he does not *know* a single transition law $\ldots$.
91
91
@@ -144,18 +144,18 @@ We will also describe how to choose $F$ to shape the sets of values.
144
144
145
145
This will involve crafting a *skinnier* set at the cost of a lower *level* (at least for low values of entropy).
146
146
147
-
### Inspiring Video
147
+
### Inspiring video
148
148
149
149
If you want to understand more about why one serious quantitative researcher is interested in this approach, we recommend [Lars Peter Hansen's Nobel lecture](https://www.nobelprize.org/prizes/economic-sciences/2013/hansen/lecture/).
150
150
151
-
### Other References
151
+
### Other references
152
152
153
153
Our discussion in this lecture is based on
154
154
155
155
* {cite}`HansenSargent2000`
156
156
* {cite}`HansenSargent2008`
157
157
158
-
## The Model
158
+
## The model
159
159
160
160
For simplicity, we present ideas in the context of a class of problems with linear transition laws and quadratic objective functions.
Soon we'll quantify the quality of a model specification in terms of the maximal size of the discounted sum $\sum_{t=0}^{\infty} \beta^{t+1}w_{t+1}' w_{t+1}$.
212
212
213
-
## Constructing More Robust Policies
213
+
## Constructing more robust policies
214
214
215
215
If our agent takes $\{ w_t \}$ as a given deterministic sequence, then, drawing on ideas in earlier lectures on dynamic programming, we can anticipate Bellman equations such as
216
216
@@ -253,7 +253,7 @@ By raising $\theta$ more and more, we more and more limit the ability of maximiz
253
253
254
254
So bigger $\theta$ is implicitly associated with smaller distortion sequences $\{w_t \}$.
255
255
256
-
### Analyzing the Bellman Equation
256
+
### Analyzing the Bellman equation
257
257
258
258
So what does $J$ in {eq}`rb_wcb0` look like?
259
259
@@ -353,7 +353,7 @@ Furthermore, when $\theta$ is large, $\hat K$ is approximately equal to zero.
353
353
354
354
Conversely, smaller $\theta$ is associated with greater fear of model misspecification and greater concern for robustness.
355
355
356
-
## Robustness as Outcome of a Two-Person Zero-Sum Game
356
+
## Robustness as outcome of a two-person zero-sum game
357
357
358
358
What we have done above can be interpreted in terms of a two-person zero-sum game in which $\hat F, \hat K$ are Nash equilibrium objects.
359
359
@@ -366,7 +366,7 @@ Agent 2's malevolence helps the original agent to compute bounds on his value fu
366
366
We begin with agent 2's problem.
367
367
368
368
(rb_a2)=
369
-
### Agent 2's Problem
369
+
### Agent 2's problem
370
370
371
371
Agent 2
372
372
@@ -464,9 +464,9 @@ The remaining step for agent 2's problem is to set $\theta$ to enforce the const
464
464
Here $x_t$ is given by {eq}`rob_lomf` --- which in this case becomes $x_{t+1} = (A - B F + CK(F, \theta)) x_t$.
465
465
466
466
(rb_a1)=
467
-
### Using Agent 2's Problem to Construct Bounds on Value Sets
467
+
### Using Agent 2's problem to construct bounds on value sets
468
468
469
-
#### The Lower Bound
469
+
#### The lower bound
470
470
471
471
Define the minimized object on the right side of problem {eq}`rb_a2o` as $R_\theta(x_0, F)$.
472
472
@@ -518,7 +518,7 @@ To construct the *lower bound* on the set of values associated with all perturba
518
518
This procedure sweeps out a set of separating hyperplanes indexed by different values for the Lagrange multiplier $\theta$.
519
519
```
520
520
521
-
#### The Upper Bound
521
+
#### The upper bound
522
522
523
523
To construct an *upper bound* we use a very similar procedure.
524
524
@@ -575,11 +575,11 @@ To construct the *upper bound* on the set of values associated all perturbations
575
575
* Compute the upper bound on the value function $V_{\tilde \theta}(x_0, F) + \tilde \theta \ {\rm ent}$ and plot it against ${\rm ent}$.
576
576
* Repeat the preceding three steps for a range of values of $\tilde \theta$ to trace out the upper bound.
577
577
578
-
#### Reshaping the Set of Values
578
+
#### Reshaping the set of values
579
579
580
580
Now in the interest of *reshaping* these sets of values by choosing $F$, we turn to agent 1's problem.
581
581
582
-
### Agent 1's Problem
582
+
### Agent 1's problem
583
583
584
584
Now we turn to agent 1, who solves
585
585
@@ -618,7 +618,7 @@ Once again, the expression for the optimal policy can be found [here](https://py
618
618
it by $\tilde F$.
619
619
620
620
(rb_eq)=
621
-
### Nash Equilibrium
621
+
### Nash equilibrium
622
622
623
623
Clearly, the $\tilde F$ we have obtained depends on $K$, which, in agent 2's problem,
624
624
depended on an initial policy $F$.
@@ -644,7 +644,7 @@ In particular, for any given $\theta$,
644
644
645
645
A sketch of the proof is given in {ref}`the appendix <rb_appendix>`.
646
646
647
-
## The Stochastic Case
647
+
## The stochastic case
648
648
649
649
Now we turn to the stochastic case, where the sequence $\{w_t\}$ is treated as an IID sequence of random vectors.
650
650
@@ -687,7 +687,7 @@ The distribution $\phi$ is chosen as the least desirable conditional distributio
687
687
688
688
This penalty term plays a role analogous to the one played by the deterministic penalty $\theta w'w$ in {eq}`rb_wcb0`, since it discourages large deviations from the benchmark.
689
689
690
-
### Solving the Model
690
+
### Solving the model
691
691
692
692
The maximization problem in {eq}`rb_wcb1` appears highly nontrivial --- after all,
693
693
we are maximizing over an infinite dimensional space consisting of the entire set of densities.
@@ -777,11 +777,11 @@ where $\hat K$ is given by {eq}`rb_kd`.
777
777
Note that the mean of the worst-case shock distribution is equal to the same worst-case $w_{t+1}$ as in the earlier deterministic setting.
778
778
779
779
(rb_coq)=
780
-
### Computing Other Quantities
780
+
### Computing other quantities
781
781
782
782
Before turning to implementation, we briefly outline how to compute several other quantities of interest.
783
783
784
-
#### Worst-Case Value of a Policy
784
+
#### Worst-case value of a policy
785
785
786
786
One thing we will be interested in doing is holding a policy fixed and
787
787
computing the discounted loss associated with that policy.
0 commit comments