Skip to content

Commit af67682

Browse files
Tom's Nov 4 edits of calvo.md lecture
1 parent fe277e3 commit af67682

File tree

1 file changed

+20
-30
lines changed

1 file changed

+20
-30
lines changed

lectures/calvo.md

Lines changed: 20 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ We'll use ideas from papers by Cagan {cite}`Cagan`, Calvo {cite}`Calvo1978`, an
6464
well as from chapter 19 of {cite}`Ljungqvist2012`.
6565

6666
In addition, we'll use ideas from linear-quadratic dynamic programming
67-
described in [Linear Quadratic Control](https://python-intro.quantecon.org/lqcontrol.html) as applied to Ramsey problems in {doc}`Stackelberg problems <dyn_stack>`.
67+
described in [Linear Quadratic Control](https://python-intro.quantecon.org/lqcontrol.html) as applied to Ramsey problems in {doc}`Stackelberg plans <dyn_stack>`.
6868

6969
We specify model fundamentals in ways that allow us to use
7070
linear-quadratic discounted dynamic programming to compute an optimal government
@@ -104,8 +104,7 @@ Let:
104104
- $\theta_t = p_{t+1} - p_t$ be the net rate of inflation between $t$ and $t+1$
105105
- $\mu_t = m_{t+1} - m_t$ be the net rate of growth of nominal balances
106106

107-
The demand for real balances is governed by a perfect foresight
108-
version of a Cagan {cite}`Cagan` demand function for real balances:
107+
The demand for real balances is governed by a discrete time version of Sargent and Wallace's {cite}`sargent1973stability` perfect foresight version of a Cagan {cite}`Cagan` demand function for real balances:
109108

110109
```{math}
111110
:label: eq_old1
@@ -119,9 +118,9 @@ Equation {eq}`eq_old1` asserts that the demand for real balances is inversely
119118
related to the public's expected rate of inflation, which equals
120119
the actual rate of inflation because there is no uncertainty here.
121120

122-
(When there is no uncertainty, an assumption of **rational expectations** that becomes equivalent to **perfect foresight**).
121+
(When there is no uncertainty, an assumption of **rational expectations** becomes equivalent to **perfect foresight**).
123122

124-
(See {cite}`Sargent77hyper` for a rational expectations version of the model when there is uncertainty.)
123+
({cite}`Sargent77hyper` presents a rational expectations version of the model when there is uncertainty.)
125124

126125
Subtracting the demand function {eq}`eq_old1` at time $t$ from the demand
127126
function at $t+1$ gives:
@@ -204,7 +203,7 @@ as it ordinarily would be in the state-space model described in our lecture on
204203
205204
206205
207-
We use form {eq}`eq_old4` because we want to apply an approach described in our lecture on {doc}`Stackelberg problems <dyn_stack>`.
206+
We use form {eq}`eq_old4` because we want to apply an approach described in our lecture on {doc}`Stackelberg plans <dyn_stack>`.
208207
209208
Notice that $\frac{1+\alpha}{\alpha} > 1$ is an eigenvalue of transition matrix $A$ that threatens to destabilize the state-space system.
210209
@@ -224,14 +223,9 @@ $$
224223
U(-\alpha \theta_t) = u_0 + u_1 (-\alpha \theta_t) -\frac{u_2}{2}(-\alpha \theta_t)^2 .
225224
$$ (eq_old5a)
226225
227-
The ``bliss level`` of real balances is $\frac{u_1}{u_2}$ and the inflation rate that attains
228-
it is $-\frac{u_1}{u_2 \alpha}$.
229-
230-
(TO TOM: the first sentece in the next section is very similar to the sentence above.)
231-
232226
## Friedman's Optimal Rate of Deflation
233227
234-
According to {eq}`eq_old5a`, the "bliss level" of real balances is $\frac{u_1}{u_2}$ and the inflation rate that attains it is
228+
According to {eq}`eq_old5a`, the ``bliss level`` of real balances is $\frac{u_1}{u_2}$ and the inflation rate that attains it is
235229
236230
237231
$$
@@ -324,9 +318,10 @@ for all $t \geq 0$.
324318
Values of $V(\bar \mu)$ computed according to formula {eq}`eq:barvdef` for three different values of $\bar \mu$ will play important roles below.
325319
326320
* $V(\mu^{MP})$ is the value of attained by the government in a **Markov perfect equilibrium**
321+
* $V(\mu^R_\infty)$ is the value that a continuation Ramsey planner attains at $t \rightarrow +\infty$
322+
* We shall discover that $V(\mu^R_\infty)$ is the worst continuation value attained along a Ramsey plan
327323
* $V(\mu^{CR})$ is the value of attained by the government in a **constrained to constant $\mu$ equilibrium**
328-
* $V(\mu^R_\infty)$ is the limiting value attained by a continuation Ramsey planner under a Ramsey plan.
329-
* We shall see that $V(\mu^R_\infty)$ is a worst continuation value attained along a Ramsey plan
324+
330325
331326
## Structure
332327
@@ -414,7 +409,7 @@ The models are distinguished by their having either
414409
$\{\mu_t\}_{t=0}^\infty$ once and for all at time $0$
415410
subject to the constraint that $\mu_t = \mu$ for all
416411
$t \geq 0$; or
417-
- A sequence indexed by $t =0, 1, 2, \ldots$ of separate policymakers
412+
- A sequence of distinct policymakers indexed by $t =0, 1, 2, \ldots$
418413
- a time $t$ policymaker chooses $\mu_t$ only and forecasts that future government decisions are unaffected by its choice.
419414
420415
@@ -438,7 +433,7 @@ The relationship between outcomes in the first (Ramsey) timing protocol and th
438433
We'll begin with the timing protocol associated with a Ramsey plan and deploy
439434
an application of what we nickname **dynamic programming squared**.
440435
441-
The nickname refers to the feature that a value satisfying one Bellman equation appears as an argument in a second Bellman equation.
436+
The nickname refers to the feature that a value satisfying one Bellman equation appears as an argument in a value function associated with a second Bellman equation.
442437
443438
Thus, our models have involved two Bellman equations:
444439
@@ -455,7 +450,7 @@ Here we consider a Ramsey planner that chooses
455450
$\{\mu_t, \theta_t\}_{t=0}^\infty$ to maximize {eq}`eq_old7`
456451
subject to the law of motion {eq}`eq_old4`.
457452
458-
We can split this problem into two stages, as in {doc}`Stackelberg problems <dyn_stack>` and {cite}`Ljungqvist2012` Chapter 19.
453+
We can split this problem into two stages, as in the lecture {doc}`Stackelberg plans <dyn_stack>` and {cite}`Ljungqvist2012` Chapter 19.
459454
460455
In the first stage, we take the initial inflation rate $\theta_0$ as given
461456
and solve what looks like an ordinary LQ discounted dynamic programming problem.
@@ -491,7 +486,7 @@ $$
491486
x' = Ax + B\mu
492487
$$
493488
494-
As in {doc}`Stackelberg problems <dyn_stack>`, we can map this problem into a linear-quadratic control problem and deduce an optimal value function $J(x)$.
489+
As in the lecture {doc}`Stackelberg plans <dyn_stack>`, we can map this problem into a linear-quadratic control problem and deduce an optimal value function $J(x)$.
495490
496491
Guessing that $J(x) = - x'Px$ and substituting into the Bellman
497492
equation gives rise to the algebraic matrix Riccati equation:
@@ -698,7 +693,7 @@ about dynamic or time inconsistency.
698693
699694
## Time inconsistency
700695
701-
As discussed in {doc}`Stackelberg problems <dyn_stack>` and {doc}`Optimal taxation with state-contingent debt <opt_tax_recur>`, a continuation Ramsey plan is not a Ramsey plan.
696+
As discussed in {doc}`Stackelberg plans <dyn_stack>` and {doc}`Optimal taxation with state-contingent debt <opt_tax_recur>`, a continuation Ramsey plan is not a Ramsey plan.
702697
703698
This is a concise way of characterizing the time inconsistency of a Ramsey plan.
704699
@@ -1351,11 +1346,10 @@ $$
13511346
\begin{aligned}
13521347
\theta^{CR} & = - \frac{\alpha u_1}{\alpha^2 u_2 + c } \\
13531348
\theta^{MPE} & = - \frac{\alpha u_1}{\alpha^2 u_2 + (1+\alpha)c} \\
1354-
\theta^{MPE} & = - \frac{\alpha u_1}{\alpha^2 u_2 + (1+\alpha)c}
1349+
\theta^{*} & = -\frac{u_1}{u_2 \alpha}
13551350
\end{aligned}
13561351
$$
13571352
1358-
(TO TOM: $\theta^{MPE}$ is repeated in the above equations. Should one of them be $\theta^*$?)
13591353
13601354
13611355
But let's see what happens when we change $c$.
@@ -1376,7 +1370,7 @@ generate_table(clqs, dig=4)
13761370
The above table and figures show how
13771371
changes in $c$ alter $\theta_\infty^R$
13781372
and $\theta_0^R$ as well as $\theta^{CR}$ and $\theta^{MPE}$, but not
1379-
$\theta^*$, again in accord with formulas
1373+
$\theta^*,$ again in accord with formulas
13801374
{eq}`eq:Friedmantheta`, {eq}`eq:muRamseyconstrained`, and {eq}`eq:Markovperfectmu`.
13811375
13821376
Notice that as $c $ gets larger and larger, $\theta_\infty^R, \theta_0^R$
@@ -1529,20 +1523,16 @@ A constrained-to-constant-$\mu$ Ramsey plan is time consistent by constructio
15291523
15301524
### Implausibility of Ramsey Plan
15311525
1532-
In settings in which governments actually choose sequentially, many economists
1533-
regard a time inconsistent plan as implausible because of the incentives to
1534-
deviate that are presented along the plan.
1535-
1536-
(TO TOM: In our meeting, you suggested that we can improve the sentence above.)
1526+
Many economists regard a time inconsistent plan as implausible because they question the plausibility of timing protocol in
1527+
which a plan for setting a sequence of policy variables is chosen once-and-for-all at time $0$.
15371528
1538-
A way to state this reaction is to say that a Ramsey plan is not credible because there are persistent incentives for policymakers to deviate from it.
15391529
15401530
For that reason, the Markov perfect equilibrium concept attracts many
15411531
economists.
15421532
1543-
* A Markov perfect equilibrium plan is constructed to insure that government policymakers who choose sequentially do not want to deviate from it.
1533+
* A Markov perfect equilibrium plan is constructed to insure that a sequence of government policymakers who choose sequentially do not want to deviate from it.
15441534
1545-
The *no incentive to deviate from the plan* property is what makes the Markov perfect equilibrium concept attractive.
1535+
The property of a Markov perfect equilibrium that there is *no incentive to deviate from the plan* makes it attractive.
15461536
15471537
15481538
## Comparison of Equilibrium Values

0 commit comments

Comments
 (0)