首页    期刊浏览 2025年12月24日 星期三
登录注册

文章基本信息

  • 标题:Policymaking under uncertainty: gradualism and robustness.
  • 作者:Barlevy, Gadi
  • 期刊名称:Economic Perspectives
  • 印刷版ISSN:1048-115X
  • 出版年度:2009
  • 期号:June
  • 语种:English
  • 出版社:Federal Reserve Bank of Chicago
  • 摘要:Policymakers are often required to make decisions in the face of uncertainty. For example, they may lack the timely data needed to choose the most appropriate course of action at a given point in time. Alternatively, they may be unable to gauge whether the models they rely on to guide their decisions can account for all of the issues that are relevant to their decisions. These concerns obviously arise in the formulation of monetary policy, where the real-time data relevant for deciding on policy are limited and the macroeconomic models used to guide policy are at best crude simplifications. Not surprisingly, a long-standing practical question for monetary authorities concerns how to adjust their actions given the uncertainty they face.
  • 关键词:Gradualism (Political philosophy);Monetary policy;Uncertainty

Policymaking under uncertainty: gradualism and robustness.


Barlevy, Gadi


Introduction and summary

Policymakers are often required to make decisions in the face of uncertainty. For example, they may lack the timely data needed to choose the most appropriate course of action at a given point in time. Alternatively, they may be unable to gauge whether the models they rely on to guide their decisions can account for all of the issues that are relevant to their decisions. These concerns obviously arise in the formulation of monetary policy, where the real-time data relevant for deciding on policy are limited and the macroeconomic models used to guide policy are at best crude simplifications. Not surprisingly, a long-standing practical question for monetary authorities concerns how to adjust their actions given the uncertainty they face.

The way economists typically model decision-making under uncertainty assumes that policymakers can assign probabilities to the various scenarios they might face. Given these probabilities, they can compute an expected loss for each policy--that is, the expected social cost of the outcomes implied by each policy. The presumption is that policymakers would prefer the policy associated with the smallest expected loss. One of the most influential works on monetary policy under uncertainty based on this approach is Brainard (1967). That paper considered a monetary authority trying to meet some target--for example, an inflation target or an output target. Brainard showed that under certain conditions, policymakers who face uncertainty about their economic environment should react less to news that they are likely to miss their target than policymakers who are fully informed about their environment. This prescription is often referred to as "gradual" policy. Over the years, gradualism has come to be viewed as synonymous with caution. After all, it seems intuitive that if policymakers are unsure about their environment, they should avoid reacting too much to whatever information they do receive, given that they have only limited knowledge about the rest of the environment.

Although minimizing expected loss is a widely used criterion for choosing policy, in some situations it may be difficult for policymakers to assign expected losses to competing policy choices. This is because it is hard to assign probabilities to rare events that offer little historical precedent by which to judge their exact likelihood. For this reason, some economists have considered an alternative approach to policymaking in the face of uncertainty that does not require knowing the probability associated with all possible scenarios. This approach is largely inspired by work on robust control of systems in engineering. Like policymakers, engineers must deal with significant uncertainty--specifically, about the systems they design; thus they are equally concerned with how to account for such uncertainty in their models. Economic applications based on this approach are discussed in a recent book by Hansen and Sargent (2008). The policy recommendations that emerge from this alternative approach are referred to as robust policies, reflecting the fact that this approach favors policies that avoid large losses in all relevant scenarios, regardless of how likely they are. Interestingly, early applications of robust control to monetary policy seemed to contradict the gradualist prescription articulated by Brainard (1967), suggesting that policymakers facing uncertainty should respond more aggressively to news that they are likely to miss their target than policymakers facing no uncertainty. Examples of such findings include Sargent (1999), Giannoni (2002), and Onatski and Stock (2002); their results contradict the conventional wisdom based on Brainard (1967), which may help to explain the tepid response to robust control in some monetary policy circles.

In this article, I argue that aggressiveness is not a general feature of robust control and that the results from early work on robust monetary policy stem from particular features of the economic environments those papers studied. (1) Similarly, gradualism is not a generic feature of the traditional approach to dealing with uncertainty based on minimizing expected losses--a point Brainard (1967) himself was careful to make. I explain that the way policymakers should adjust their response to news that they are likely to miss their target depends on asymmetries in the uncertain environment in which they operate. As we shall see, both the traditional and robust control approaches dictate gradualism in the environment Brainard (1967) considered, while both dictate being more aggressive in other environments that capture elements in the more recent work on robust control.

My article is organized as follows. First, I review Brainard's (1967) original result. Then, I describe the robust control approach, including a discussion of some of its critiques. Next, I apply the robust control approach to a variant of Brainard's model and show that it implies a gradualist prescription for that environment--just as Brainard found when he derived optimal policy, assuming policymakers seek to minimize expected losses. Finally, using simple models that contain some of the features from the early work on robust monetary policy, I show why robustness can recommend aggressive policymaking under some conditions.

Brainard's model and gradualism

I begin my analysis by reviewing Brainard's (1967) model. Brainard considered the problem of a policymaker who wants to target some variable that he can influence so that the variable will equal some prespecified level. For example, suppose the policymaker wants to maintain inflation at some target rate or steer output growth toward its natural rate. Various economic models suggest that monetary policy can affect these variables, at least over short horizons, but that other factors beyond the control of monetary authorities can also influence these variables. Meeting the desired target will thus require the monetary authority to intervene in a way that offsets changes in these factors. Brainard focused on the question of how this intervention should be conducted when the monetary authority is uncertain about the economic environment it faces but can assign probabilities to all possible scenarios it could encounter and acts to minimize expected losses computed using these probabilities.

Formally, let us refer to the variable the monetary authority wants to target as y. Without loss of generality, we can assume the monetary authority wants to target this variable to equal zero. The variable y is affected by a policy variable set by the policymaker, which I denote by r. In addition, y is affected by some variable x that the policymaker can observe prior to setting r. For simplicity, suppose y depends on these two variables linearly; that is,

1) y = x - kr,

where k measures the effect of changes in r on y and is assumed to be positive. For example, y could reflect inflation, and r could reflect the short-term nominal interest rate set by the monetary authority. Equation then implies that raising the nominal interest rate would lower inflation, but that inflation is also determined by other variables, as summarized by x. These variables could include shocks to productivity or the velocity of money. If x rises, the monetary authority would simply have to set r to equal x/k to restore y to its target level of 0.

To incorporate uncertainty into the policymaker's problem, suppose y is also affected by random variables whose values the policymaker does not know, but whose distributions are known to him in advance. Thus, let us replace equation 1 with

2) y = x - (k + [[Epsilon].sub.k])r + [[Epsilon].sub.u],

where [[Epsilon].sub.k] and [[Epsilon].sub.u] are independent random variables with means 0 and variances [[sigma].sup.2.sub.k] and , [[sigma].sup.2.sub.u] respectively. This formulation assumes the policymaker is uncertain both about the effect of his policy, as captured by the [[Epsilon].sub.k]) term that multiplies his choice of r, and about factors that directly affect y, as captured by the additive term [[Epsilon].sub.u]). The optimal policy depends on how much loss the policymaker incurs from missing his target. Suppose the loss is quadratic in the deviation between the actual value of y and its target--that is, the loss is equal to [y.sup.2]. The policymaker will then choose r so as to minimize his expected loss, that is, to solve

3) [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

Brainard (1967) showed that the solution to this problem is given by

4) r = x / k + [[sigma].sup.2.sub.k] / k

Equation 4 is derived in appendix 1. Uncertainty about the effect of policy will lead the policymaker to attenuate his response to x relative to the case where he knows the effect of r on y with certainty. In particular, when [[sigma].sup.2.sub.k] = 0, the policymaker will set r to undo the effect of x by setting r = x/k. But when [[sigma].sup.2.sub.k] > 0, the policy will not fully offset x. This is what is commonly referred to as gradualism: A policymaker who is unsure about the effect of his policy will react less to news about missing the target than he would if he were fully informed. By contrast, the degree of uncertainty about [[Epsilon].sub.u], as captured by [[sigma].sup.2.sub.u], has no effect on policy, as evident from the fact that the optimal rule for r in equation 4 is identical regardless of [[sigma].sup.2.sub.u].

To understand this result, note that the expected loss in equation 3 is essentially the variance of y. Hence, a policy that leads y to be more volatile will be considered undesirable given the objective of solving equation 3. From equation 2, the variance of y is equal to [r.sup.2] [[sigma].sup.2.sub.k] + [[sigma].sup.2.sub.u] , which is increasing in the absolute value of r. An activist (aggressive) policy that uses r to offset nonzero values of x thus implies a more volatile outcome for y, while a passive (gradual) policy that sets r = 0 implies a less volatile outcome for y. This asymmetry introduces a bias toward less activist policies. Even though a less aggressive response to x would cause the policymaker to miss the target on average, he is willing to do so in order to make y less volatile. Absent this asymmetry, there would be no reason to attenuate policy. This explains why uncertainty in [[Epsilon].sub.u] has no effect on policy: It does not involve any asymmetry between being aggressive and being gradual, since neither affects volatility. Although Brainard (1967) was careful to point out that gradualism is an optimal reaction to certain types of uncertainty, his result is sometimes misleadingly cited as a general rule for coping with uncertainty regardless of its nature.

The robust control approach

An important assumption underlying Brainard's (1967) analysis is that the policymaker knows the probability distribution of the variables that he is uncertain about. More recent work on policy under uncertainty is instead motivated by the notion that policymakers may not know what probability to attach to scenarios they are uncertain about. For example, there may not be enough historical data to infer the likelihood of various situations, especially those that have yet to be observed but remain theoretically possible. Without knowing these probabilities, it will be impossible to compute an expected loss for different policy choices as in equation 3. This necessitates an alternative criterion to judge what constitutes a good policy. The robust control approach argues for picking the policy that minimizes the damage that the policy could possibly inflict--that is, the policy under which the largest possible loss across all potential outcomes is smaller than the largest possible loss under any alternative policy. A policy chosen under this criterion is known as a robust policy (or a robust strategy). Such a policy ensures the policymaker will not incur a bigger loss than the unavoidable bare minimum. This rule is often associated with Wald 0950, p. 18), who argued that this approach, known as the minimax (or minmax) rule, is "a reasonable solution of the decision problem when an a priori distribution...does not exist or is unknown." For a discussion of economic applications of robust control as well as related references, refer to Hansen and Sargent (2008).

Before I consider the consequences of adopting the robust control approach for choosing a target as in Brainard (1967), I first consider an example of an application of robust control both to help illustrate what it means for a policy to be robust in this manner and to discuss some of the critiques of this approach. The example is known as the "lost in a forest" problem, which was first posed by Bellman (1956) and which spawned a subsequent literature that is surveyed in Finch and Wetzel (2004). (2) Although this example differs in several respects from typical applications of robust control in economics, it remains an instructive introduction to robust control. I will point out some of these differences throughout my discussion when relevant.

The lost in a forest problem can be described as follows. A hiker treks into a dense forest. He starts his trip from the main road that cuts through the forest, and he travels in a straight line for one mile into the forest. He then lies down to take a nap, but when he wakes up he realizes he forgot which direction he came from. He wishes to return to the road--not necessarily to the point where he started, but anywhere on the road where he can flag down a car and head back to town. He would like to do so using the shortest possible route, which if he knew the location of his starting point would be exactly one mile. But he does not know where the road lies, and because the forest is dense with trees, he cannot see the road from afar. So, he must physically reach the road in order to find it. What strategy should he follow in searching for the road? A geometric description of the problem is provided in box 1, although these details are not essential for following the remainder of this discussion.

Solving this problem requires establishing a criterion by which a strategy can qualify as "best" among all possible strategies. In principle, if the hiker knew his propensity to lie down in a particular orientation relative to the direction he travelled from, he could assign a probability that his starting point could be found in any given direction. In that case, an obvious candidate for the optimal strategy is the one that minimizes the expected distance to reach some point on the road. But most people would be unlikely to know their likelihood of lying down in any particular direction or the odds they don't turn in their sleep. While it might seem tempting to simply treat all locations as equally likely, this effectively amounts to making assumptions on just these likelihoods. It is therefore arguable that we cannot assign probabilities that the starting point lies in any particular direction, implying we cannot calculate an expected travel distance for each strategy and choose an optimum. As an alternative criterion, Bellman (1956) proposed choosing the strategy that minimizes the amount of walking required to ensure reaching the road regardless of where it is located. That is, for any strategy, we can compute the longest distance one would have to walk to make sure he reaches the main road regardless of where it is located. We then pick the strategy for which this distance is shortest. This rule ensures we do not have to walk any more than is absolutely necessary to reach the road. While other criteria have been proposed for the lost in a forest problem, many have found the criterion of walking no more than is absolutely necessary to be intuitively appealing. But this is precisely the robust control approach. The worst-case scenario for any search strategy involves exhaustively searching through every wrong location before reaching the true location. Bellman's suggestion thus amounts to using the strategy whose worst-case scenario requires less walking than the worst-case scenario of any other strategy. In other words, the "best" strategy is the one that minimizes the amount of walking needed to run through the gamut of all possible locations for the hiker's original starting point.

Although Bellman (1956) first proposed this rule as a way of solving the lost in a forest problem, it was Isbell (1957) who derived the strategy that meets this criterion. His solution is presented in box 1. The hiker can ensure he finds the road by walking out one mile and, if he doesn't reach the road, continue walking along the circle of radius one mile around where he woke up. While this strategy ensures finding the road eventually, it turns out that deviating from this scheme in a particular way still ensures finding the road eventually, but with less walking.

Note that in the lost in a forest problem, the set of possibilities the hiker must consider to compute the worst-case scenario is an objective feature of the environment: The main road must lie somewhere along a circle of radius one mile around where the hiker fell asleep (we just don't know exactly where). By contrast, in most economic applications, the region that a decision-maker is uncertain about is not an objective feature of the environment but an artificial construct. In particular, the decision-maker is assumed to contemplate the worst-case scenario from a restricted set of economic models that he believes can capture his environment. This setup has the decision-maker ruling out some models with certainty even as he admits other arbitrarily close models that would be hard to distinguish empirically from those he rejected. Unfortunately, changing the admissible set of models often affects the worst-case scenario and thus the implied policy recommendation. The lost in a forest problem provides a relatively clean motivating example in which we can apply the robust control approach, although this problem obscures important issues that arise in economic applications, such as how to construct the set of scenarios from which a decision-maker calculates the worst case.

As noted previously, many mathematicians regard the robust strategy as a satisfactory solution for the lost in a forest problem. However, this strategy has been criticized in ways that mirror the criticisms of robust control applications in economics. One such critique is that the robust policy is narrowly tailored to do well in particular scenarios rather than in most scenarios. This critique is sometimes described as "perfection being the enemy of the good": The robust strategy is chosen because it does well in the one state of the world that corresponds to the worst-case scenario, even if that state is unlikely and even if the strategy performs much worse than alternative strategies in most if not all remaining states of the world. (3) In the lost in a forest problem, the worst-case scenario for any search strategy involves guessing each and every one of the wrong locations first before finding the road. Arguably, guessing wrong at each possible turn is rather unlikely. But the robust policy is tailored to this scenario, and because of this, the hiker does not take advantage of shortcuts that allow him to search through many locations without having to walk a great distance. As discussed in box 1, such shortcuts exist, but they would involve walking a longer distance if the spot on the road where the hiker started from happened to be situated at the last possible location he tries, and so the robust strategy avoids them. Viewed this way, the robust strategy might seem less appealing.

The problem with this critique is that the lost in a forest problem assumes it is not possible to assign a probability distribution to which direction the nearest point on the road lies. Absent such a distribution, one cannot argue that exhaustively searching through all other paths is an unlikely scenario, since to make this statement precise requires a probability distribution as to where the road is located. One might argue that, even absent an exact probability distribution, we can infer from common experience that we do not often run through all possibilities before we find what we are searching for, so we can view this outcome as remote even without attaching an exact probability to this event. But such intuitive arguments are tricky. Consider the popularity of the adage known as Murphy's Law, which states that whatever can go wrong will go wrong. The fact that people view things going wrong at every turn as a sufficiently common experience to be humorously compared with a scientific law suggests they might not view looking through all of the wrong locations first as such a remote possibility. Moreover, in neither the lost in a forest problem nor many economic applications is it common that the robust strategy performs poorly in all scenarios other than the worst-case one. By continuity, the strategy that is optimal in the worst-case scenario will be approximately optimal in similar situations--for example, exhausting most but not all possible locations before reaching the road in the lost in a forest problem. Such continuity is common to many economic applications. Hence, even if the probability of the worst-case scenario is low, there may be other nearby states that are not as infrequent where the policy remains approximately optimal. In the lost in a forest problem, it also turns out that the robust strategy does well if the road lies in one of the regions to be explored first. But if the road lies in neither the first nor last regions to be explored, the robust strategy involves walking an unnecessarily long distance. Criticizing a policy because it performs poorly in some states imposes an impossible burden on policy. Even a policy designed to do well on average (assuming the distribution of outcomes is known) may perform poorly in certain states of the world.
BOX 1
The lost in a forest problem

The lost in a forest problem formally amounts to
choosing a path starting from an initial point (the hiker's
location when he wakes up) that must ultimately intersect
with an infinite straight line (the road that cuts
through the forest) whose closest distance to the initial
point is one mile. That is, we need to choose a path
starting from the center of a circle of radius one mile
to any point on a particular straight line that is tangent
to this circle. The fact that the hiker forgot where he
came from corresponds to the stipulation that the location
of the tangency point on the circle is unknown.
This situation is illustrated graphically in panel A
of figure B1, which shows three of the continuum
of possible locations for the road.

[FIGURE B1 OMITTED]

Because the forest is dense with trees, the hiker
is assumed not to know where the main road is until
he actually reaches it. Bellman (1956) was the first to
suggest choosing the path that minimizes the longest
distance needed to reach the line with absolute certainty
regardless of the location of the tangency point
on the unit circle. To better appreciate this criterion,
consider the strategy of walking a straight path for
one mile until reaching the circle along which the tangency
point must be located, then travelling counterclockwise
along the circle until reaching the road. This
strategy will reach the road with certainty, and the
longest distance needed to ensure reaching the road
regardless of its location corresponds to walking a mile
plus the entire distance of the circle, that is, 1 + 2[pi]
[approximately equal to] 7.28 miles. But it is possible to
ensure we will reach
the road regardless of its location with an even shorter
path. To see this, suppose we walk out a mile and
proceed to walk counterclockwise along the circle as
before, but after travelling for three-fourths of the
circle, rather than continuing to walk along the circle,
we instead walk straight ahead for a mile, as shown
in panel B of figure B1. This approach also ensures
we will reach the road with certainty regardless of
where it is located, but it involves walking at most
1 + [3/2] [pi] + 1 [approximately equal to] 6.71 miles.
It turns out that it is possible
to do even better than this. The optimal path, derived
by Isbell (1957), is illustrated in panel C of figure B1.
This procedure involves walking out 2/[square root of (3)] = 1.15 miles,
then turning clockwise 60 degrees and walking back
toward the circle for another 1/[square root of (3)] = 0.58 miles until
reaching the perimeter of the circle of radius one mile
around where the hike started, walking 210 degrees
along the circle, and then walking a mile along the
tangent to the circle at this point. The strategy requires
walking at most 1+2[square root of (3)] + [7/6] [pi] +
1 [approximately equal to]  6.40 miles. This is the
absolute minimum one would have to walk and still
ensure reaching the road regardless of its location.

When he originally posed the lost in a forest
problem, Bellman (1956) suggested as an alternative
strategy the path that minimizes the expected distance
of reaching the road, assuming the tangency point was
+distributed uniformly over the unit circle. To the best
of my knowledge, this problem has yet to be solved
analytically. However, Gluss (1961) provided some
intuition as to the nature of this solution by numerically
solving for the optimal path among a parameterized
set of possible strategies. He showed that the robust
path in panel C of figure B1 does not minimize the
expected distance, and he demonstrated various strategies
that improve upon it. The general shape Gluss
found to perform well among the paths he considered
is demonstrated in panel D of figure B1. In order to
minimize the expected distance, it turns out that it will
be better to eventually stray outside the circle rather
than always hewing close to it as Isbell's (1957) path
does. The reason is that one can cover more possibilities
walking outside the circle and reaching the road at a
nontangency point than hewing to the circle and searching
for tangency points. This can be seen in panel D of
figure B1, where walking along the proposed path up
to point B covers all possible locations for the tangency
point in the arc AC, whereas walking along the arc AC
would have required walking a significantly longer
distance. The drawback of straying from the circle this
way is that if the road happens to be located at the last
possible location, the hiker would have to cover a much
greater distance to reach that point. But since the probability
that the road will lie in the last possible location to
be searched is small under the uniform distribution, it
is worth taking this risk if the goal is to minimize the
expected travel time rather than the maximal travel time.


Another critique of robust control holds that, rather than choosing a policy that is robust, decision-makers should act like Bayesians; that is, they should assign subjective beliefs to the various possibilities they contemplate, compute an implied expected loss for each strategy, and then choose the strategy that minimizes the expected loss. For example, Sims (2001) argued decision-makers should avoid rules that violate the sure-thing principle, which holds that if one action is preferred to another action regardless of which event is known to occur, it should remain preferred if the event were unknown. The robust control approach can violate this principle, while subjective expected utility does not. The notion of assigning subjective probabilities to different scenarios is especially compelling in the lost in a forest problem, where assigning equal probabilities to all locations seems natural given there is no information to suggest that any one direction is more likely than the other. In fact, when Bellman (1956) originally posed his question, he suggested both minimizing the longest path (minimax) and minimizing the expected path assuming a uniform prior (min-mean) as ways of solving this problem. This approach need not contradict the policy recommendation that emerges from the robust control approach. In particular, if the cost of effort involved in walking rises steeply with the distance one has to walk, a policy that eliminates the possibility of walking very long distances would naturally emerge as desirable. But at shorter distances, assigning a probability distribution to the location of the road might lead to a different strategy from the robust one. This critique is not aimed at a particular strategy per se, but against using robustness as a criterion for choosing which policy to pursue. (4)

The problem with this critique is that it is not clear that decision-makers would always agree with the recommendation that they assign subjective probabilities to scenarios whose likelihood they do not know. As an example, assigning a distribution to the location of the road in the lost in a forest problem is incompatible with the notion of Murphy's Law. Inherent in Murphy's Law is the notion that the location of the road depends on where the hiker chooses to search. But the Bayesian approach assumes a fixed distribution regardless of what action the hiker chooses. Thus, to a person who finds Murphy's Law appealing, proceeding like a Bayesian would ring false. As another example, consider the "Ellsberg paradox," which is due to Ellsberg (1961). This paradox is based on a thought experiment in which people are asked to choose between a lottery with a known probability of winning and another lottery featuring identical prizes but with an unknown probability of winning. Ellsberg argued that most people would prefer to avoid the lottery whose probability of winning they do not know and would not choose as if they assigned a fixed subjective probability to the lottery with an unknown probability of winning. In other words, the preferences exhibited by most people would seem paradoxical to someone who behaved like a Bayesian. Subsequent researchers who conducted experiments offering these choices to real-life test subjects, starting with Becker and Brownson (1964), confirmed this conjecture. The saliency of these findings suggests that the failure to behave like a Bayesian may reflect genuine discomfort by test subjects with the Bayesian approach of assigning subjective probabilities to outcomes whose probabilities they do not know. But if this is so, we cannot objectively fault decision-makers for not adopting the Bayesian approach, since any recommendation we make to them would have to respect their preferences.

Of course, even accepting that policymakers may not always find the Bayesian approach appealing, it does not automatically follow that they should favor the robust control approach in particular. The relevant question is whether there is a compelling reason for decision-makers to specifically prefer the robust policy. One result often cited by advocates of robust control is the work of Gilboa and Schmeidler (1989). They show that if decision-makers' preferences over lotteries satisfy a particular set of restrictions, it will be possible to represent their choices as if they chose the action that minimizes the largest possible expected loss across a particular set of probability distributions. However, this result is not an entirely satisfactory argument for why policymakers should adopt the robust control approach. First, there is little evidence to suggest that standard preferences obey the various restrictions derived by Gilboa and Schmeidler (1989). While the Ellsberg paradox suggests many people have preferences different from those that would lead them to behave like Bayesians, it does not by itself confirm that preferences accord with each one of the restrictions in Gilboa and Schmeidler (1989). Second, Gilboa and Schmeidler (1989) show that the set of scenarios from which the worst case is calculated depends on the preferences of the decision-makers. This is not equivalent to arguing that policymakers, once they restrict the set of admissible models that could potentially account for the data they observe, should always choose the action that minimizes the worst-case outcome from this set.

In the lost in a forest problem, Gilboa and Schmeidler's (1989) result only tells us that if an individual exhibited particular preferences toward lotteries whose outcomes dictate distances he would have to walk, he would in fact prefer the minimax solution to this problem. It does not say that whenever he faces uncertainty more generally--for example, if he also forgot how far away he was from the road when he lay down--that he would still choose the strategy dictated by the robust control approach. In short, Gilboa and Schmeidler (1989) show that opting for a robust strategy is coherent in that we can find well-posed preferences that rationalize this behavior, but their analysis does not imply such preferences are common or that robustness is a desirable criterion whenever one is at a loss to assign probabilities to various possible scenarios. (5)

The theme that runs through the discussion thus far is that if the decision-makers cannot assign probabilities to scenarios they are uncertain about, there is no inherently correct criterion on how to choose a policy. As Manski (2000, p. 421) put it, "there is no compelling reason why the decision maker should or should not use the maximin rule when [the probability distribution] is a fixed but unknown objective function. In this setting, the appeal of the maximin rule is a personal rather than normative matter. Some decision makers may deem it essential to protect against worstcase scenarios, while others may not." (6) One can point to unappealing elements about robust control, but these do not definitively rule out this approach. Conversely, individuals with particular preferences toward lotteries might behave as if they were following a minimax rule, but this does not imply that they will adopt such a rule whenever they are unable to assign probabilities to possible scenarios. Among engineers, the notion of designing systems that minimize the worst-case scenario among the set of possible states whose exact probability is unknown has carried some appeal. Interestingly, Murphy's Law is also an export from the field of engineering. (7) The two observations may be related: If worst-case outcomes are viewed not as rare events but as common experiences, robustness would naturally seem like an appealing criterion. Policymakers who are nervous about worst-case outcomes would presumably find appeal in the notion of keeping the potential risk exposure to the bare minimum. More generally, studying robust policies can help us to understand the costs and benefits of maximally aggressive risk management so that policymakers can contemplate their desirability.

Recasting the Brainard model as a robust control problem

Now that I have described what it means for a policy to be robust, I can return to the question of how a policymaker concerned about robustness should act when trying to target a variable in an uncertain environment. In particular, I will now revisit the environment that Brainard (1967) considered, but with one key difference: The policymaker is assumed to be unable to assign probabilities to the scenarios he is uncertain about. In what follows, I introduce uncertainty in a way that Hansen and Sargent (2008) and Williams (2008) describe as structured uncertainty; that is, I assume the policymaker knows the model but is uncertain about the exact value of one of its parameters. More precisely, he knows that the parameter lies in some range, but he cannot ascribe a probability distribution to the values within this range. By contrast, unstructured uncertainty corresponds to the case where a model is defined as a probability distribution over outcomes, and where the policymaker is unsure about which probability distribution from some set represents the true distribution from which the data are drawn. (8)

Once again, I begin by assuming that the variable in question, y, is affected linearly by a policy variable, r; various factors that the policymaker can observe prior to setting policy, x; and other factors that the policymaker cannot observe prior to setting his policy but whose distribution is known, [[epsilon].sub.u]:

5) y=x-kr+[[epsilon].sub.u].

As before, I assume [[epsilon].sub.u] has mean 0 and variance [[sigma].sup.2.sub.u]. To capture uncertainty about the effect of policy, I modify the coefficient on r to allow for uncertainty:

6) y=x-(k+[[epsilon].sub.k])r+[[epsilon].sub.u].

In contrast to Brainard's (1967) setup, I assume that rather than knowing the distribution of [[epsilon].sub.k], the policymaker only knows that its support is restricted to the interval [[[epsilon].bar][bar.[epsilon]]] that includes 0, that is, [[epsilon].bar] < 0 < [bar.[epsilon]]. In other words, the effect of r on y can be less than, equal to, or higher than k. Beyond this, he will not be able to assign probabilities to particular values within this interval.

Since the support of [[epsilon].sub.k] will figure prominently in formulating the robust strategy, it is worth commenting on where it might come from. In practice, information about k + [[epsilon].sub.k] is presumably compiled from past data. That is, given time-series data on y, x, and r, we can estimate k + [[epsilon].sub.k] by using standard regression techniques. With a finite history, our estimate would necessarily be noisy due to variation from [[epsilon].sub.u]. However, we might still be able to reject some values for k + [[epsilon].sub.k] as implausible--for example, values that are several standard errors away from our point estimate. Still, there is something seemingly arbitrary in classifying some values of k as possible while treating virtually identical values as impossible. Although this may be consistent with the common practice of "rejecting" hypotheses whose likelihood falls below some set cutoff, it is hard to rationalize such dogmatic rules for including or omitting possible scenarios in constructing worst-case outcomes. In what follows, I will treat the support for [[epsilon].sub.k] as given, sidestepping these concerns. (9)

The robust control approach in this environment can be cast as a two-step process. First, for each value of r, we compute its worst-case scenario over all values [[epsilon].sub.k][member of][[epsilon].bar][bar.[epsilon]], or the largest expected loss the policymaker could incur. Define this expected loss as W(r); that is,

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

Second, we choose the policy r that implies the smallest value for W(r). The robust strategy is defined as the value of r that solves [min.sub.r] W(r); that is,

7)[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

I explicitly solve equation 7 in appendix 2. In what follows, I limit myself to describing the robust strategy and providing some of the intuition behind it. It turns out that the robust policy hinges on the lowest value that [[epsilon].sub.k] can assume. If [[epsilon].bar] < -k, which implies that the coefficient k + [[epsilon].sub.k] can assume either positive or negative values, the solution to equation 7 is given by

8) r = 0.

If instead [[epsilon].bar] > -k, so the policymaker is certain that the coefficient k + [[epsilon].sub.k] is positive (but is still unsure

of its value), the solution to equation 7 is given by

9) r = x/k + ([[epsilon].bar] + [bar.[epsilon]])/2.

Thus, if the policymaker knows the sign of the effect of r on y, he will respond to changes in x in a way that depends on the extreme values [[epsilon].sub.k] can assume, that is, the endpoints of the interval [[[epsilon].bar] + [bar.[epsilon]]]. But if the policymaker is unsure about the sign of the effect of policy on y, he will not respond to changes in x at all.

To better understand why concerns about robustness lead to this rule, consider first the result that if the policymaker is uncertain about the sign of k + [[epsilon].sub.k], he should altogether abstain from responding to x. This is related to Brainard's (1967) original attenuation result: There is an inherent asymmetry in that a passive policy where r = 0 leaves the policymaker unexposed to risk from [[epsilon].sub.k], while a policy that sets r [not equal to] 0 leaves him exposed to such risk. When the policymaker is sufficiently concerned about the risk from [[epsilon].sub.k], which turns out to hinge on whether he knows the sign of the coefficient on r, he is better off resorting to a passive policy that protects him from this risk than trying to offset nonzero values of x. However, the attenuation here is both more extreme and more abrupt than what Brainard found. In Brainard's formulation, the policymaker will always act to offset x, at least in part, but he will moderate his response to x continuously with [[sigma].sup.2.sub.k]. By contrast, robustness considerations imply a threshold level for the lower support of [[epsilon].sub.k], which, if crossed, leads the policymaker to radically shift from actively offsetting x to passively not responding to it at all.

The abrupt shift in policy in response to small changes in [[epsilon].sub.k] demonstrates one of the criticisms of robust control cited earlier--namely, that this approach formulates policy based on how it performs in specific states of the world rather than how it performs in general. When [[epsilon].bar] is close to -k, it turns out that the policymaker is almost indifferent among a large set of policies that achieve roughly the same worst-case loss. When [[epsilon].bar] is just below -k, setting r = 0 performs slightly better under the worst-case scenario than setting r according to equation 9. When [[epsilon].bar] is just above -k, setting r according to equation 9 performs slightly better under the worst-case scenario than setting r = 0. When [[epsilon].bar] is exactly equal to -k, both strategies perform equally well in the worst-case scenario, as does any other value of r. However, the two strategies lead to different payoffs in scenarios other than the worst case, that is, for values of [[epsilon].sub.k] that are between [[epsilon].bar] and [bar.[epsilon]]. Hence, concerns for robustness might advocate dramatic changes in policy to eke out small gains under the worst-case scenario, even if these changes result in substantially larger losses in most other scenarios. A dire pessimist would feel perfectly comfortable guarding against the worst-case scenario in this way. But in situations such as this, where the policymaker chooses his policy based on minor differences in how the policies perform in one particular case even when the policies result in enormous differences in other cases, the robust control approach has a certain tail-wagging-the-dog aspect to it that makes it seem less appealing.

Next, consider what robustness considerations dictate when the policymaker knows the sign of k + [[epsilon].sub.k] but not its precise magnitude. To see why r depends on the endpoints of the interval [[[epsilon].bar] + [bar.[epsilon]]], consider figure 1. This figure depicts the expected loss [[(x -(k + [[epsilon].sub.k])r).sup.2] + [[sigma].sup.2.sub.u]] for a fixed r against different values of [[epsilon].sub.k]. The loss function is quadratic and convex, which implies the largest loss will occur at one of the two extreme values for [[epsilon].sub.k]. Panel A of figure 1 illustrates a case in which the expected losses at [[epsilon].sub.k] = [[epsilon].bar] and [[epsilon].sub.k] = [bar.[epsilon]] are unequal: The expected loss is larger for [[epsilon].sub.k] = [bar.[epsilon]]. But if the losses are unequal under some rule r, that value of r fails to minimize the worst-case scenario. This is because, as illustrated in panel B of figure l, changing r will shift the loss function to the left or the right (it might also change the shape of the loss function, although this can effectively be ignored). The policymaker should thus be able to reduce the largest possible loss over all values of [[epsilon].sub.k] in [[[epsilon].bar], [bar.[epsilon]]]. Although shifting r would lead to a greater loss if [[epsilon].sub.k] happened to equal [[epsilon].bar], since the goal of a robust policy is to reduce the largest possible loss, shifting r in this direction is desirable. Robustness concerns would therefore lead the policymaker to adjust r until the losses at the two extreme values were balanced, that is, until the loss associated with the policy being maximally effective was exactly equal to the loss associated with the policy being minimally effective.

When there is no uncertainty, that is, when [[epsilon].bar] = [bar.[epsilon]] = 0, the policymaker would set r = x/k, since this would set y exactly equal to its target. When there is uncertainty, whether the robust policy will respond to x more or less aggressively than this benchmark depends on how the lower and upper bounds are located relative to 0. If the region of uncertainty is symmetric around 0 so that [[epsilon].bar] = -[bar.[epsilon]], uncertainty has no effect on policy. To see this, note that if we were to set r = x/k, the expected loss would reduce to [(x/k).sup.2][[epsilon].sup.2.sub.k] + [[sigma].sup.2.sub.u], which is symmetric in [[epsilon].sub.k]. Hence, setting r to offset x would naturally balance the loss at the two extremes. But if the region of uncertainty is asymmetric around 0, setting r = x/k would fail to balance the expected losses at the two extremes, and r would have to be adjusted so that it either responds more or less to x than in the case of complete certainty. In particular, the response to x will be attenuated if [bar.[epsilon]] > -[[epsilon].bar], that is, if the potential for an overly powerful stimulus is greater than the potential for an overly weak stimulus, and will be amplified in the opposite scenario.

This result begs the question of when the support for [[epsilon].sub.k] will be symmetric or asymmetric in a particular direction. If the region of uncertainty is constructed using past data on y, x, and r, any asymmetry would have to be driven by differences in detection probabilities across different scenarios--for example, if it is more difficult to detect k when its value is large than when it is small. This may occur if the distribution of [[epsilon].sub.u] were skewed in a particular direction. But if the distribution of [[epsilon].sub.u] were symmetric around 0, policymakers who rely on past data should find it equally difficult to detect deviations in either direction, and the robust policy would likely react to shocks in the same way as if k were known with certainty.

[FIGURE 1 OMITTED]

Deriving the robust strategy in Brainard's (1967) setting reveals two important insights. The first is that the robustness criterion does not inherently imply that policy should be more aggressive in the face of uncertainty. Quite to the contrary, the robust policy exhibits a more extreme form of the same attenuation principle that Brainard demonstrated, for essentially the same reason: The asymmetry between how passive and active possibilities leave the policymaker exposed to risk tends to favor passive policies. More generally, whether facing uncertainty about the economic environment leads to a more gradual policy or a more aggressive policy depends on asymmetries in the underlying environment. If the policymaker entertains the possibility that policy can be far too effective but not that it will be very ineffective, he will naturally tend to attenuate his policy. But if his beliefs are reversed, he will tend to magnify his response to news about potential deviations from the target level. (10)

The second insight is that, at least in some circumstances, the robust strategy can be described as one that balances the losses from different risks. Consider the case where [[epsilon].bar] > -k: In that case, the robust strategy will be the one that equates the loss from the risk of policy being overly effective with the loss from the risk of policy being insufficiently effective. This suggests that, in some cases, robustness amounts to a recommendation of keeping opposing risks in balance, in line with what monetary authorities often cite as the principle that guides their policies in practice. That said, the notion that concerns for robustness amount to balancing losses in this way is not common to all environments. The next section presents an example in which robustness considerations would recommend proceeding as if the policymaker knew the worst-case scenario to be true rather than to keep different risks in balance. Moreover, the notion of balancing losses or risks is implicit in other approaches to modeling decision-making under uncertainty. For example, choosing a policy to minimize expected losses will typically call on the policymaker to equate expected marginal losses across states of the world or to otherwise balance expected costs and benefits of particular policies. Hence, robust control is neither uniquely nor fundamentally a recommendation to balance risks. Nevertheless, in some circumstances it involves balancing opposing risks in a way that mirrors some of the stated objectives of monetary authorities.

Robustness and aggressive rules

Since robustness considerations can lead to policies that are either more gradual or more aggressive, depending on the underlying asymmetry of the environment, it seems natural to ask which asymmetries tended to favor aggressive policies in the original work on robust monetary policy. The papers cited earlier consider different environments, and their results are not driven by one common feature. I now offer two examples inspired by these papers to illustrate some of the relevant asymmetries. The first assumes the policymaker is uncertain about the persistence of shocks, following Sargent (1999). The second assumes the policymaker is uncertain about the trade-off between competing objectives, following Giannoni (2002). As I show next, both of these features could tilt a policymaker who is uncertain about his environment in the direction of overreacting to news about missing a particular target, albeit for different reasons.

Uncertain persistence

One of the first to argue that concerns for robustness could dictate a more aggressive policy under uncertainty than under certainty was Sargent (1999). In discussing a model proposed by Ball (1999), Sargent asked how optimal policy would be affected when we account for the possibility that the model is misspecified--in particular that the specification errors are serially correlated. To gain insight into this question, I adapt the model of trying to meet a target I described earlier to allow for the possibility that the policymaker is uncertain about the persistence of the shocks he faces, and I examine the implied robust policy. I show that there is an asymmetry in the loss from underreacting to very persistent shocks and the loss from overreacting to moderately persistent shocks. Other things being equal, this tilts the policymaker toward reacting more to past observable shocks when he is uncertain about the exact degree of persistence.

Formally, consider a policymaker who wants to target a variable that is affected by both policy and other factors. Although Sargent (1999) considers a model in which the policymaker is concerned about multiple variables, it will be simpler to assume there is only one variable he cares about. Let [y.sub.t] denote the value at date t of the variable that the policymaker wishes to target to 0. As in equation 5, I assume [y.sub.t] is linear in the policy variable [r.sub.t] and in an exogenous shock term [x.sub.t]:

10) [y.sub.t] = [x.sub.t] - [kr.sub.t].

Here, I no longer assume the policymaker is uncertain about the effect of his policy on y. As such, it will be convenient to normalize k to 1. However, I now assume he is uncertain about the way [x.sub.t] is correlated over time. Suppose

11) [x.sub.t] = [rho][x.sub.t-1] + [[epsilon].sub.t],

where [[epsilon].sub.t] are independent and identically distributed over time with mean 0 and variance [[sigma].sup.2.sub.[epsilon]]. At each date t, the policymaker can observe [x.sub.t-1] and condition his policy on its realization. However, he must set [r.sub.t] before observing [x.sub.t]. He will be uncertain about [x.sub.t] for two reasons: He must act before getting to observe [[epsilon].sub.t], and he may not know the value of p with certainty.

I assume the policymaker discounts future losses at rate [beta] < 1 so that his expected loss is given by

E[[infinity].summation over (t=0)][[beta].sup.t][y.sup.2.sub.t]] = E [infinity].summation over (t=0)][[beta].sup.t][([x.sub.t] - [r.sub.t].sup.2])].

If the policymaker knew 9 with certainty, his optimal strategy would be to set [r.sub.t] = [rho][x.sub.t-1], which is the expected value of [x.sub.t+1]. Suppose instead that be knew 0 fell in some interval [[rho].bar][bar.[rho]]. Let [[rho].sup.*] denote the midpoint of this interval; that is.

[[rho].sup.*] = ([[rho].bar] + [bar.[rho]])/2.

To emphasize asymmetries inherent to the loss function as opposed to the region of uncertainty, suppose the interval of uncertainty is symmetric around the certainty benchmark; that is, in assessing whether the robust policy is more aggressive, we will compare it to the policy the monetary authority would pursue if it knew [rho] = [[rho].sup.*]. An important and empirically plausible assumption in what follows is that [[rho].sup.*] > 0; that is, the beliefs of the monetary authority are centered around the possibility that shocks are positively correlated.

Once again, we can derive the robust strategy in two steps. First, for each rule [r.sub.t], define W([r.sub.t]) as the biggest loss possible among the different values of [rho]; that is,

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII].

We then choose the policy rule [r.sub.t] that minimizes W([r.sub.t]); that is, we solve

12)[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII].

Following Sargent (1999), I assume the policymaker is restricted in the type of policies [r.sub.t] he can carry out: The policymaker must choose a rule of the form [r.sub.t] = [ax.sub.t-1], where a is a constant that cannot vary over time. This restriction is meant to capture the notion that the policymaker cannot learn about the parameters about which he is uncertain and then change the way policy reacts to information as be observes [x.sub.t] over time and potentially infers [rho]. I further assume the expectation in equation 12 is the unconditional expectation of future losses; that is, the policymaker calculates his expected loss from the perspective of date 0. To simplify the calculations, I assume [x.sub.0] is drawn from the stationary distribution for [x.sub.t].

The solution to equation 12, subject to the constraint that [r.sub.t] = [ax.sub.t-1], is derived in appendix 3. The key result shown in that appendix is that as long as [[rho].sup.*] > 0, the robust policy would set a to a value in the interval [[[rho].bar], [bar.[rho]]] that is strictly greater than the midpoint [[rho].sup.*]. In other words, starting with the case in which the policymaker knows [rho] = [[rho].sup.*], if we introduce a little bit of uncertainty in a symmetric fashion, so the degree of persistence can deviate equally in either direction, the robust policy would react more to a change in [x.sub.t-1] in the face of uncertainty than it would react to such a change if the degree of persistence were known with certainty.

To understand this result, suppose the policymaker instead set [rho] = [[rho].sup.*]. As in the previous section, the loss function is convex in [rho], so the worst-case scenario will occur when [rho] assumes one of its two extreme values, that is, either when [rho] = [bar.[rho]] or [rho] = [[rho].bar]. It turns out that when a = [[rho].sup.*], setting [rho] = [bar.[rho]] imposes a bigger cost on the policymaker than setting [rho] = [rho]. Intuitively, for any given [rho], setting a = [[rho].sup.*] will imply [y.sub.t] = ([rho] - [[rho].sup.*])[x.sub.t-1] + [[epsilon].sub.t]. The expected deviation of [y.sub.t] from its target given [x.sub.t-1] will have the same expected magnitude in both cases; that is, |([rho] - [[rho].sup.*])[x.sub.t-1]| will be the same when [rho] = [bar.[rho]] and [rho] = [[rho].bar], given [[[rho].bar], [bar.[rho]]] is symmetric around [[rho].sup.*]. However, the process [x.sub.t] will be more persistent when [rho] is higher, and so deviations from the target will be more persistent when [rho] = [bar.[rho]] than when [rho] = [bar.[rho]]. More persistent deviations imply more volatile y, and hence a larger expected loss. Since the robust policy tries to balance the losses at the two extreme values of [rho], the policymaker should choose a higher value for a to reduce the loss when [rho] = [bar.[rho]].

The basic insight is that, while the loss function for the policymaker is symmetric in p around [rho] = 0, if we focus on an interval that is centered in either direction of [rho] = 0, the loss function will be asymmetric. This asymmetry will tend to favor policies that react more to past shocks. In fact, this feature is not unique to policies guided by the robustness criterion: If we assumed the policymaker assigned symmetric probabilities to values of [rho] in the interval [[[rho].bar],[bar.[rho]] and acted to minimize expected losses, the asymmetry in the loss function would tilt his policy toward being more aggressive; that is, the value of a that solves [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII] would also exceed [[rho].sup.*] if [[rho].sup.*] > O.

While my example highlights a force that generally favors more aggressive policies, it should be emphasized that my exercise is not quite equivalent to the one in Sargent (1999). Sargent allowed the policymaker to entertain stochastic processes that are correlated as in my example, but he also allowed the policymaker to entertain the possibility that the mean of the process is different from zero. In addition, the region of uncertainty Sargent posited was not required to be symmetric around the certainty benchmark; rather, it was constructed based on which types of processes are easier to distinguish from the certainty benchmark case. The analysis here reveals an asymmetry in the loss function that magnifies concerns about more persistent processes and thus encourages reacting more to past shocks, but the nature of the robust policy depends crucially on the set of scenarios from which the worst-case is constructed.

Uncertain trade-off parameters

Following Sargent (1999), other papers also argued that robust policies tended to be aggressive. While these papers reached the same conclusion as Sargent (1999), they considered different environments. For example, Giannoni (2002) assumed that the policymakers know the persistence of shocks but are uncertain about parameters of the economic model that dictate the effect of these shocks on other economic variables. This leaves policymakers uncertain about the trade-off between their competing objectives. In this environment, Giannoni too found that the robust policy is more aggressive in the face of uncertainty than when policymakers know the trade-off parameters with certainty.

To provide some insight behind these results, consider the following simplified version of Giarmoni's (2002) model. Suppose the monetary authority cares about two variables, denoted y and n; in Giannoni's model, y and n are the output gap and inflation, respectively. The monetary authority has a quadratic loss function:

13) [alpha][y.sup.2] + [[pi].sup.2].

The variables y and n are related linearly

14) [pi] = [lambda]y + x,

where x is an observable shock. (11) This relationship implies a trade-off between [pi] and y. If we set n = 0 as desired, then y would vary with x and deviate from 0. If we set y = 0, then [pi] would vary with x and deviate from 0. Substituting equation 14 into the loss function allows us to express the policymaker's problem as choosing [pi] to minimize the loss

[alpha][([pi] - x).sup.2]/[[lambda].sup.2] + [[pi].sup.2].

Taking the first-order condition with respect to [pi] gives us the optimal choices for y and [pi] as

15) [pi] = [alpha]x/[alpha] + [[lambda].sup.2], y = [lambda]x/[alpha] + [[lambda].sup.2].

Suppose the policymaker were uncertain about [lambda], knowing only that it lies in some interval [[lambda].bar],[bar.[lambda]]. Given a choice of [pi], the worst-case scenario over this range of [lambda] is given by

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII].

The worst case always corresponds to [lambda] = [[lambda].bar], except when [pi] = x, in which case the value of [lambda] has no effect on the loss function. The robust strategy, therefore, is to set [pi] and y to their values in equation 15 as if the policymaker knew [lambda] = [[lambda].bar], the lowest value [lambda] can assume. Thus, as long as the certainty benchmark [lambda] lies in the interior of the uncertainty interval [[[lambda].bar],[bar.[lambda]]], concerns for robustness will lead the policymaker to have y respond less to x, as well as [pi] respond more to x. The robustness criterion leads the policymaker to stabilize y more aggressively against shocks to x and stabilize n less aggressively against these same shocks. The reason is that when a shock x causes [pi] to deviate from its target, a lower value of [lambda] implies that pushing [pi] back to its target would require y to deviate from its target by a greater amount. The worst-case scenario is if y deviates to the largest extent possible, and so the robust policy advocates stabilizing y more aggressively while loosening up on [pi].

Figure 2 illustrates this result graphically. The problem facing the policymaker is to choose a point from the line given by [pi] = [lambda]y + x. Ideally, it would like to move toward the origin, where [pi] = y = 0. Changing [lambda] will rotate the line from which the policymaker must choose as depicted in the figure. A lower value of [lambda] corresponds to a flatter curve. Given the policymaker prefers to be close to the origin, a flatter curve leaves the policymaker with distinctly worse options that are farther from the origin, since one can show that the policymaker would only choose points in the upper left quadrant of the figure. This explains why the worst-case scenario corresponds to the flattest curve possible. If we assume the policymaker must choose his relative position on the line before knowing the slope of the line (that is, before knowing [lambda]), then the flatter the line could be, the greater his incentive will be to locate close to the x-axis rather than risk deviating from his target on both variables, as indicated by the path with the arrow. This corresponds to more aggressively insulating y from x.

Robustness concerns thus encourage the policymaker to proceed as if he knew [lambda], was equal to its lowest possible value. Note the difference from the two earlier models, in which the robust policy recommended balancing losses associated with two opposing risks. Here, by contrast, the policy equates two marginal losses (the loss from letting y deviate a little more from its target and the loss from letting [pi] deviate a little more) for a particular risk, namely, that [lambda], will be low. Thus, robustness does not in general amount to balancing losses from different risks as in my two previous examples.

Conclusion

In recent years, economists have paid increasing attention to the problem of formulating policy under uncertainty, particularly when it is not possible to attach probabilities to the scenarios that concern policymakers. One recommendation for policy in these circumstances, often attributed to Wald (1950), is the robust control approach, which argues for choosing the policy that achieves the most favorable worst-case outcome. Recent work has applied this notion to economic questions, especially in dynamic environments. However, there seem to be only limited references to this literature in monetary policy circles.

One reason for the limited impact of this literature appears to be that early applications of robust control to monetary problems focused on applications in which policymakers should amplify rather than attenuate their responses to shocks, in contrast with the theme emphasized by Brainard (1967) in his model. One of the points of my article is that comparing these applications of robust control to Brainard's result is somewhat misleading, since the policymaker is guarding against different types of uncertainty in the two models. Brainard examined the case of a policymaker who was unsure as to the effect of his policy on the variable he wanted to target; the policymaker found that attenuating his response to shocks would be optimal. But applying a robustness criterion in the same environment would suggest attenuating policy even more drastically, not responding at all to shocks when the range of possibilities that the policymaker is uncertain about is large.

[FIGURE 2 OMITTED]

By contrast, early applications of robust control were concerned with uncertainty over the persistence of shocks or parameters that govern the trade-off between conflicting policy objectives. In that case, robustness concerns suggest amplifying the response of policy to shocks. But so would the expected utility criterion that Brainard considered. Whether the policymaker should change his policies in the face of uncertainty depends on the nature of that uncertainty--specifically whether it involves any inherent asymmetries that would tilt policy from its benchmark when the policymaker is fully informed. These considerations can be as important as the criterion by which a policy is judged to be optimal.

Ultimately, whether policymakers will find the robust control approach appealing depends on their preferences and on the particular application at hand. If policymakers are pessimistic and find an appeal in the dictum of Murphy's Law, which holds that things that can go wrong quite often do go wrong, they will find minimizing the worst-case scenario appealing. But in an economic environment where monetary policy has little impact on the worst-case scenario and has substantial impact on other scenarios, as in one of the examples presented here, even relatively pessimistic policymakers might find alternative approaches to robust control preferable for guiding the formulation of monetary policy.

APPENDIX 1. DERIVING THE OPTIMAL RULE IN BRAINARD'S (1967) MODEL

This appendix derives the optimal policy that solves equation 3 (on p. 39). Using the fact that d/dr E[[y.sup.2]] = E [d/dr [y.sup.2]], we can write the first-order condition for the problem in equation 3 as E[2y X dy/dr] = 0. This implies

E [-2 (x - (k + [[epsilon].sub.k]) r + [[epsilon].sub.u]) (k + [[epsilon].sub.k])] = 0.

Expanding the term inside the expectation operator, we get

E [2(-x (k + [[epsilon].sub.k]) + [(k + [[epsilon].sub.k]).sup.2]r - [[epsilon].sub.k] (k + [[epsilon].sub.k]))] = 0.

Using the fact that E[[[epsilon].sub.k]] = E[[[epsilon].sub.u]] = 0 and that [[epsilon].sub.k] and [[epsilon].sub.u] are independent so E [[[epsilon].sub.k] [[epsilon].sub.u]] = E [[[epsilon].sub.u] E [[[epsilon].sub.k]] = 0, the preceding equation reduces to

- 2xk + 2r([k.sup.2] + [[sigma].sup.2.sub.k])= 0.

Rearranging and setting this expression to 0 yields the value of r described in the text (on p. 39):

r = x/k+[[sigma].sup.2.sub.k]/k.

Note that dr/dx is decreasing in [[sigma].sup.2.sub.k]); that is, greater uncertainty leads the policymaker to attenuate his response to shocks.

APPENDIX 2. DERIVING THE ROBUST STRATEGY IN BRAINARD'S (1967) MODEL

This appendix derives the policy that solves equation 7 (on p. 46), that is,

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

For ease of exposition, let us rewrite this problem as

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

where

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

Since the second derivative of (x - [(k + [[epsilon].sub.k])r).sup.2] with respect to [[epsilon].sub.k] is just 2[r.sup.2] [greater than or equal to] 0, the function W(r) is convex in [[epsilon].sub.k]. It follows that the maximum value must occur at one of the two endpoints of the support, that is, either when [[epsilon].sub.k] = [[epsilon].bar] or when [[epsilon].sub.k] = [bar.[epsilon]].

Next, I argue that if [r.sup.*] solves equation 7, then we can assume without loss of generality that W([r.sup.*]) takes on the same value when [[epsilon].sub.k] = [[epsilon].bar] as when [[epsilon].sub.k] = [[epsilon].bar]. Suppose instead that at the value [r.sup.*] that solves equation 7, W([r.sup.*]) is unequal at these two values. Let us begin first with the case where [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]. This implies

[(x-(k + [[epsilon].bar]) [r.sup.*]).sup.2] > [(x-(k + [bar.[epsilon]]) [r.sup.*]).sup.2].

I obtain a contradiction by establishing there exists an r [not equal to] [r.sup.*] that achieves a lower value of W(r) than W([r.sup.*])= [(x -(k + [[epsilon].bar])[r.sup.*]).sup.2] + [[sigma].sup.2.sub.u]. lf this is true, then [r.sup.*] could hot have been the solution to equation 7, since the solution requires that W([r.sup.*]) [less than or equal to] W(r) for all r.

Differentiate [(x -(k + [[epsilon].bar])r).sup.2] with respect to r and evaluate this derivative at r = [r.sup.*]. If this derivative, 2(k+[[epsilon].bar]_)((k+[[epsilon].bar])[r.sup.*]-x), is different from 0, then we can change r in a way that lowers [(x -(k + [[epsilon].bar])r).sup.2]. By continuity, we can find an r close enough to [r.sup.*] to ensure that

[(x-(k + [[epsilon].bar])r).sup.2] > [(x-(k + [bar.[epsilon]])r).sup.2].

It follows that there exists an r [not equal to] [r.sup.*] such that W(r) = [(x-(k + [[epsilon].bar])r).sup.2] + [[sigma].sup.2.sub.u] < W([r.sup.*]); this is a contradiction. This leaves us with the case where 2(k + [[epsilon].bar]) x ((k + [[epsilon].bar])[r.sup.*]-x) is equal to 0. If (k + [[epsilon].bar])[r.sup.*] = x, then we have

0 = [(x-(k + [[epsilon].bar])r).sup.2] > [(x-(k + [bar.[epsilon]])r).sup.2] [greater than or equal to] 0,

which once again is a contradiction. The last remaining case involves k + [[epsilon].bar] = 0. In that case, W(r) = [chi square]. But we can achieve this value by setting r = 0, and since W(0) does not depend on [[epsilon].sub.k], the statement follows trivially. That is, when k + [[epsilon].bar] = 0 and [[epsilon].sub.k] = [[epsilon].bar] solves the maximization problem that underlies W(r), there will be multiple solutions to equation 7, including one that satisfies the desired property.

The case where [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII] leads to a similar result, without the complication that k + [bar.[epsilon]] might vanish to 0.

Equating the losses at [[epsilon].sub.k] = [[epsilon].bar] and [[epsilon].sub.k] = [bar.[epsilon]] we have

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

The roots of this quadratic equation are r = 0 and r = x/k + ([[epsilon].bar] + [bar.[epsilon]])/2. Substituting in each of these two values yields

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

Define

[PHI] = [bar.[epsilon]] - [epsilon]/2k+[[epsilon].bar]+ [bar.[epsilon]]

Of the two candidate values for r, the one where r = 0 will minimize W(r) if [absolute value of [PHI]] > 1 and the one where

r = x/k+([[epsilon].bar]+[bar.[epsilon]])/2 will minimize W(r) if [absolute value of [PHI][ < 1.

Since [[epsilon].bar] < 0 < [bar.[epsilon]] the numerator for [PHI] is always positive. Hence, [PHI] will be negative if and only if 2k + [[epsilon].bar] + [bar.[epsilon]] < 0. But if 2k + [[epsilon].bar] + [bar.[epsilon]] < 0, then [PHI] < -1. The solution to equation 7 will thus set r = 0 when [PHI] < 0. Since [bar.[epsilon]] > 0, a necessary condition for [PHI] < 0 is for

[[epsilon].bar] -2k.

The only case in which the optimal policy will set r [not equal to] 0 is if 0 [less than or equal to] [PHI] < 1. This in turn will only be true if 2k + [[epsilon].bar] + [bar.[epsilon]] > [bar.[epsilon]] [[epsilon].bar], which reduces to

[[epsilon].bar] > -k.

This implies the robust strategy will set r = 0 whenever [[epsilon].bar] > -k but not otherwise.

APPENDIX 3. DERIVING THE ROBUST RULE WITH UNCERTAIN PERSISTENCE

This appendix derives the policy that solves equation 12 (on p. 49). Substituting in for [y.sub.t], [x.sub.t], and [r.sub.t] = [ax.sub.t-1], we can rewrite equation 12 as

A1) [MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

Since [x.sub.0] is assumed to be drawn from the stationary distribution, the unconditional distribution of [x.sub.t] is the same for all t. In addition, [x.sub.t-1] and [[epsilon].sub.t] are independent. Hence, equation A 1 can be rewritten as

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

or just

[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]

By differentiating the function [([rho] - a/1 - [[rho].sup.2]).sup.2], one can show that it is convex in [rho] for [rho] [member of] [-1,1] and a [member of] [-1,1]. Hence, the biggest loss will occur at the extreme values, that is, when [rho] is equal to either [[rho].bar] or [bar.[rho]]. By the same argument as in appendix 2, the robust strategy must set a in order to equate the losses at these two extremes, which are given by [([bar.[rho]] - a/1 - [[bar.[rho]].sup.2]).sup.2] when [rho] = [bar.[rho]] and to [([[rho].bar]] - a.sup.2]/1 - [[[rho].bar].sup.2] when [rho] = [[rho].bar]. If we equate these two expressions and rearrange, the condition a would have to satisfy

A2) [([[rho].bar]/ [bar.[rho]]).sup.2] = 1 - [[bar.[rho]].sup.2]/1 - [[[rho].bar].sup.2].

Given that [[rho].bar] and [bar.[rho]] are symmetric around [[rho].sup.*] > 0, the expression on the right-hand side of equation A2 is less than 1. Then the left-hand side of equation A2 must be less than l as well. Using the fact that a [member of] [[[rho].bar], [bar.[rho]], it follows that this in turn requires

[bar.[rho]] - a < a - [[rho].bar]

or, upon rearranging,

a > ([[rho].bar] + [bar.[rho]])/2 = [[rho].sup.*]

By contrast, if the policymaker knew [rho] = [[rho].sup.*] with certainty, he would set a = [[rho].sup.*]. The policymaker thus responds more aggressively to past shocks with uncertainty than if he knew [rho] was equal to the midpoint of the set with certainty.

Finally, note that if [[rho].sup.*] = 0, the symmetry requirement would imply [[rho].bar] = -[bar.[rho]], in which case the solution to equation A2 would imply a = 0 = [[rho].sup.*]. Hence, the aggressive response stems from the fact that [[rho].sup.*] is assumed to be strictly positive; that is, the region of uncertainty is parameterized to be asymmetric with respect to no autocorrelation. This asymmetry drives the result that the robust strategy is more aggressive under uncertainty than under certainty.

REFERENCES

Ball, Laurence, 1999, "Policy rules for open economies," Monetary Policy Rules, John Taylor (ed.), National Bureau of Economic Research Studies in Business Cycles, Vol. 31, Chicago: University of Chicago Press, pp. 127-144.

Becker, Selwyn, and Fred Brownson, 1964, "What price ambiguity? Or the role of ambiguity in decisionmaking," Journal of Political Economy, Vol. 72, No. 1, February, pp. 62-73.

Bellman, Richard, 1956, "Minimization problem," Bulletin of the American Mathematical Society, Vol. 62, No. 3, p. 270.

Bernanke, Ben, 2007, "Monetary policy under uncertainty," speech at the 32nd Annual Economic Policy Conference, Federal Reserve Bank of St. Louis, October 19, available at www.federalreserve.gov/ newsevents/speech/bernanke20071019a.htm.

Brainard, William, 1967, "Uncertainty and the effectiveness of policy," American Economic Review, Vol. 57, No. 2, May, pp. 411-425.

Ellsberg, Daniel, 1961, "Risk, ambiguity, and the Savage axioms," Quarterly Journal of Economics, Vol. 75, No. 4, November, pp. 643-669.

Epstein, Larry, and Martin Schneider, 2003, "Recursive multiple-priors," Journal of Economic Theory, Vol. 113, No. 1, November, pp. 1-31.

Finch, Steven, and John Wetzel, 2004, "Lost in a forest," American Mathematical Monthly, Vol. 111, No. 8, October, pp. 645-654.

Giannoni, Mare, 2002, "Does model uncertainty justify caution? Robust optimal monetary policy in a forward-looking model," Macroeconomic Dynamics, Vol. 6, No. 1, February, pp. 111-144.

Gilboa, Itzhak, and David Schmeidler, 1989, "Maxmin expected utility with nonunique prior," Journal of Mathematical Economics, Vol. 18, No. 2, April, pp. 141-153.

Gluss, Brian, 1961, "An alternative solution to the 'lost at sea' problem," Naval Research Logistics Quarterly, Vol. 8, No. 1, March, pp. 117-121.

Hansen, Lars, and Thomas Sargent, 2008, Robustness, Princeton, NJ: Princeton University Press.

Isbell, John, 1957, "An optimal search pattern," Naval Research Logistics Quarterly, Vol. 4, No. 4, December, pp. 357-359.

Maccheroni, Fabio, Massimo Marinacei, and Aldo Rustichini, 2006, "Dynamic variational preferences," Journal of Economic Theory, Vol. 128, No. 1, May, pp. 4-44.

Manski, Charles, 2000, "Identification problems and decisions under ambiguity: Empirical analysis of treatment response and normative analysis of treatment choice," Journal of Econometrics, Vol. 95, No. 2, April, pp. 415-442.

Onatski, Alexi, and James Stock, 2002, "Robust monetary policy under model uncertainty in a small model of the U.S. economy," Macroeconomic Dynamics, Vol. 6, No. 1, February, pp. 85-110.

Rustem, Berc, Volker Wieland, and Stan Zakovic, 2007, "Stochastic optimization and worst-case analysis in monetary policy design," Computational Economics, Vol. 30, No. 4, November, pp. 329-347.

Sargent, Thomas, 1999, "Comment on 'Policy rules for open economies' by Laurence Ball," in Monetary Policy Rules, John Taylor (ed.), National Bureau of Economic Research Studies in Business Cycles, Vol. 31, Chicago: University of Chicago Press, pp. 144-154.

Sims, Christopher, 2001, "Pitfalls of a minimax approach to model uncertainty," American Economic Review, Vol. 91, No. 2, May, pp. 51-54.

Spark, Nick, 2006, A History of Murphy's Law, Los Angeles: Periscope Film.

Svensson, Lars, 2007, "Robust control made simple: Lecture notes," Princeton University, February.

Wald, Abraham, 1950, Statistical Decision Functions, New York: Wiley.

Williams, Noah, 2008, "Robust control," in The New Palgrave Dictionary of Economics, Steven Durlauf and Lawrence Blume (eds.), 2nd ed., 8 vols., New York: Palgrave Macmillan.

NOTES

(1) Onatski and Stock (2002) also argue that aggressiveness is not a generic feature of robust policies, but they note that in their framework, aggressiveness will arise for many classes of perturbations.

(2) The problem is sometimes referred to as the "lost at sea" problem. Richard Bellman, who originally posed the problem, is well-known among economists for his work on dynamic programming, a tool used by economists to analyze sequential decision problems that unfold over time. I refer to this example only because of its intuitive appeal and not to draw any analogies between the practice of conducting monetary policy and the situation of being lost.

(3) An example of this critique can be found in Svensson (2007), who writes "if a Bayesian prior probability measure were to be assigned to the feasible set of models, one might find that the probability assigned to the models on the boundary are exceedingly small. Thus, highly unlikely models can come to dominate the outcome of robust control." Bernanke (2007) also refers to (but does not invoke) this critique of robust control.

(4) In this spirit, Sims (2001) suggested using robust control to aid Bayesian decision-makers who for convenience rely on procedural rules rather than explicit optimization. The idea is to figure out under what prior beliefs about the models it would be optimal to pursue the robust strategy, and then reflect on whether this prior distribution seems sensible. As Sims notes, deriving the robust strategy "may alert decision-makers to forms of prior that, on reflection, do not seem far from what they actually might believe, yet imply decisions very different from that arrived at by other simple procedures." Interestingly, Sims' prescription cannot be applied to the lost in a forest problem: There is no distribution over the location of the road for which the minimax path minimizes expected distance. However, an analogous argument can be made regarding the cost of effort needed to walk various distances. If the hiker relies on procedural rules, he can back out how steeply the cost of effort must rise with distance to justify the minimax path. This will alert him to paths that are different from recommendations arrived at by other simple procedures but rely on plausible cost of effort functions.

(5) Strictly speaking, Gilboa and Schmeidler (1989) only consider a static one-shot decision, while the lost in a forest problem is a dynamic problem in which the hiker must constantly choose how to search given the results of his search at each point in time. However, the problem can be represented as a static decision in which the hiker chooses his search algorithm before starting to search. This is because he would never choose to revise his plans as a result of what he learns; if he did, he could have designed his search algorithm that way before starting to search. This will not be true in many economic applications, where there may be problems with time inconsistency that complicate the task of how to extend the minimax notion to such choices. For a discussion of axiomatic representations of minimax behavior in dynamic environments, see Epstein and Schneider (2003) and Maccheroni, Marinacci, and Rustichini (2006).

(6) For the purposes of this article, the terms "minimax rule" and "robust strategy" can be viewed as interchangeable with the term "maximin rule" that Manski (2000) uses.

(7) According to Spark (2006), the law is named after aerospace engineer Edward Murphy, who complained after a technician attached a pair of sensors in a precisely incorrect configuration during a crash test Murphy was observing. Engineers on the team Murphy was working with began referring to the notion that things will inevitably go wrong as Murphy's Law, and the expression gained public notoriety after one of the engineers used it in a press conference.

(8) For example, the set of distributions the policymaker is allowed to entertain might be those whose relative entropy to the distribution of some benchmark model falls below some threshold. There are other papers that similarly focus on structured uncertainty because of its simplicity--for example, Svensson (2007).

(9) One way to avoid this issue is to model concerns for robustness in a different way. In particular, rather than restrict the set of scenarios policymakers can entertain to some set, we can allow the set to be unrestricted but introduce a penalty function that punishes scenarios continuously depending on how much they differ from some benchmark scenario--for example, the point estimates from an empirical regression. This formulation is known as multiplier preferences, since the penalty function is scaled by a multiplicative parameter that captures concern for model misspecification. See Hansen and Sargent (2008) for a more detailed discussion.

(10) Rustem, Wieland, and Zakovic (2007) also consider robust control in asymmetric models, although they do not discuss the implications of this for gradual policy versus aggressive policy.

(11) This relationship is a simplistic representation of the New Keynesian Phillips curve Giannoni (2002) uses, in which inflation at date t depends on the output gap at date t, expected inflation at date t + 1, and a shock term; that is, n = [[pi]sub.1], + [lambda][y.sub.1] + [beta][E.sub.1][[pi].sub.1+1]. + [x.sub.1] If [x.sub.1] is independent and identically distributed over time, expected inflation would just enter as a constant, and the results would be identical to those in the simplified model I use.

Gadi Barlevy is a senior economist and economic advisor in the Economic Research Department at the Federal Reserve Bank of Chicago. He thanks Marco Bassetto, Lars Hansen, Charles Manski, and Anna Paulson for their comments on earlier drafts.
联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有