首页    期刊浏览 2024年09月21日 星期六
登录注册

文章基本信息

  • 标题:Policy evaluation in a time of austerity: introduction.
  • 作者:Bryson, Alex ; Dorsett, Richard ; Portes, Jonathan
  • 期刊名称:National Institute Economic Review
  • 印刷版ISSN:0027-9501
  • 出版年度:2012
  • 期号:January
  • 语种:English
  • 出版社:National Institute of Economic and Social Research
  • 摘要:The approach of the new government appears somewhat different. Although the Treasury has maintained its public commitment to evidence-based policymaking, research and evaluation budgets have been severely squeezed. Faced with a choice between service delivery and policy evaluation, it may be tempting to regard evaluation as a luxury that government can ill afford, at least in the short term.
  • 关键词:Education

Policy evaluation in a time of austerity: introduction.


Bryson, Alex ; Dorsett, Richard ; Portes, Jonathan 等


As Phil Davies (pp. R41-52) notes, the last government made a series of high-profile commitments to the idea of evidence-based policy. Central to this philosophy was a culture of evaluation whereby potential and actual policies were scrutinised and appraised in order to build and refine a knowledge base of 'what works'. While this was far from consistently applied, and the results of evaluations were not always comfortable reading for the government (for example, the very mixed results from the National Evaluation of Sure Start), on the whole there was a presumption that major social policy initiatives should be subject to independent evaluation.

The approach of the new government appears somewhat different. Although the Treasury has maintained its public commitment to evidence-based policymaking, research and evaluation budgets have been severely squeezed. Faced with a choice between service delivery and policy evaluation, it may be tempting to regard evaluation as a luxury that government can ill afford, at least in the short term.

We reflect briefly on why such a response may be shortsighted and offer some thoughts on how the approach to policy evaluation may adapt to these more straitened times. This is a theme that is taken up in a number of places by the papers in this issue.

The key reason why evaluation continues to be relevant in times of financial austerity is that it provides the evidence needed for informed decision-making. This is especially relevant when resources are scarce, a point made forcefully by Rachel Glennerster (pp. R4-14) in her review of policy evaluation in developing countries. Evaluation evidence, including the assessment of the costs and benefits of policy, offers the most direct insight to inform such decisions, as Haroon Chowdry (pp. R53-64) makes clear. Policies found through rigorous analysis to be both effective and cost-effective are strong candidates for investment. Furthermore, broader explorations--including qualitative analyses such as the one discussed by Heather Rolfe (pp. R65-76)--can provide a deeper understanding of the way policy is implemented and how and why a policy seems (not) to work, sometimes suggesting ways of adapting policies to better achieve their aims.

Evaluation can be costly but it is often a means of saving money. The recently-completed Employment Retention and Advancement (ERA) demonstration (Hendra et al., 2011) provides an edifying example of how a thorough evaluation can help guide policymakers away from wasteful spending. Davies and Chowdry outline the features of the intervention. The evaluation revealed that impacts for the first two years were large and significant but subsequently disappeared. Overall, the intervention failed the cost-benefit test for lone parents. The evidence cautions strongly against national roll-out of the policy (Chowdry). Evaluations of labour market programmes typically confine themselves to a shorter period over which to observe impacts. Had the ERA evaluation only observed outcomes for two years, say, the conclusion would have been very different, with significant financial implications.

Properly conducted, evaluations offer an objective assessment of the appropriateness of policy, thus enabling governments to deploy available funds in the most cost-effective manner. When the Work and Pensions Committee asked the employment minister, the Rt Hon Chris Grayling MP, why the Department for Work and Pensions (DWP) had decided not to commission an independent evaluation of the Future Jobs Fund (FJF) he cautioned of the "... danger in financially straitened times that we spend vast amounts of money evaluating things that we are no longer going to do." (1) The decision to discontinue the FJF was not based on evaluation evidence and, consequently, it is impossible to establish its cost-effectiveness. One argument for ending the FJF was that, from June 2011, the Work Programme has provided a unified mechanism for helping individuals move into sustained employment. Unfortunately, little allowance has been made for a robust quantitative evaluation of this new policy. The way it was introduced--nationwide and with no trialling--means that attempts to estimate its effectiveness face very difficult challenges. Given that both the FJF and the Work Programme involve the expenditure of billions of pounds of public money, this may be a false economy.

Convincing and impartial evaluations often require the involvement of external researchers. This can be achieved in various ways. Over the past decade, a number of commissioning departments have set up framework agreements as a means of channelling such input. Under this arrangement, external contractors compete to become framework members who can then be invited--on either a competitive or non-competitive basis--to submit proposals for discrete evaluation (or other) projects. Alternatively, departments can provide support for external centres. Stephen Machin and Sandra McNally (pp. R15-25) report on the contribution of the Centre for the Economics of Education (jointly funded, until recently, by the Department for Education and the Department for Business, Innovation and Skills) to the evaluation of schools policies in England. The Low Pay Commission (LPC) is an independent statutory body set up under the National Minimum Wage Act 1998. Tim Butcher (pp. R26-40) reflects on how the LPC's programme of evaluation research has influenced the evolution of the National Minimum Wage, while Justin van de Ven (pp. 77-89) reports on analyses exploring the potential impact of a new pension scheme for those on modest incomes, which was funded under the LPC's research programme. A recent interesting model for providing funding is the Education Endowment Foundation, a charity founded by the Sutton Trust and the Impetus Trust and responsible for using a large grant from the Department for Education to commission projects that will identify and evaluate ways of improving achievement levels among disadvantaged children.

Of course, the funds available to support evaluation are limited and there is a separate point about whether and how there might be scope to deliver robust evaluation evidence at lower cost. There are no universal laws here and each policy or pilot must be considered in its own right. However, it is important to keep in mind opportunities for efficiencies in this regard.

The ERA evaluation mentioned above was perhaps as comprehensive an evaluation as it is possible to imagine. The intervention itself was costly (involving substantial financial payments, training of special advisers and time-consuming delivery of support and guidance) but the evaluation involved a lengthy design phase, close monitoring of implementation, detailed manipulation of administrative data, large longitudinal survey data collection, extensive qualitative interviewing with recipients and providers of the service, careful estimation of the impact of the programme and, lastly, a cost-benefit analysis (which itself involved a substantial cost-collection exercise). Comprehensive and expensive evaluations such as this are unlikely to be commissioned in the foreseeable future. However, smaller evaluations can also provide high-quality evidence. As Rachel Glennerster points out, even random assignment evaluations, which are generally regarded as being more credible than alternative approaches, can be afforded in certain circumstances (often, it is not the randomisation itself which is costly). A good example is Dorsett and Smeaton (2008) which randomly assigned to older benefit claimants the requirement to participate fully in an active labour market programme. By tracking employment and benefit outcomes using administrative data, compelling evidence of programme effectiveness was provided. The point here is that the nature of the intervention was such that a much more limited evaluation was feasible and could provide reliable impact estimates at a fraction of the cost of the ERA demonstration. Whether such a pared-down approach is appropriate to other evaluations depends on a number of factors, including whether the programme is new (as opposed to an existing programme being extended to a new client group) and whether the outcomes of interest exist in readily-available administrative data. Certain conditions still must be met, of course. Rolfe reports here on a random assignment evaluation that was less successful.

One legitimate criticism often heard from politicians and senior policymakers is that evaluation evidence, while interesting, comes too late to influence policy as directly as it might. There is thus a premium on researchers working with policymakers to ensure timely feedback. Of course, certain types of evidence simply cannot be produced on short order. The ERA demonstration is again a good example; estimating longer-term impacts required observing what happens to people over a five-year period. However, some types of evidence can be provided more quickly. Qualitative evidence, for instance, can provide early insights and, importantly, highlight practical issues that have a clear relevance to the effectiveness of policy. Furthermore, in some cases (such as van de Ven in this issue), it may be possible to use economic models to estimate in advance what the effect of a particular intervention may be. Such an approach requires that the policy change is amenable to economic modelling. Changes to the tax/transfer system, for instance, involve an amendment to individuals' incentives to work that can be explored ex ante (see, for example, Brewer et al., 2006, for such an evaluation of Working Families' Tax Credit). However, there is an unavoidable fact that good evaluation evidence often takes time to materialise. While ministers may be impatient to see results, the quality and usefulness of the evidence base in the UK would be improved were this key point borne in mind, not least in the commissioning of research.

The academic and research community is interested in the effectiveness of government policies and wants to carry out evaluation studies even without government funding. Such studies are of clear public benefit and should be of interest to government. However, the data required to carry out evaluation studies are often subject to rigid access restrictions. This is entirely appropriate given the potential for disclosure when working with individual-level data. However, while issues of data security and anonymity remain paramount, this should not be an argument for stasis. Recently, there have been some improvements in facilitating safe access to individual-level datasets for research purposes. We very much welcome this and hope that it is the beginning of a trend that would extend to ever more government-held datasets. This offers a means of democratising research and allowing the data to achieve their potential as a public good. Taken together, the papers in this edition make a compelling case for robust, independent evaluation being at the heart of government efforts to deploy policies that improve citizens' welfare.

doi: 10.1177/002795011221900101

REFERENCES

Brewer, M., Duncan, A., Shephard, A. and Suarez, M. (2006), 'Did working families' tax credit work? The impact of in-work support on labour supply in Great Britain', Labor Economics December, 13(6), pp. 699-720.

Dorsett, R. and Smeaton, D. (2008), Mandating Intensive Activity Period for Jobseekers aged 50+: Final Report of the Quantitative Evaluation, DWP Research Report 500.

Hendra, R., Riccio, J., Dorsett, R., Greenberg, D., Knight, G., Phillips, J., Robins, P., Vegeris, S. and Walter, J., with Hill, A., Ray, K. and Smith, J. (2011), Breaking the Low-pay, No-pay cycle: Final Evidence from the UK Employment Retention and Advancement (ERA) Demonstration, DWP Research Report 765.

NOTE

(1) House of Commons Work and Pensions Committee, Youth Unemployment and the Future Jobs Fund, First report of 2010/-11, HC 472.

Alex Bryson, Richard Dorsett and Jonathan Portes *

* National Institute of Economic and Social Research. E-mail: a.bryson@niesr.ac.uk
联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有