Electronic Submission Margins Except for the running head see belowleave margins of one inch at the top and bottom and on both sides of the text. Text Formatting Always choose an easily readable typeface Times New Roman is just one example in which the regular type style contrasts clearly with the italic, and set it to a standard size, such as 12 points.
One program tested in Kenya jumped out, and the Rwandan government wanted to know whether it would likely work in Rwanda as well. A randomized controlled trial RCT found that showing eighth-grade girls and boys a minute video and statistics on the higher rates of HIV among older men dramatically changed behavior: The number of teen girls who became pregnant with an older man within the following 12 months fell by more than 60 percent.
Random assignment determined which girls received the risk awareness program and which girls continued to receive the standard curriculum.
Our government partners could thereby have confidence that the reduction in risky behavior was actually caused by the program.
But if they replicated this approach in a new context, could they expect the impact to be similar? Policy makers repeatedly face this generalizability puzzle—whether the results of a specific program generalize to other contexts—and there has been a long-standing debate among policy makers about the appropriate response.
But the discussion is often framed by confusing and unhelpful questions, such as: Should policy makers rely on less rigorous evidence from a local context or more rigorous evidence from elsewhere? And must a new experiment always be done locally before a program is scaled up? These questions present false choices.
Rigorous impact evaluations are designed not to replace the need for local data but to enhance their value. This complementarity between detailed knowledge of local institutions and global knowledge of common behavioral relationships is fundamental to the philosophy and practice of our work at the Abdul Latif Jameel Poverty Action Lab J-PALa center at the Massachusetts Institute of Technology founded in with a network of affiliated professors and professional staff around the world.
Four Misguided Approaches To give a sense of our philosophy, it may help to first examine four common, but misguided, approaches about evidence-based policy making that our work seeks to resolve.
Can a study inform policy only in the location in which it was undertaken? Kaushik Basu has argued that an impact evaluation done in Kenya can never tell us anything useful about what to do in Rwanda because we do not know with certainty that the results will generalize to Rwanda.
Describing general behaviors that are found across settings and time is particularly important for informing policy. The best impact evaluations are designed to test these general propositions about human behavior. Should we use only whatever evidence we have from our specific location?
In an effort to ensure that a program or policy makes sense locally, researchers such as Lant Pritchett and Justin Sandefur argue that policy makers should mainly rely on whatever evidence is available locally, even if it is not of very good quality. The challenge is to pair local information with global evidence and use each piece of evidence to help understand, interpret, and complement the other.
Should a new local randomized evaluation always precede scale up? One response to the concern for local relevance is to use the global evidence base as a source for policy ideas but always to test a policy with a randomized evaluation locally before scaling it up.
With limited resources and evaluation expertise, we cannot rigorously test every policy in every country in the world. We need to prioritize. For example, there have been more than 30 analyses of 10 randomized evaluations in nine low- and middle- income countries on the effects of conditional cash transfers.
While there is still much that could be learned about the optimal design of these programs, it is unlikely to be the best use of limited funds to do a randomized impact evaluation for every new conditional cash transfer program when there are many other aspects of antipoverty policy that have not yet been rigorously tested.
Must an identical program or policy be replicated a specific number of times before it is scaled up? One of the most common questions we get asked is how many times a study needs to be replicated in different contexts before a decision maker can rely on evidence from other contexts. We think this is the wrong way to think about evidence.
There are examples of the same program being tested at multiple sites: For example, a coordinated set of seven randomized trials of an intensive graduation program to support the ultra-poor in seven countries found positive impacts in the majority of cases. This type of evidence should be weighted highly in our decision making.
But if we only draw on results from studies that have been replicated many times, we throw away a lot of potentially relevant information.
Focus on Mechanisms These four misguided approaches would have blocked a useful path forward in deciding whether to introduce the HIV information program in Rwanda. This is because they ignore the key insight from an evaluation: First, such a focus draws attention to more relevant evidence.
When considering whether to implement a specific policy or program, we may not have much existing evidence about that exact program. But we may have a deep evidence base to draw from if we ask a more general question about behavior. For example, imagine a public health agency that would like to encourage health-care providers to promote flu vaccinations.
A review of the literature may produce few, if any, rigorous evaluations of this specific approach. Second, underlying human behaviors are more likely to generalize than specific programs.Long-Term Effects of Teacher Performance Pay: Experimental Evidence from India Karthik Muralidharan† 10 April * Abstract: We present results from a five-year long randomized evaluation of group and individual teacher performance pay programs implemented across a large representative sample of government-run.
Founded in , Macmillan Publishers is one of the largest global trade book publishers and home to numerous bestselling and award-winning fiction, nonfiction, and children’s books, from St.
Martin’s Press, Tor Books, Farrar, Straus & Giroux, Henry Holt, Picador, Flatiron Books, Celadon Books, and Macmillan . In this paper we describe some of our findings in the context of Latour and Woolgar’s seminal work on the incentive systems that motivate publishing scientists.
We suggest that incentives in Wikipedia and similar online communities. In what ways does the incentive. “Incentives, Motivation and Workplace Performance: Research & Best Practices,” conducted by researchers for the International Society of Performance Improvement, and funded with a grant by the SITE Foundation, was designed to analyze the complete body of scientific research on incentive programs, determine what if any research-supported.
What You'll Find in this Article: 1. Instructions for how to (and how not to) pick a topic. 2. Lists of topic ideas (in the categories of food and health, obesity and dieting, recycling and the environment, families and relationships, and science and technology, with videos and many links to research and student essay examples.
Program on Education Policy and Governance Working Papers Series You get what you pay for: Incentives and Selection in the Education System Armin Falk.