Using the as a building block, the U.S. Department of Education is taking the next formal step to make research and evidence far more important factors as it awards competitive grants.
The goal is twofold: to reward projects that already have established a research-based track record of success and to encourage grant winners to produce rigorous evidence detailing the extent to which their project does鈥攐r does not鈥攚ork.
To make that happen, the Education Department is proposing significant changes to , or the Education Department General Administrative Regulations, as part of a governmentwide push to introduce more evidence into decisionmaking.
Those rule changes, which are open for public comment, will serve as an umbrella over all the department鈥檚 competitive programs, potentially governing more than $2 billion in grants. They would not apply to large, hallmark grants鈥攕uch as Title I for disadvantaged students鈥攖hat are given to states under a formula set by law.
But the proposed rules nonetheless signal to educators and policymakers that evidence matters.
Staying Power
鈥淭his has the potential to be a major step forward,鈥 said Jon Baron, the president of the Coalition for Evidence-Based Policy, a Washington-based group of researchers working to increase government effectiveness through evidence. 鈥淭he question is the extent to which this would be used. But for the first time, in many department programs, it would begin to institutionalize the development and the use of rigorous evidence.鈥
It鈥檚 not that the Education Department could not have done this before on a piecemeal basis, but placing those rules in EDGAR gives the ideas more staying power. They can now be applied to any competitive-grant program without having to go through the cumbersome, rulemaking process each time a new competition is launched.
The department has not yet indicated what competitive-grant programs, beyond i3, might begin to use the standards.
Still, 鈥渢his shows, again, that the Education Department has a preference for funding programs that have a very strong evidence base,鈥 said Michele McLaughlin, the president of the Knowledge Alliance, a national association of research groups based in Washington.
The groundwork for the rules was laid early in the Obama administration with the launch in 2009 of the relatively small, yet important, $650 million Investing in Innovation contest. Known as i3, it rewarded not just promising ideas but projects with strong evidence of past success. The contest was the first effort in years by the department to award grants based, in part, on results proved by sound research.
And last year, the to federal agencies directing them to include with their fiscal 2014 budget requests their 鈥渕ost innovative uses of evidence and evaluation.鈥 A lack of emphasis on evidence has been considered a problem across agencies, not just education, particularly when it comes to social policy.
The proposed EDGAR changes are out for public comment until Feb. 12; the timeline for getting them on the books depends on how many comments are received.
Once they are finalized, the rules would:
鈥 Create tiers of evidence. This recognizes that depending on the competition and its goals, more or less evidence may be appropriate.
鈥 Standardize definitions of what it means to have 鈥渟trong鈥 or 鈥渕oderate鈥 evidence, or 鈥渆vidence of promise鈥 and 鈥渟trong theory.鈥
鈥 Define what kind of evidence a project must produce at the end of a grant period. This seeks to ensure some projects produce strong evidence that meets the 鈥榮 evidence standards, in which the gold standard is an experiment that randomly assigns students to either an intervention or a control group. (The online clearinghouse is an initiative of the Education Department to vet education research against rigorous standards.)
鈥 Extend the official project time frame for a grant winner. This would enable some winners to collect data on a project鈥檚 impact long after the grant officially ends.
Link to Outcomes
Several groups that monitor the use of evidence in federal grantmaking, including Mr. Baron鈥檚 coalition, are urging Education Department officials to make one change in their proposal: to make clear that strong evidence should be linked to outcomes that matter鈥攕uch as high school graduation rates鈥攁nd not an outcome that may be statistically significant but not related to improving student achievement.
The department said it will consider, and respond to all comments, once the deadline passes.
The Investing in Innovation program 鈥渨as a way of seeing how this could actually work,鈥 said Jim Shelton, the department鈥檚 assistant deputy secretary for innovation and improvement, referring to evidence-based grantmaking. 鈥淣ow, we want to institutionalize the evidence framework for use by other programs now and in the future.鈥
Robert E. Slavin, the director of the Center for Research and Reform in Education at Johns Hopkins University, in Baltimore, pointed out that in any federal contests, there are often bonus points鈥攁lso called competitive preferences鈥攇iven for focusing a grant proposal on a rural area, or for doing something in the stem (science, technology, engineering, and math) subjects, for example.
鈥淥ftentimes the difference between a successful proposal and an unsuccessful proposal is a couple of points,鈥 said Mr. Slavin, whose school improvement organization has won two i3 grants and is involved in a third. 鈥淭hese competitive preferences can be hugely important. Well why not ask whether projects actually work?鈥
Success for All has been so successful in winning i3 grants in part because it started building an evidence base for its work when it was founded in 1987, Mr. Slavin said.
鈥淧art of what we were trying to do is build a good program,鈥 he said, 鈥渂ut also build the idea that rigorously evaluated programs were a way forward for education in general.鈥