澳门跑狗论坛

School & District Management What the Research Says

What Does 鈥楨vidence-Based鈥 Mean? A Study Finds Wide Variation.

By Sarah D. Sparks 鈥 January 10, 2024 5 min read
photograph of a magnifying glass on an open book
  • Save to favorites
  • Print
Email Copy URL

What makes an education intervention evidence-based?

Over the last 20 years, that question has moved front and center as more federal and state agencies require programs to show evidence of effectiveness, and more education leaders look for proof that interventions used in other districts will help their own students.

But a new finds wide variation on the kind of evidence that is required to show an education intervention is effective. In fact, large research clearinghouses, set up to review the evidence bases of programs for practitioners, reach the same conclusion on less than a third of the education programs they review.

What that means is that a teacher or principal trying to choose a reading curriculum or tutoring program for students may find it recommended by one clearinghouse and rejected by another.

Researchers from George Washington and Northwestern universities analyzed the evidence standards for 10 common research clearinghouses, including the federal What Works Clearinghouse and the National Dropout Prevention Center in New York, part of the nonprofit Successful Practices Network. The majority of these groups are supported by public agencies or nonprofit foundations.

The team, led by postdoctoral researchers Mansi Wadhwa and Jingwen Zheng of George Washington, compared evidence reviews for nearly 1,360 pre-K-20 education programs and interventions whose evidence base had been reviewed by at least one of the 10 clearinghouses.

The study found 83 percent of the education programs reviewed had only ever been rated by one clearinghouse. Of the programs with multiple ratings, fewer than 1 in 3 had consistent ratings across clearinghouses.

The clearinghouses were more likely to agree about what didn鈥檛 work; more than 80 percent of programs with at least two similar reviews were deemed ineffective by both. Less than 18 percent of programs had at least two 鈥渆ffective鈥 ratings, and many had mixed reviews.

For example, five different clearinghouses reviewed the evidence for Peer-Assisted Learning Strategies, a peer-tutoring program focused on math and reading. One clearinghouse recommended the program as a whole and another found it promising.

Two others reviewed the program separately for each subject; one recommended the math program, while the other didn鈥檛 find the program promising in either math or reading.

And the final clearinghouse reviewed the PALs program effectiveness on a variety of different outcomes, finding evidence to recommend it for some purposes and not for others.

One reason for the disagreement is that standards differ from clearinghouse to clearinghouse, on what kinds of outcomes can be used to judge program effectiveness, how large a sample of students must be studied, and for how long.

They also differ on whether studies must use randomized controlled experiments, in which students are randomly assigned to a study or control condition, or other designs. Randomized studies are generally considered the most rigorous, but they are difficult and expensive to conduct in educational settings.

鈥淏ecause [research clearinghouses] do not agree on such criteria for acceptable evidence, and because they are important enough to lead to different judgments about program effectiveness, 鈥榚vidence-based鈥 seems to be an idea [with limited use] despite [clearinghouses] being funded precisely to identify which programs are most evidence based,鈥 they conclude.

Tough to build consensus

The nonprofit Successful Practices Network, one of the clearinghouses in the study which reviews research on issues like dropout recovery and career and technical education, doesn鈥檛 try to align how it defines evidence quality with other groups, according to Bill Dagget, the network鈥檚 founder.

鈥淚f you鈥檙e trying to define 鈥榚vidence-based,鈥 it鈥檚 very difficult to incorporate any of the skills that are harder to measure,鈥 like critical thinking, collaboration, or social-emotional development, Dagget said.

鈥淲hen you begin to look at these broader skills, you can鈥檛 evaluate those with a written test. Typically you have got to do some type of rubric,鈥 Dagget said. 鈥淭he problem with that is any time you use a rubric, I don鈥檛 care how carefully you train, the people using them are always somewhat subjective.鈥

In a , Jean Stockard, an emerita professor at the University of Oregon, found that half of the What Works Clearinghouse鈥檚 intervention reports were based on a single study. Stockard, who was not part of the new study in the journal Review of Educational Research, that out of more than 120 different studies of one broadly used literacy program, Reading Mastery, those that included evidence beyond randomized controlled studies had more consistent and precise reviews.

The effects of education interventions often fade over time, and the researchers said there鈥檚 little agreement on how long and how much follow-up should be done on evaluations. Clearinghouses most often required researchers to follow up a year after an intervention is used, but some allowed shorter follow-ups.

While national research groups have begun to advocate for more researchers to verify a program鈥檚 effectiveness, 鈥渆ducation research isn鈥檛 in limitless supply,鈥 said Julie Brosnan of the National Student Support Accelerator at Stanford university, which collects and conducts research related to tutoring programs.

鈥淔or instance, to use tutoring as an example, it is neither feasible nor cost-effective to have every tutoring program engage in a randomized controlled trial to test effectiveness given that there is such a strong evidence base,鈥 Brosnan said. 鈥淓ducation leaders need to ensure the program characteristics align with those that have evidence behind them, while also monitoring implementation and collecting ongoing data.鈥

Stockard and Brosnan agreed that to build an evidence base for a given program, studies need to include more details about how and for whom the program was used, as well as more analyses of multiple studies to tease out individual aspects of an intervention that may work for different groups of students.

鈥淚f the evidence isn鈥檛 the right evidence, then the study isn鈥檛 of much value,鈥 Dagget said. 鈥淪o the essential question is, what鈥檚 our purpose? Is our purpose to prepare kids for the next grade and the next test and the next level of education? Or is it to prepare them for the world beyond school?鈥

A version of this article appeared in the January 31, 2024 edition of 澳门跑狗论坛 as What Does 鈥楨vidence-Based鈥 Mean? A Study Finds Wide Variation

Events

Artificial Intelligence K-12 Essentials Forum Big AI Questions for Schools. How They Should Respond鈥
Join this free virtual event to unpack some of the big questions around the use of AI in K-12 education.
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of 澳门跑狗论坛's editorial staff.
Sponsor
School & District Management Webinar
Harnessing AI to Address Chronic Absenteeism in Schools
Learn how AI can help your district improve student attendance and boost academic outcomes.
Content provided by 
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of 澳门跑狗论坛's editorial staff.
Sponsor
Science Webinar
Spark Minds, Reignite Students & Teachers: STEM鈥檚 Role in Supporting Presence and Engagement
Is your district struggling with chronic absenteeism? Discover how STEM can reignite students' and teachers' passion for learning.
Content provided by 

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide 鈥 elementary, middle, high school and more.
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.

Read Next

School & District Management Principals Polled: Where School Leaders Stand on 10 Big Issues
A look at how principals responded to questions on Halloween costumes, snow days, teacher morale, and more.
4 min read
Illustration of speech/thought bubbles.
DigitalVision Vectors
School & District Management Opinion You鈥檙e the Principal, and Your Teachers Hate a New District Policy. What Now?
This school leader committed to being a bridge between his district and school staff this year. Here鈥檚 what he learned.
Ian Knox
4 min read
A district liaison bridging the gap between 2 sides.
Vanessa Solis/澳门跑狗论坛 via Canva
School & District Management The 4 District Leaders Who Could Be the Next Superintendent of the Year
Four district leaders are finalists for the national honor. They've emphasized CTE, student safety, financial sustainability, and more.
4 min read
Clockwise from upper left: Sharon Desmoulin-Kherat, superintendent of the Peoria Public School District 150; Walter Gonsoulin, superintendent of Jefferson County Schools; Debbie Jones, superintendent of the Bentonville School District; David Moore, superintendent of the School District of Indian River County.
Clockwise from upper left: Sharon Desmoulin-Kherat, superintendent of the Peoria school district in Illinois; Walter Gonsoulin, superintendent of Jefferson County schools in Alabama; Debbie Jones, superintendent of the Bentonville, Ark., school district; and David Moore, superintendent in Indian River County, Fla. The four have been named finalists for national Superintendent of the Year. AASA will announce the winner in March 2025.
Courtesy of AASA, the School Superintendent's Association
School & District Management 3 Tips for Districts to Maximize FEMA Funding After a Natural Disaster
District leaders who have been through natural disasters stress the need for thorough documentation, even if it seems excessive.
5 min read
Close up of FEMA paperwork
iStock/Getty