Much of the testing discussion in the United States today is grounded on several widely accepted notions: that we first must get the actual assessment instrument right, that there is an important distinction between 鈥渇ormative鈥 and 鈥渟ummative鈥 assessment, that teachers need to understand the language of assessment, and that we should drop tests on schools like 鈥減recision bombs鈥 for the purpose of measuring a student鈥檚 performance and progress.
These notions are misguided, as decades of research from around the world on what matters most in student learning demonstrates. In fact, the major purpose of assessment in schools should be to provide interpretative information to teachers and school leaders about their impact on students, so that these educators have the best information possible about what steps to take with instruction and how they need to change and adapt.
So often we use assessment in schools to inform students of their progress and attainment. Of course this is important, but it is more critical to use this information to inform teachers about their impact on students. Using assessments as feedback for teachers is powerful. And this power is truly maximized when the assessments are timely, informative, and related to what teachers are actually teaching.
We hardly need more data鈥攕chools are awash with data. Instead, we need better interpretations of these data. When we built the New Zealand school assessment system 15 years ago, we started by designing the reports, rather than the tests. The reports were piloted to ensure we met our own two criteria:
鈥 Did the teachers and school leaders interpret the reports correctly? (And if not, we needed to change the reports.)
鈥 Was there a consequential action from the reports?
After many iterations of the reports, we were able to meet these criteria, and then, and only then, we backfilled them with great assessment items. The reports, not the tests, are what matter to teachers and school leaders; they are hungry for information to help them know their impact on their students. At the system level, policymakers are happy to receive continual information on the status of learning in the country without all the negativity that comes with high-stakes testing. Given that the New Zealand assessment system is voluntary, it鈥檚 worth noting that a majority of teachers and schools still use it today, 15 years later.
Students can be beneficiaries of assessment, provided we start by asking, 'Who owns the test data?'
Until we see tests as aids to enhance teaching and learning, and not primarily as barometers of how much a student knows now, on this day, on this test, then developing more tests will add little, and will remain an expensive distraction. Educators need to understand what each student already knows, and where that student needs to go next in the teaching process. They need to be experts in using an array of interventions to help get their students to success, and to evaluate the impact that they鈥檝e made.
As an education researcher, I know this because I spent more than 15 years collecting nearly 1,200 meta-analyses of 65,000 education studies focused on the learning of more than 250 million students around the world. Because almost any education intervention can claim some positive effect on learning, I developed a threshold to determine effect.
Far and away the most effective teaching intervention we found was what I call 鈥渧isible learning": raising the quality of the feedback teachers receive about their impact. Expert teachers assess the visible impact they have on their students, constantly monitor learning and seek feedback about their teaching, and then evaluate and adjust their teaching methods based on these findings. Teachers show students how to self-assess their own performance, and discuss how they can improve. These are the major influences for raising achievement.
Students can be beneficiaries of assessment, provided we start by asking, 鈥淲ho owns the test data?鈥 The answer must be the students鈥攊t is their schooling, their lives, their futures that are at stake in classrooms鈥攁nd so the focus should be first on developing 鈥渟tudent assessment capabilities.鈥 It is then incumbent on teachers to teach students how to know when the best time to assess their learning is, and how to interpret the information from the tests. This is the core idea of what we call formative interpretation鈥攕tudents learning to know what to do next in light of their progress.
The difference between formative and summative is important. As the University of Illinois assessment expert Bob Stake once put it, 鈥淲hen the cook tastes the soup, it is formative; when the guests taste the soup, it is summative.鈥 Both can be valuable, and it is crucial to note that they do not refer to the tests. The same measure can be used for formative or summative interpretations.
So to speak of formative assessment is misleading, as it is the interpretation during the learning that is critical. But if we instead jump straight to 鈥渟tudent assessment capabilities,鈥 we avoid false distinctions, we place the students back at the center, and we are happy only when students understand their own learning based on test evidence. This is a powerful call to action for teachers鈥攖o learn about their own impact from assessments, and to teach students to do what they are asked themselves to do in their teaching, which is to evaluate the impact of learning.
This also means we do not need the notion of 鈥渢eacher assessment literacy"鈥攚hy should teachers be asked to learn the language of measurement people? Instead, we measurement people should learn how to speak in the language of learning and teaching and provide interpretations that are in turn correctly interpreted by teachers, with consequential actions and decisions. Similarly, we need reports from student assessments that help students understand their own progress in learning鈥攚hat they can do, what they cannot yet do, where to go next. If only tests were created with these aims.
Assessment can be powerful in classes, schools, and school systems, but we need to stop the high-stakes methods and the overreliance on developing tests to maximize precision. We need to instead focus on the power of assessment as feedback to help teachers maximize their impact, and ask teachers to teach students how to interpret their own information from assessments.