Includes updates and/or revisions.
It鈥檚 a daunting job for two big groups of states to design multilayered assessment systems by 2014, and a panel of experts has made it even more daunting, composing a long list of concerns about what it will take to make the venture a success.
On its list, the panel included high-level, long-range items such as integrating the tests into systems of instruction, and nitty-gritty, immediate worries such as making sure the tests鈥 computer demands don鈥檛 blow schools鈥 electrical circuits.
The to-do list was sketched out during a six-hour hearing convened April 15 by the U.S. Department of Education. It was the first in a series aimed at advising the two state collaboratives as they design tests for the new common standards in mathematics and English/language arts that have been adopted by all but six states, using $360 million in federal Race to the Top money. Forty-five states are participating in the assessment consortia.
The meeting focused on the technological challenges states and districts might face in the 2014-15 school year, when the federal government expects the tests to be fully operational. Questions of technological capacity loom, since both consortia plan assessments that are largely computer-based.
Presiding over the hearing, Ann Whalen, a top adviser to U.S. Secretary of Education Arne Duncan, called the technological questions 鈥渟ticky.鈥 By day鈥檚 end, the long list of cautions led one panelist to extend his sympathies to the two consortia, each of which was represented by a small battery of technical experts.
鈥淚f I were sitting in your shoes right now, I鈥檇 be feeling a bit nervous,鈥 said Michael Russell, who directs the Nimble Innovation lab at Measured Progress, a Dover, N.H.-based assessment company. As chuckles rippled around the room, Mr. Russell hurried to add an optimistic spin, expressing confidence that the test-design teams would adjust to the challenges ahead.
鈥淭wenty years from now, we鈥檒l look back, and it鈥檒l be second nature,鈥 he said.
The two consortia plan to work jointly to size up districts鈥 capacity to handle large-scale computer-based testing. They said they would soon issue a joint request for proposals from vendors to design a technology-readiness assessment that would give the consortia a sense of districts鈥 capacity and signal them about what they might need to transition to online testing.
States and districts must soberly assess what鈥檚 needed to make that shift, since the consortia鈥檚 鈥済reat visions of grandeur鈥 involve not just summative tests, but systems that include interim or formative tools, banks of test items, portals through which test results can be accessed and analyzed, and more, said John Jesse, the assessment director in Utah, which belongs to the .
The most problematic capacity issues will be at the schools themselves, Mr. Russell said. If an Internet router can鈥檛 handle 60 or 70 computers at once, for instance, problems could arise if a social studies teacher decides to stream video for her class while large groups of students are taking tests elsewhere in the building, he said.
Experts also warned the SBAC and the other consortium, the , or PARCC, to be cautious about administering tests on multiple kinds of devices. Schools might not be able to be 鈥渄evice agnostic,鈥 Mr. Russell said, if a student can鈥檛 demonstrate achievement as well on a tablet computer as on a desktop computer. It might not be possible to deliver assessments the same way on different devices 鈥渨ithout measurement effects,鈥 he said.
Lessons Learned
A trio of officials from Virginia was on hand to share the state鈥檚 experience, since that state was on the leading edge when it ventured into online assessment in 2000. Sarah Susbury, who oversees test administration, scoring, and reporting, noted that Virginia had six years to phase in its online tests, a luxury the consortia won鈥檛 have, with their deadlines only three years away. Virginia did not adopt the common standards or join either assessment consortium.
In moving to online testing, a key lesson for Virginia was that it鈥檚 impossible to separate assessment and technology, Ms. Susbury said. Experts in both areas must leave their traditional silos and work closely together, she said.
Virginia officials emphasized the importance of having one portal that could serve as the hub for the testing system, from entering and viewing student data to training teachers and posting status updates about problems with the system. Especially during testing windows, it鈥檚 crucial to have one place for districts to check for problems, since the state education department and its contracted help desk can鈥檛 answer every phone call, said Shelley Loving-Ryder, Virginia鈥檚 assistant superintendent for assessment and school improvement.
If that experience is any guide, the consortia should be prepared to do 鈥渁 lot of hand-holding鈥 in training educators at all levels on the new systems, Ms. Loving-Ryder said. As Virginia prepares to pilot online writing assessments, 5th grade teachers are particularly concerned because they fear 10-year-olds鈥 keyboarding skills could affect test outcomes, she said. The education department is sharing the test interface ahead of time to give teachers time to practice, she said.
Security Concerns
Online testing also prompts new questions about security of test data, several panelists said. Ms. Susbury warned that 鈥渆ncryption and security are critical鈥 as states endeavor to protect test data. The state learned the hard way that it had to plan for unimagined security emergencies. In the first month of its program, a backhoe doing work on school property during a test severed a key computer line, prompting fear that students鈥 online responses wouldn鈥檛 be preserved, she recalled.
Mobile devices and 鈥渙pen source鈥 systems pose challenges as well, without adequate attention to security, said Denny Way, a senior research scientist for psychometric and research services at Pearson.
One approach to security for certain kinds of test items, he said, would be to produce and publicly share a huge number of items. Releasing hundreds of essays in English/language arts, for instance, would make it impossible to prepare for so many essays, making cheating very difficult.
Wes Bruce, Indiana鈥檚 chief assessment officer, warned against that approach. He said he heard that one district in his state released so many test items鈥攊n an apparent bid to help students prepare for the test鈥攖hat it constituted what he viewed as 鈥渁 crime against kids.鈥
Concerns expressed by the experts ranged from the big-picture to the very down-to-earth.
Richard F. Rozzelle, the president of the Center for Educational Leadership and Technology, which helps districts and states manage information technology, urged states to use the new testing systems as an opportunity to build a new 鈥渋nformation architecture鈥 that would integrate all pieces of the education spectrum, from curriculum design to management of assessment data.
Virginia officials urged the test-design teams to size up even the most basic forms of capacity at the school level. They recounted how one rural district decided to charge all its wheeled carts of laptop computers overnight, overloading the electrical circuits and shutting off heat in all its buildings.
Stories like that prompted Ken Wagner, the assistant commissioner for data systems and information and reporting services in New York state, to note that the 鈥渂oring, mundane details鈥 can easily sink an ambitious assessment system. 鈥淚f we don鈥檛 start talking about specific details,鈥 he said, 鈥渨e鈥檙e going to regret it.鈥