Includes updates and/or revisions.
All states and the District of Columbia have now ushered in new English-language-proficiency tests to comply with No Child Left Behind Act requirements for those still learning the language. Still, it鈥檚 too early to tell if a passing score predicts that a student will do well in a regular classroom or on other mandatory state tests, according to a nationwide examination of such assessments released today.
鈥淲e see systematic improvement鈥 in the new generation of English-proficiency tests, Jamal Abedi, a professor of education at the University of California, Davis, and the editor of the report, said in a phone interview. 鈥淏ut we don鈥檛 know how they actually translate into performance of English-language learners.鈥
The report, released by Mr. Abedi鈥檚 university, is a collaboration by 32 experts on testing and ELLs to provide an overview of the new era of testing for English proficiency.
Under the nearly 6-year-old NCLB law, states were, for the first time, required to gauge the progress of second-language learners in grades K-12 every year in learning English. The federal education law requires states to assess those students in reading, writing, speaking, and listening鈥攚hile the previous generation of tests were mostly designed to assess only speaking and listening.
The tests are better than the ones that were commonly used prior to passage of the NCLB law because they are aligned with state standards for English-proficiency鈥攚hich many states had to craft to comply with the act鈥攁nd state content standards, such as for mathematics, Mr. Abedi explained.
In addition, he said, the new generation of tests are designed to assess 鈥渁cademic English鈥濃攖he version of the language children need in order to learn subjects in school鈥攔ather than the social English used on the playground.
Prodded by NCLB
Experts say the federal education law鈥檚 requirements to test English progress in a comprehensive way is a positive move, even though much more work needs to be done to ensure that the tests are valid and meaningful.
鈥淭hese tests wouldn鈥檛 be as good as they are without those [NCLB] requirements,鈥 said Stanley N. Rabinowitz, the director of assessment and standards-development services for WestEd, a San Francisco-based research agency. He is conducting a separate examination of a number of the new English-proficiency tests.
Mr. Rabinowitz agreed with Mr. Abedi that the states have done a good job in coming up with tests that evaluate academic English. At the same time, he said, one limitation of the UC-Davis report is that it is written primarily by the people who devised the new tests. 鈥淚 won鈥檛 call it public relations, but it is the best foot forward,鈥 he said. 鈥淣ot all tests have been received and are working as well as this report indicates.鈥
In the case of Florida, noted Candace A. Harper, an associate professor of education at the University of Florida, Gainesville, the report鈥檚 description of the state鈥檚 English-proficiency test is much more 鈥済lowing鈥 than what teachers had to say about the test in her state.
The new generation of tests adopted by states is more rigorous and comprehensive than the previous tests. Typical is the Assessing Comprehension and Communication State to State for English Language Learners test, or ACCESS for ELLs, now used by 15 states to replace a variety of older tests.
THE OLD TESTS
鈥 Not based on standards
鈥 Nonsecure, low-stakes tests
鈥 Social language emphasized
鈥 Not linked with content standards
鈥 Not compliant with NCLB
鈥 Static, with irregular updates
THE NEW TEST**
鈥 Based on English-language-proficiency standards
鈥 Secure, high-stakes test
鈥 Academic language emphasized
鈥 Aligned with content standards
鈥 Compliant with NCLB
鈥 Updated every year
**Devised by the World-Class Instructional Design and Assessment Consortium, the test has been adopted by Alabama, Delaware, the District of Columbia, Georgia, Illinois, Kentucky, Maine, New Hampshire, New Jersey, North Dakota, Oklahoma, Pennsylvania, Rhode Island, Vermont, and Wisconsin. Virginia plans to implement it in the next school year.
SOURCE: 鈥淓nglish Language Proficiency Assessment in the Nation: Current Status and Future Practice鈥
For example, she and several graduate students interviewed 12 teachers and 16 students about Florida鈥檚 test, the Comprehensive English Language Learning Assessment. They found that the listening section was broadcast over the public address system in one school鈥檚 cafeteria, while at another school, an English-as-a-second-language teacher administered that section to students in small groups. Such inconsistencies in implementation raise questions about the validity of the scores, Ms. Harper said.
Other experts also said it鈥檚 important to learn more about whether the English-proficiency tests are working as the developers intended and how useful they are to educators.
鈥淲hat do you do with the scores?鈥 said Diane Staehr Fenner, a research scientist at the Center for Equity and Excellence in Education at George Washington University in the nation鈥檚 capital.
鈥淵ou see all these numbers. Is the teacher trained on how to read these scores? How do you translate the scores into meaningful classroom instruction?鈥 said Ms. Staehr Fenner, who previously was an assessment specialist for ELLs in the Fairfax County, Va., schools.
Federal Efforts
The UC-Davis report is being released as the U.S. Department of Education is trying to respond to state officials鈥 pleas for help in writing and using English-proficiency tests and for setting targets under the federal education law, called 鈥渁nnual measurable achievement objectives,鈥 for students to progress in English and attain proficiency in the language.
Separately, state officials and federal education officials have, at times, been at odds over how best to include English-learners in the regular mathematics and reading assessments that all students must take. (鈥淭ussle Over English-Language Learners,鈥 Jan. 31, 2007.)
Meanwhile, states鈥攚ith federal assistance鈥攈ave quietly been taking steps to put English-proficiency tests in place. To date, the Education Department has provided a total of $10 million to four consortia of states to craft new tests, which about half are using, and has followed up with some additional grants. The rest of the states are using tests designed specifically for their states or by commercial developers.
The Education Department released in October a draft of a framework for creating English-language-proficiency standards and tests. Federal officials soon will invite states to participate in a pilot project to use the framework in evaluating their English proficiency standards and tests, according to Kathryn M. Doherty, a special assistant to the department鈥檚 deputy secretary.
In addition, Ms. Doherty said, federal officials are writing official guidance for states on how to set targets for the achievement objectives, which federal officials expect to release in the spring.
At the state level, educators involved in the implementation of English-proficiency tests said the UC-Davis report helps to put into perspective what is happening with such tests in their own states.
鈥淗aving a report of this type earlier could have conserved resources so that states would not have had to independently collect current information to make decisions about which English-language-proficiency assessment they would select,鈥 said Teddi Predaris, the director of the office of services for English-language learners in the 164,000-student Fairfax County, Va., school system.
Ms. Predaris served on a committee to select a new English-proficiency test for Virginia to replace the one that most districts in the state are now using, the Stanford English Language Proficiency Test, to comply with the NCLB law. After comparing four tests, the committee recommended switching to a test known as Assessing Comprehension and Communication State to State for English Language Learners, or ACCESS for ELLs, starting next school year.
The Virginia board of education approved that selection in September, making Virginia the 16th state to adopt the test produced by the World-Class Instructional Design and Assessment Consortium. It is the English-proficiency test most widely used by states.
Veteran 澳门跑狗论坛 reporter Mary Ann Zehr tackles difficult policy questions, explores learning innovations, and shares stories about different cultural groups on her beat.
.
鈥淭he ACCESS for ELLs test is the best thing out there right now because it is carefully aligned with the content-area assessments and is an accurate, valid, and reliable assessment,鈥 Ms. Predaris said.
In Florida, Ms. Harper said she was concerned by the UC-Davis report鈥檚 finding that Florida is behind many other states in putting an English-proficiency test in place. While Florida just carried out its English-proficiency test for the first time last school year, California, for instance, administered such a test before passage of the NCLB law.
Ms. Harper, who serves on an advisory committee to the Florida education department for implementation of Florida鈥檚 test, believes it is premature to require that the test be used by schools to make decisions such as when to release children from English-acquisition programs.
鈥淗ow do these scores reflect what students can do in content-area classrooms?鈥 she said, 鈥淲e don鈥檛 know that. We have very shaky data. Even the assessment people [on her committee] were saying, 鈥楾his is like building on mud.鈥 鈥
Ms. Harper鈥檚 concerns are warranted, said Mr. Abedi, the report鈥檚 editor. A state should have at least two years鈥 worth of data based on full implementation of its test before using it for any high-stakes decisions, he added. Now, Mr. Abedi said, most of the data available is from field-testing or pilot-testing, not based on full implementation across a state over a significant amount of time.