This is the accessible text file for GAO report number GAO-06-815 entitled 'No Child Left Behind Act: Assistance from Education Could Help States Better Measure Progress of Students with Limited English Proficiency' which was released on July 26, 2006. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: United States Government Accountability Office: GAO: July 2006: No Child Left Behind Act: Assistance from Education Could Help States Better Measure Progress of Students with Limited English Proficiency: No Child Left Behind Act: GAO-06-815: GAO Highlights: Highlights of GAO-06-815, a report to congressional requesters Why GAO Did This Study: The No Child Left Behind Act of 2001 (NCLBA) focused attention on the academic achievement of more than 5 million students with limited English proficiency. Obtaining valid test results for these students is challenging, given their language barriers. This report describes (1) the extent to which these students are meeting annual academic progress goals, (2) what states have done to ensure the validity of their academic assessments, (3) what states are doing to ensure the validity of their English language proficiency assessments, and (4) how the U.S. Department of Education (Education) is supporting states’ efforts to meet NCLBA’s assessment requirements for these students. To collect this information, we convened a group of experts and studied five states (California, Nebraska, New York, North Carolina, and Texas). We also conducted a state survey and reviewed state and Education documents. What GAO Found: In the 2003-2004 school year, state data showed that the percentage of students with limited English proficiency scoring proficient on a state’s language arts and mathematics tests was lower than the state’s annual progress goals in nearly two-thirds of the 48 states for which we obtained data. Further, our review of data 49 states submitted to Education showed that in most states, these students generally did not perform as well as other student groups on state mathematics tests. Factors other than student knowledge, such as how a state establishes its annual progress goals, can influence whether states meet their goals. For their academic assessments, officials in our five study states reported taking steps to follow generally accepted test development procedures and to ensure the validity and reliability of these tests for students with limited English proficiency, such as reviewing test questions for bias. However, our group of experts expressed concerns about whether all states are assessing these students in a valid manner, noting that some states lack the resources and technical expertise to take appropriate steps to ensure the validity of tests for these students. Further, Education’s peer reviews of assessments in 38 states found that 25 states did not provide adequate evidence to ensure the validity or reliability of academic test results for these students. To improve the validity of these test results, most states offer accommodations, such as a bilingual dictionary. However, our experts reported that research is lacking on what accommodations are effective in mitigating language barriers. A minority of states used native language or alternate assessments for students with limited English proficiency, but these tests are costly to develop and are not appropriate for all students. Many states are implementing new English language proficiency assessments in 2006 to meet NCLBA requirements; as a result, complete information on their validity and reliability is not yet available. In 2006, 22 states used tests developed by one of four state consortia. Consortia and state officials reported taking steps to ensure the validity of these tests, such as conducting field tests. A 2005 Education-funded technical review of available documentation for 17 English language proficiency tests found insufficient documentation of the validity of these assessments’ results. Education has offered a variety of technical assistance to help states assess students with limited English proficiency, such as peer reviews of states’ academic assessments. However, Education has issued little written guidance to states on developing English language proficiency tests. Officials in one-third of the 33 states we visited or directly contacted told us they wanted more guidance about how to develop tests that meet NCLBA requirements. Education has offered states some flexibility in how they assess students with limited English proficiency, but officials in our study states told us that additional flexibility is needed to ensure that progress measures appropriately track the academic progress of these students. What GAO Recommends: GAO recommends that the Secretary of Education (1) support research on accommodations, (2) identify and provide technical support states need to ensure the validity of academic assessments, (3) publish additional guidance on requirements for assessing English language proficiency, and (4) explore ways to provide additional flexibility for measuring annual progress for these students. Education generally agreed with our recommendations. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-06-815]. To view the full product, including the scope and methodology, click on the link above. For more information, contact Marnie S. Shaul at (202) 512-7215 or shaulm@gao.gov [End of Section] Contents: Letter: Results in Brief: Background: Students with Limited English Proficiency Performed below Progress Goals in 2004 in Two-Thirds of States, but States We Studied Are Working to Improve Student Academic Performance: Selected States Considered Language Issues When Developing Academic Assessments, but Validity and Reliability Concerns Remain: Most States Implemented New English Language Proficiency Assessments but Faced Challenges Establishing Their Validity and Reliability: Education Has Provided Assistance, but States Reported Need for Additional Guidance and Flexibility: Conclusions: Recommendations for Executive Action: Agency Comments: Appendix I: GAO's Group of Experts on Assessing the Academic Knowledge of Students with Limited English Proficiency: Appendix II: Determining Adequate Yearly Progress for Student Groups: Appendix III: Percentage of Districts Making AYP Goals for Mathematics for Students with Limited English Proficiency: Appendix IV: Proficiency Scores on Mathematics Tests for All Students and Students with Limited English Proficiency: Appendix V: Enhanced Assessment Consortia Participation: Appendix VI: English Language Proficiency Assessments Used in the 2005- 2006 School Year, by State: Appendix VII: Comments from the Department of Education: Appendix VIII: GAO Contacts and Acknowledgments: Related GAO Products: Tables: Table 1: Selected Provisions from Title I of NCLBA: Table 2: Selected Provisions from Title III of NCLBA: Table 3: Examples of English Language Proficiency and Language Arts Standards for a Fifth-Grade Student: Table 4: Percentage of Elementary Students Scoring at the Proficient Level or Above on State Mathematics Assessment for Selected Student Groups, School Year 2003-2004: Table 5: Examples of Issues Relating to Assessing Students with Limited English Proficiency Raised in Education's Peer Review Reports: Table 6: Most Frequently Cited Accommodations in 42 States: Figures: Figure 1: NCLBA's Requirements for Students with Limited English Proficiency under Title I and Title III: Figure 2: School Year 2003-2004 Comparison of Percentage of Students with Limited English Proficiency Who Achieved Proficient Scores in Language Arts and Mathematics with State-Established Progress Goals: Figure 3: Percentage of Districts in 18 Selected States Reporting Adequate Yearly Progress Results in School Year 2003-2004 for Students with Limited English Proficiency: Figure 4: Use of Native Language and Alternate Assessments for Students with Limited English Proficiency: Figure 5: Type of English Language Proficiency Assessment Administered in 2005-2006 School Year: Figure 6: Movement of Students In and Out of Limited English Proficient Student Group and Other Student Groups: Figure 7: Process for Determining Adequate Yearly Progress for a Student Group: Abbreviations: AYP: adequate yearly progress: CELLA: Comprehensive English Language Learning Assessment: ELDA: English Language Development Assessment: ESL: English as a second language: GAO: Government Accountability Office: MWAC: Mountain West Assessment Consortium: NCLBA: No Child Left Behind Act: PA EAG: Pennsylvania Enhanced Assessment Grant: SCASS: State Collaborative on Assessment and Student Standards: WIDA: World-Class Instructional Design and Assessment: United States Government Accountability Office: Washington, DC 20548: July 26, 2006: The Honorable George Miller: Ranking Minority Member: Committee on Education and the Workforce: House of Representatives: The Honorable Rubén Hinojosa: Ranking Minority Member: Subcommittee on Select Education: Committee on Education and the Workforce: House of Representatives: The Honorable Lynn Woolsey: Ranking Minority Member: Subcommittee on Education Reform: Committee on Education and the Workforce: House of Representatives: The Honorable Raúl Grijalva: House of Representatives: An estimated 5 million children with limited English proficiency were enrolled in U.S. public schools during the 2003-2004 school year, representing about 10 percent of the total school population. They speak over 400 languages, with almost 80 percent of students with limited English proficiency speaking Spanish. These students have difficulties in speaking, reading, writing, or understanding English that interfere with their ability to successfully participate in school. Because of these language barriers, obtaining information on the academic knowledge of these students from an assessment that is valid and reliable (i.e., it measures what it is designed to measure in a consistent manner) presents challenges. As a result, students with limited English proficiency have historically been excluded from statewide assessments, leaving states and districts with little information about how these students are performing academically. In 1994, the enactment of the Improving America's Schools Act required states to assess these students, to the extent practicable, in the manner most likely to yield accurate information about their academic knowledge. Subsequently, Congress passed the No Child Left Behind Act of 2001 (NCLBA) with the goal of increasing academic achievement and closing achievement gaps among different student groups. Specifically, NCLBA required states to demonstrate that all students have reached the "proficient" level on a state's language arts and mathematics assessments by 2014. States are obligated to demonstrate "adequate yearly progress" toward this goal each year--that is, they must show that increasing percentages of students are reaching proficient achievement levels over time. Students with limited English proficiency, along with other targeted student groups, must separately meet the same academic progress goals as other students. Further, NCLBA required states to annually assess the English proficiency of these students and to demonstrate that they are making progress in becoming proficient in English. Because these students are defined by a temporary characteristic--unlike other student groups targeted in NCLBA--once a state determines that students with limited English proficiency have attained English proficiency, they are no longer included in the group of students with limited English proficiency, although Education has given states some flexibility in this area. Given your interest in the academic achievement of these students and the validity and reliability of assessments used to measure their performance, we are providing information on (1) the extent to which students with limited English proficiency are meeting adequate yearly progress goals and what selected states and districts are doing to support the improved academic performance of these students, (2) what states have done to ensure that results from language arts and mathematics assessments are valid and reliable for students with limited English proficiency, (3) how states are assessing English proficiency and what they are doing to address the validity and reliability of these assessment results, and (4) how the Department of Education (Education) is supporting states' efforts to meet NCLBA's assessment requirements for these students. To determine the extent to which students with limited English proficiency are meeting adequate yearly progress goals, we collected school year 2003-2004 state-level data for 48 states, including the District of Columbia. We obtained the majority of our data from state Web sites and, when necessary, contacted state officials for these data. Three states did not publish data in a format that allowed us to determine if students with limited English proficiency had met the state's adequate yearly progress goals. We also collected additional achievement data for 2003-2004 at the school district level from 18 states. We chose a nonrandom sample of states with the largest percentage of the national population of students with limited English proficiency, states with the largest percentage increases in these students between 1990 and 2000, and included at least 2 states from each region represented by Education's regional education laboratories (with the exception of one region that included only one state). When district-level achievement data for school year 2003-2004 were not available on a state's Web site or a state had more than 500 districts, we requested the data directly from state officials; 2 states did not respond to our request for these data. We determined that the state and district data were sufficiently reliable for our purposes. We studied 5 states in depth (California, Nebraska, New York, North Carolina, and Texas) to collect detailed information from state and district officials on their assessment practices, efforts to ensure the validity and reliability of their assessments for students with limited English proficiency, and their approaches to improve the performance of these students. These 5 states had relatively large percentages of students with limited English proficiency or had experienced large increases in their populations of these students. In addition, we selected these particular states to ensure variation in geography, types of English language proficiency tests used, and use of different approaches to assessing the content knowledge of this student group. To obtain information on the assessments used by other states, we reviewed accountability workbooks and other documents that states submit to Education, available reports from state monitoring visits conducted by Education, and available peer review reports from 38 states on their assessment and accountability systems. In addition to studying 5 states, we directly contacted officials in 28 states to confirm what English language proficiency assessment they planned to administer in 2005-2006 and to discuss what guidance Education had provided regarding these assessments. We also interviewed officials from major test development companies, from state consortia that are developing English language proficiency assessments, and from Education. To assess state efforts to ensure the validity and reliability of their assessments, we reviewed national assessment standards developed by professional organizations and convened a group of experts to discuss states' efforts to develop and implement valid and reliable academic assessments for students with limited English proficiency (see app. I for more information about these experts). Finally, we obtained information from the 50 states and the District of Columbia on their use of native language assessments using a short e-mail survey. We conducted our review between June 2005 and June 2006 in accordance with generally accepted government auditing standards. Results in Brief: In school year 2003-2004, state data showed that the percentage of students with limited English proficiency scoring proficient on a state's language arts and mathematics tests was lower than the state's annual progress goals in nearly two-thirds of the 48 states for which we obtained data. To help these students progress academically, state and district officials in the 5 states we visited reported using a variety of strategies, including training teachers to incorporate language development into academic classes. Further, our review of data 49 states submitted to Education showed that the performance of students with limited English proficiency on states' mathematics assessments for elementary school students was lower than that of the total student population in all of these states but 1. Although the student groups are not mutually exclusive, in most of the 49 states, the performance of students with limited English proficiency was generally lower than that of other groups, such as economically disadvantaged students. Factors other than student academic knowledge, however, can influence whether states and districts meet their academic progress goals for students with limited English proficiency, such as how a state establishes its annual progress goals. To support improved academic progress for these students, district and state officials we spoke with in our 5 study states reported using strategies similar to those considered good practices for all students. In particular, they cited providing teacher training focused on these students, having school leadership focused on their needs and using data to target interventions as key to the success of these students. For assessments of academic knowledge in language arts and mathematics, we found that our 5 study states have taken some steps to address the specific challenges associated with assessing students with limited English proficiency. Although officials in these states reported taking steps to follow generally accepted test development procedures to ensure the validity and reliability of results from these assessments for the general student population, these assessments may not provide valid results for students with limited English proficiency. Our group of experts expressed concerns about whether all states are assessing these students in a valid and reliable manner, noting that states are not taking all the critical steps needed to do so. Although states have been required to include these students in their assessments since 1994, Education's recent peer reviews of 38 states cited 25 for not providing sufficient evidence on the validity or reliability of results for students with limited English proficiency. In 1 state, for example, procedures to develop test questions did not include an adequate check for language biases. To increase the validity and reliability of assessment results for this population, most states offered accommodations, such as offering extra time to complete the test and using a bilingual dictionary. While most states offered some accommodations, our experts indicated that research is lacking on what specific accommodations are appropriate for students with limited English proficiency. Our survey of states and review of state documents found that 16 states used statewide native language assessments for some grades and 13 states used statewide alternate assessments (such as portfolios of classroom work) in 2005 to better accommodate certain students with limited English proficiency. While such assessments may improve the validity of test scores, our group of experts noted that developing native language and alternate assessments requires resources and expertise that not all states have. Further, our experts told us that native language assessments may not provide valid results for students who are not receiving instruction in their native language. In addition, developing assessments in all languages spoken by students with limited English proficiency would likely not be practicable for most states. With respect to English language proficiency assessments, many states are still in the preliminary phases of developing and administering new assessments to measure students' progress in learning English, as required by NCLBA; as a result, complete information on the validity and reliability of these assessments is not yet available. To assess these students in the 2005-2006 school year, 22 states used new assessments developed by one of four state consortia; 8 states used customized, off-the-shelf assessments offered by testing companies; 14 states used off-the-shelf assessments; and 7 states used state- developed assessments. While a few states already had the required English language proficiency assessments in place, others will be administering these assessments for the first time in 2006; as a result, states and test developers are still collecting evidence to document the validity and reliability of the results for most of these tests. An Education-funded study by a national education research organization reviewed the available documentation for the English language proficiency assessments used by 33 states in the 2005-2006 school year and found insufficient documentation on the validity and reliability of results from these assessments. Education has offered states support and technical assistance in a variety of ways to help them appropriately assess students with limited English proficiency, such as providing training, conducting peer reviews of states' academic assessments, and providing flexibility in assessing these students. However, Education has issued little written guidance to states on developing English language proficiency assessments that meet NCLBA's requirements and on tracking the progress of students in acquiring English. Officials in about one-third of the 33 states we visited or contacted told us that they were uncertain about Education's requirements for the new English language proficiency assessments and that they wanted more guidance. In addition, our group of experts reported that some states need more assistance to develop language arts and mathematics assessments that provide valid measures of the academic knowledge of this group of students. To support states' efforts to incorporate these students into their accountability systems for academic performance, Education has offered states some flexibilities in how they track progress goals for these students. For example, students who have been in the United States for less than 1 year do not have to be assessed for language arts. However, several state and district officials in the 5 states we studied told us that additional flexibility, such as excluding students from testing for a longer period, is needed to ensure that adequate yearly progress measures accurately track the academic progress of these students. To help states improve their assessment of students with limited English proficiency, we are recommending that the Secretary of Education (1) support additional research on accommodations, (2) identify and provide additional technical support states need to ensure the validity and reliability of academic assessments for these students, (3) publish more detailed guidance on assessing the English language proficiency of these students, and (4) explore ways to provide additional flexibility with respect to measuring annual progress for these students. In its comments, Education generally agreed with our recommendations. Background: Students with limited English proficiency are a diverse and complex group. They speak many languages and have a tremendous range of educational needs and include refugees with little formal schooling and students who are literate in their native languages. Accurately assessing the academic knowledge of these students in English is challenging. If a student responds incorrectly to a test item, it may not be clear if the student did not know the answer or misunderstood the question because of language barriers. Several approaches are available to allow students to demonstrate their academic knowledge while they are becoming proficient in English, although each poses challenges. First, a state can offer assessments in a student's native language. However, vocabulary in English is not necessarily equivalent in difficulty to the vocabulary in another language. As a result, a test translated from English may not have the same level of difficulty as the English version. If a state chooses to develop a completely different test in another language instead of translating the English version, the assessment should measure the same standards and reflect the same level of difficulty as the English version of the test to ensure its validity. Second, states can also offer accommodations, such as providing extra time to take a test, allowing the use of a bilingual dictionary, reading test directions aloud in a student's native language, or administering the test in a less distracting environment. Accommodations alter the way a regular assessment is administered, with the goal of minimizing the language impediments faced by students with limited English proficiency; they are intended to level the playing field without providing an unfair advantage to these students. Finally, states can use alternate assessments that measure the same things as the regular assessment while minimizing the language burden placed on the student. For example, an alternate assessment can be a traditional standardized test that uses simplified English or relies more on pictures and diagrams. It can also be a portfolio of a student's class work that demonstrates academic knowledge. In either case, studies would be needed to demonstrate that the alternate assessment is equivalent to the regular assessment. NCLBA Requirements: Title I of NCLBA seeks to ensure that all children have a fair and equal opportunity to obtain a high-quality education and become proficient in academic subjects. It requires states to administer tests in language arts and mathematics to all students in certain grades and to use these tests as the primary means of determining the annual performance of states, districts, and schools. These assessments must be aligned with the state's academic standards--that is, they must measure how well a student has demonstrated his or her knowledge of the academic content represented in these standards. States are to show that increasing percentages of students are reaching the proficient level on these state tests over time. NCLBA also requires that students with limited English proficiency receive reasonable accommodations and be assessed, to the extent practicable, in the language and form most likely to yield accurate data on their academic knowledge. Somewhat similar versions of these provisions, such as reporting testing results for different student groups, had been included in legislation enacted in 1994. One new NCLBA requirement was for states to annually assess the English language proficiency of students identified as having limited English proficiency. Table 1 summarizes some key Title I provisions from NCLBA. Table 1: Selected Provisions from Title I of NCLBA: State academic assessments; Beginning in the 2005-2006 school year, states must implement annual, high-quality state assessments in reading and mathematics in grades 3-8 and at least once in high school.[A] These assessments must be aligned with challenging state academic standards and must be "consistent with relevant, nationally recognized professional and technical standards for such assessments" and used in ways that are valid and reliable. States must provide for the participation of all students, including those with limited English proficiency. Academic assessment provisions related to students with limited English proficiency; Students with limited English proficiency are to be assessed in a valid and reliable manner. In addition, they must be provided with reasonable accommodations and be assessed, to the extent practicable, "in the language and form most likely to yield accurate data" on their academic knowledge. In addition, for language arts, students who have been in U.S. schools for 3 years or more generally must be assessed in English. Adequate yearly progress; States must set annual goals that lead to all students achieving proficiency in language arts and mathematics by 2014. To be deemed as having made adequate yearly progress for a given year, each district and school must show that the requisite percentage of each designated student group, as well as the student population as a whole, met the state proficiency goal (that is, the percentage of students who have achieved the proficient level on the state's assessments). Schools must also show that at least 95 percent of students in each designated student group participated in these assessments. Further, schools must also demonstrate that they have met state targets on other academic indicators--graduation rates in high school or attendance or other measures in elementary and middle schools. Alternatively, a district or school can make adequate yearly progress through the "safe harbor" provision, if the percentage of students in a group considered not proficient decreased by at least 10 percent from the preceding year and the group made progress on one of the state's other academic indicators. Actions when adequate yearly progress not achieved; Schools that receive funding under Title I of NCLBA must take specified actions if they do not meet state progress goals. Specifically, schools that do not make adequate yearly progress for 2 consecutive years or more are identified for improvement and must, among other things, offer parents an opportunity to transfer students to another school and provide supplemental services (e.g., tutoring). Those that miss the annual goals for additional years are identified for successive stages of intervention, including corrective action and restructuring. State English language proficiency assessments; States must annually assess the English language proficiency of all students with limited English proficiency, measuring students' oral language, reading, and writing skills in English. Source: Pub. L. No. 107-110. [A] Beginning in school year 2007-2008, states must implement similar assessments in science. [End of table] Accurately assessing the academic knowledge of students with limited English proficiency has become more critical because NCLBA designated specific groups of students for particular focus. These four groups are students who (1) are economically disadvantaged, (2) represent major racial and ethnic groups, (3) have disabilities, and (4) are limited in English proficiency. These groups are not mutually exclusive, so that the results for a student who is economically disadvantaged, Hispanic, and has limited English proficiency could be counted in all three groups. States and school districts are required to measure the progress of all students in meeting academic proficiency goals, as well as to measure separately the progress of these designated groups. To be deemed as having made adequate yearly progress, generally each district and school must show that each of these groups met the state proficiency goal (that is, the percentage of students who have achieved the proficient level on the state's assessments) and that at least 95 percent of students in each designated group participated in these assessments. Although NCLBA placed many new requirements on states, states have broad discretion in many key areas. States establish their academic content standards and then develop their own tests to measure the academic content students are taught in school. States also set their own standards for what constitutes proficiency on these assessments. In addition, states set their own annual progress goals for the percentage of students achieving proficiency, using guidelines outlined in NCLBA.[Footnote 1] Title III of NCLBA focuses specifically on students with limited English proficiency, with the purpose of ensuring that these students attain English proficiency and meet the same academic content standards all students are expected to meet. This title established new requirements intended to hold states and districts accountable for student progress in attaining English proficiency. It requires states to establish goals to demonstrate, among other things, annual increases in (1) students making progress in learning English and (2) students attaining English proficiency. Specifically, states must establish English language proficiency standards that are aligned with a state's academic content standards. The purpose of these alignment requirements is to ensure that students are acquiring the academic language they will need to successfully participate in the classroom. Education also requires that a state's English language proficiency assessment be aligned to its English language proficiency standards. While NCLBA requires states to administer academic assessments to students in specific grades, it requires states to administer an annual English language proficiency assessment to all students with limited English proficiency, from kindergarten to grade 12. See table 2 for summary of key Title III provisions. Table 2: Selected Provisions from Title III of NCLBA: State English language proficiency standards; States must establish English language proficiency standards that are aligned with the state's challenging academic content standards. Tracking student progress in learning English; States must establish objectives for improving students' English proficiency in four areas: speaking, listening, reading, and writing.[A] States receiving grants under Title III must establish annual goals for increasing and measuring the progress of students with limited English proficiency in (1) learning English, (2) attaining English proficiency, and (3) meeting adequate yearly progress goals in attaining academic proficiency outlined in Title I. Actions when annual goals for students with limited English proficiency not met; Districts that receive funding under Title III are subject to certain consequences if they do not meet a state's annual Title III goals. If a district does not meet the goals for 2 consecutive years, it must develop an improvement plan that addresses the factors that prevented the district from meeting the goals. If a district does not meet the goals for 4 consecutive years, it must modify its curriculum and method of instruction or the state must determine whether to continue to fund the district and require the district to replace all personnel related to the district's inability to meet the goals. Source: Pub. L. No. 107-110. [A] Title I refers to oral language skills, which encompass listening and speaking. [End of table] Language arts standards define the academic skills a student is expected to master, while English language proficiency standards define progressive levels of competence in the acquisition of English necessary to participate successfully in the classroom. Examples of standards for English language proficiency and language arts are provided in table 3. Table 3: Examples of English Language Proficiency and Language Arts Standards for a Fifth-Grade Student: English language proficiency standards: The student can comprehend reading passages written in familiar or short sentence patterns and verbalize some of the main points of the passage; Language arts standards: The student can independently read and comprehend a grade- level appropriate text and write a short essay describing the main idea of the text. English language proficiency standards: The student can use acquired knowledge of the English language to learn and understand new vocabulary in context; Language arts standards: The student can apply knowledge of reading strategies to comprehend the text of the next higher level of difficulty. Source: U.S. Department of Education, "Final Non-Regulatory Guidance on the Title III State Formula Grant Program--Standards, Assessments and Accountability," February 2003. [End of table] Under NCLBA, states, districts, and schools have two sets of responsibilities for students with limited English proficiency. As shown in figure 1, they are responsible for ensuring that these students make progress in learning English under Title III and that they become proficient in language arts and mathematics under Title I. Beginning with the 2004-2005 school year, Education is required to annually review whether states have made adequate yearly progress (as defined by the state) for each of the student groups and have met their objectives for increasing the number or percentage of students who become proficient in English. Figure 1: NCLBA's Requirements for Students with Limited English Proficiency under Title I and Title III: [See PDF for image] Source: Adapted from U.S. Department of Education, "Final Non- Regulatory Guidance on the Title III State Formula Grant Program- Standards, Assessments and Accountability, February 2003. [End of figure] Test Development: NCLBA's emphasis on validity and reliability reflects the fact that these concepts are among the most important in test development. Validity refers to whether the test measures what it is intended to measure. Reliability refers to whether or not a test yields consistent results across time and location and among different sections of the test. A test cannot be considered valid if it is unreliable. The Standards for Educational and Psychological Testing provide universally accepted guidance for the development and evaluation of high-quality, psychometrically sound assessments.[Footnote 2] They outline specific standards to be considered when assessing individuals with limited English proficiency, including (1) determining when language differences produce threats to the validity and reliability of test results, (2) providing information on how to use and interpret results when tests are used with linguistically diverse individuals, and (3) collecting the same evidence to support claims of validity for each linguistic subgroup as was collected for the population as a whole. Test development begins with determining the purpose of the test and the content to be measured by the test. NCLBA outlines several purposes of statewide assessments, including determining the yearly performance of schools and districts, interpreting individual student academic needs, and tracking the achievement of several groups of students. NCLBA requires that the content of statewide assessments reflects state standards in language arts and mathematics, but the specific skills measured can vary from state to state. For example, a language arts assessment could measure a student's knowledge of vocabulary or ability to write a persuasive essay. Variations in purpose and content affect test design, as well as the analyses necessary to determine validity and reliability. After determining the purpose and content of the test, developers create test specifications, which delineate the format of the questions and responses, as well as the scoring procedures. Specifications may also indicate additional information, such as the intended difficulty of questions, the student population that will take the test, and the procedures for administering the test. These specifications subsequently guide the development of individual test questions. The quality of the questions is usually ascertained through review by knowledgeable educators and statistical analyses based on a field test of a sample of students--ideally the sample is representative of the overall target student population so the results will reflect how the questions will function when the test is administered to the population. These reviews typically evaluate a question's quality, clarity, lack of ambiguity, and sometimes its sensitivity to gender or cultural issues; they are intended to ensure that differences in student performance are related to differences in student knowledge rather than other factors, such as unnecessarily complex language. Once the quality has been established, developers assemble questions into a test that meets the requirements of the test specifications. Developers often review tests after development to ensure that they continue to produce accurate results. Education's Responsibilities: Education has responsibility for general oversight of Titles I and III of NCLBA. The department's Office of Elementary and Secondary Education oversees states' implementation of Title I requirements with respect to academic assessments and making adequate progress toward achieving academic proficiency for all students by 2014. Education's Office of English Language Acquisition, Language Enhancement and Academic Achievement for Limited English Proficient Students oversees states' Title III responsibilities, which include administering annual English language proficiency assessments to students with limited English proficiency and demonstrating student progress in attaining English language proficiency. Students with Limited English Proficiency Performed below Progress Goals in 2004 in Two-Thirds of States, but States We Studied Are Working to Improve Student Academic Performance: In school year 2003-2004, the percentage of students with limited English proficiency reported by states as scoring proficient on a state's language arts and mathematics tests was lower than the state's annual progress goals (established for all students) in nearly two- thirds of the 48 states for which we obtained data.[Footnote 3] Further, data from state mathematics tests showed that these students were generally achieving lower rates of academic proficiency than the total student population. However, factors other than student academic performance can influence whether a state meets its progress goals, such as which students a state includes in the limited English proficient group and how a state establishes its annual progress goals. Officials in our study states reported using several common approaches, including providing teacher training specific to the needs of limited English proficient students and using data to guide instruction and identify areas for improvement. In Almost Two-Thirds of States, the Percentage of Students with Limited English Proficiency Achieving Proficient Scores Was Below the State's Annual Progress Goals: In nearly two-thirds of the 48 states for which we obtained data, state data showed that the percentage of students with limited English proficiency scoring proficient on language arts and mathematics tests was below the annual progress goal set by the state for school year 2003-2004. Students with limited English proficiency met academic progress goals in language arts and mathematics in 17 states.[Footnote 4] In 31 states, state data indicated that these students missed the goals either for language arts or for both language arts and mathematics (see fig. 2). In 21 states, the percentage of proficient students in this group was below both the mathematics and the language arts proficiency goals. See appendix II for information on how adequate yearly progress measures are calculated. Figure 2: School Year 2003-2004 Comparison of Percentage of Students with Limited English Proficiency Who Achieved Proficient Scores in Language Arts and Mathematics with State-Established Progress Goals: [See PDF for image] Source: State 2003-2004 report cards available on state Web sites or data provided by state officials. Notes: We obtained data for 42 states from their state Web sites and contacted state officials in 6 states to obtain these data. Three states did not report data in a format that allowed us to determine whether the percentage of students with limited English proficiency met or exceeded the annual progress goals established by the state. When states reported proficiency data for different grades or groups of grades, we determined that students with limited English proficiency met a state's progress goals if the student group met all proficiency and participation goals for all grades reported. An Education official told us that a state could not make adequate yearly progress if it missed one of the progress goals at any grade level. All of the states on the map where the proficiency percentage for students with limited English proficiency met or exceeded the state's annual progress goal also met NCLBA's participation goals. We incorporated states' use of confidence intervals and NCLBA's safe harbor provision in determining whether the percentage of students with limited English proficiency achieving proficient scores met or exceeded a state's progress goals. If a state's published data did not explicitly include such information, we contacted state officials to ensure that the state did not meet its progress goals through the use of confidence intervals or through NCLBA's safe harbor provision. In the following seven states, the percentage of students with limited English proficiency was below the state's annual progress goal for language arts or for both language arts and mathematics, but the student group met the state's requirements for progress through the safe harbor provision: Delaware, Idaho, Maryland, Massachusetts, Oklahoma, Rhode Island, and Utah. We reported 2004-2005 school year data for Oklahoma, New Mexico and Utah because we could not obtain data for the 2003-2004 school year. Data from Iowa, Massachusetts, and Rhode Island are for the 2002-2004 school years. Rhode Island did not separately report participation rates for students with limited English proficiency. Instead, it reported that all students met the 95 percent participation goal. [End of figure] We also obtained additional data from 18 states to determine whether districts were meeting annual progress goals for students with limited English proficiency in school year 2003-2004.[Footnote 5] In 14 of the 18 states, however, we found that less than 40 percent of the districts in each state reported separate results for this group of students (see fig. 3)[Footnote 6]. Districts only have to report progress results for a student group if a minimum number of students are included in the group[Footnote 7]. In Nebraska, for example, only 4 percent of districts reported progress goals for students with limited English proficiency. Except for Florida, Hawaii, and Nevada, less than half of the districts in each state reported separate results for this group of students. Even when districts do not have to report on students with limited English proficiency, however, the test scores for these students are included in the state's overall progress measures. Figure 3: Percentage of Districts in 18 Selected States Reporting Adequate Yearly Progress Results in School Year 2003-2004 for Students with Limited English Proficiency: [See PDF for image] Source: GAO analysis of district report cards and district data provided by state officials. Notes: If a district reported annual progress results for students with limited English proficiency in either language arts proficiency or mathematics proficiency, or both, we considered that the district reported adequate yearly progress results for the student group. Hawaii has only one school district. Since the state reported separate results for students with limited English proficiency, it has been included as 100 percent of districts reporting separate results for these students. [End of figure] For those districts that reported results for students with limited English proficiency, district-level data showed that most districts in 13 of the 18 states met their mathematics progress goals for these students. For example, 67 percent of reporting districts in Nebraska and 99 percent of reporting districts in Texas met the state's goals. In 4 states, less than half of the districts reporting results for these students met the state mathematics progress goals. Specifically, 26 percent of Alaska districts, 33 percent of Nevada districts, 48 percent of Oregon districts, and 48 percent of Florida districts met these goals. (See app. III for results from each of the 18 states.) In addition to looking at whether students with limited English proficiency met annual progress goals at the state and district level, we also examined achievement levels on state assessments for this group of students compared with the total student population (which also includes students with limited English proficiency). Looking at mathematics results reported by 49 states to Education, for example, in all but one state, we found that a lower percentage of students with limited English proficiency at the elementary school level achieved proficient scores, compared to the total student population in school year 2003-2004 (see app. IV for the results reported by the 49 states). Twenty-seven states reported that the total student population outperformed students with limited English proficiency by 20 percentage points or more. The differences among groups in the percentage of students achieving proficient scores varied across states. South Dakota, for example, reported a large achievement gap, with 37 percent of limited English proficient students scoring at the proficient level, compared to 78 percent for the entire student population. The gap was less pronounced in Texas, where 75 percent of students with limited English proficiency achieved proficient scores on the mathematics assessment, while 85 percent of the total student population did. In Louisiana, these students performed about the same as the total student population, with 58 percent of limited English proficient students scoring at the proficient level on the elementary mathematics assessment, compared to 57 percent of the total student population. We also found that, in general, a lower percentage of students with limited English proficiency achieved proficient test scores than other selected student groups (see table 4). All of the 49 states reported that these students achieved lower rates of proficiency than white students.[Footnote 8] The performance of limited English proficient students relative to the other student groups varied. In 37 states, for example, economically disadvantaged students outperformed students with limited English proficiency, while students with disabilities outperformed these students in 14 states. In 12 states, all the selected student groups outperformed students with limited English proficiency. Table 4: Percentage of Elementary Students Scoring at the Proficient Level or Above on State Mathematics Assessment for Selected Student Groups, School Year 2003-2004: States: Ala; Students with Limited English Proficiency: 53; Students with Disabilities: 31; African-American: 58; Economically Disadvantaged: 62; Hispanic: 61; White: 81. States: Alaska; Students with Limited English Proficiency: 40; Students with Disabilities: 36; African-American: 50; Economically Disadvantaged: 50; Hispanic: 63; White: 77. States: Ariz.[A]; Students with Limited English Proficiency: 32; Students with Disabilities: 31; African-American: 46; Economically Disadvantaged: data not available; Hispanic: 44; White: 72. States: Ark.[B]; Students with Limited English Proficiency: 49; Students with Disabilities: 24; African-American: 38; Economically Disadvantaged: 53; Hispanic: 56; White: 74. States: Calif; Students with Limited English Proficiency: 33; Students with Disabilities: 24; African-American: 28; Economically Disadvantaged: 33; Hispanic: 33; White: 61. States: Colo.[A]; Students with Limited English Proficiency: 76; Students with Disabilities: 61; African-American: 74; Economically Disadvantaged: 79; Hispanic: 79; White: 94. States: Conn; Students with Limited English Proficiency: 47; Students with Disabilities: 49; African-American: 69; Economically Disadvantaged: 61; Hispanic: 61; White: 88. States: Del.[A]; Students with Limited English Proficiency: 70; Students with Disabilities: 47; African-American: 61; Economically Disadvantaged: 67; Hispanic: 74; White: 87. States: D.C; Students with Limited English Proficiency: 34; Students with Disabilities: 14; African-American: 49; Economically Disadvantaged: 48; Hispanic: 57; White: 89. States: Fla; Students with Limited English Proficiency: 48; Students with Disabilities: 39; African-American: 44; Economically Disadvantaged: 52; Hispanic: 60; White: 74. States: Ga; Students with Limited English Proficiency: 53; Students with Disabilities: 46; African-American: 65; Economically Disadvantaged: 66; Hispanic: 67; White: 84. States: Hawaii[A]; Students with Limited English Proficiency: 9; Students with Disabilities: 6; African-American: 19; Economically Disadvantaged: 18; Hispanic: 16; White: 36. States: Idaho; Students with Limited English Proficiency: 62; Students with Disabilities: 55; African-American: 69; Economically Disadvantaged: 76; Hispanic: 68; White: 87. States: Ill.[A]; Students with Limited English Proficiency: 53; Students with Disabilities: 57; African-American: 54; Economically Disadvantaged: 61; Hispanic: 64; White: 89. States: Ind.[A]; Students with Limited English Proficiency: 47; Students with Disabilities: 40; African-American: 54; Economically Disadvantaged: 60; Hispanic: 60; White: 75. States: Iowa; Students with Limited English Proficiency: 49; Students with Disabilities: 39; African-American: 46; Economically Disadvantaged: 62; Hispanic: 56; White: 80. States: Kans; Students with Limited English Proficiency: 58; Students with Disabilities: 67; African-American: 61; Economically Disadvantaged: 70; Hispanic: 65; White: 84. States: Ky.[A]; Students with Limited English Proficiency: 32; Students with Disabilities: 29; African-American: 28; Economically Disadvantaged: 36; Hispanic: 38; White: 51. States: La; Students with Limited English Proficiency: 58; Students with Disabilities: 30; African-American: 40; Economically Disadvantaged: 48; Hispanic: 62; White: 74. States: Maine; Students with Limited English Proficiency: 10; Students with Disabilities: 13; African-American: 15; Economically Disadvantaged: 20; Hispanic: 20; White: 33. States: Md; Students with Limited English Proficiency: 39; Students with Disabilities: 39; African-American: 52; Economically Disadvantaged: 51; Hispanic: 59; White: 83. States: Mass; Students with Limited English Proficiency: 22; Students with Disabilities: 20; African-American: 20; Economically Disadvantaged: 22; Hispanic: 19; White: 49. States: Mich; Students with Limited English Proficiency: 59; Students with Disabilities: 42; African-American: 51; Economically Disadvantaged: 57; Hispanic: 58; White: 77. States: Minn.[A]; Students with Limited English Proficiency: 38; Students with Disabilities: 45; African-American: 39; Economically Disadvantaged: 52; Hispanic: 45; White: 77. States: Miss; Students with Limited English Proficiency: 79; Students with Disabilities: 61; African-American: 69; Economically Disadvantaged: 72; Hispanic: 87; White: 91. States: Mo; Students with Limited English Proficiency: 30; Students with Disabilities: 24; African-American: 24; Economically Disadvantaged: 28; Hispanic: 29; White: 45. States: Mont; Students with Limited English Proficiency: 15; Students with Disabilities: 22; African-American: 32; Economically Disadvantaged: 33; Hispanic: 36; White: 49. States: Nebr; Students with Limited English Proficiency: 73; Students with Disabilities: 65; African-American: 72; Economically Disadvantaged: 79; Hispanic: 80; White: 90. States: Nev.[A]; Students with Limited English Proficiency: 22; Students with Disabilities: 22; African-American: 27; Economically Disadvantaged: 32; Hispanic: 32; White: 57. States: N.H.[A]; Students with Limited English Proficiency: 57; Students with Disabilities: 61; African-American: 72; Economically Disadvantaged: 71; Hispanic: 67; White: 85. States: N.J; Students with Limited English Proficiency: 47; Students with Disabilities: 46; African-American: 50; Economically Disadvantaged: 54; Hispanic: 59; White: 81. States: N.Mex; Students with Limited English Proficiency: 36; Students with Disabilities: 31; African-American: 50; Economically Disadvantaged: 47; Hispanic: 49; White: 72. States: N.C; Students with Limited English Proficiency: 86; Students with Disabilities: 75; African-American: 88; Economically Disadvantaged: 89; Hispanic: 90; White: >95. States: N.Dak; Students with Limited English Proficiency: 30; Students with Disabilities: 38; African-American: 45; Economically Disadvantaged: 52; Hispanic: 47; White: 68. States: Ohio; Students with Limited English Proficiency: 48; Students with Disabilities: 38; African-American: 39; Economically Disadvantaged: 48; Hispanic: 51; White: 72. States: Okla.[A]; Students with Limited English Proficiency: 41; Students with Disabilities: 23; African-American: 38; Economically Disadvantaged: 46; Hispanic: 47; White: 61. States: Oreg.[A]; Students with Limited English Proficiency: 61; Students with Disabilities: 57; African-American: 71; Economically Disadvantaged: 73; Hispanic: 63; White: 86. States: Pa.[A]; Students with Limited English Proficiency: 34; Students with Disabilities: 27; African-American: 30; Economically Disadvantaged: 42; Hispanic: 38; White: 70. States: R.I; Students with Limited English Proficiency: 23; Students with Disabilities: 34; African-American: 32; Economically Disadvantaged: 36; Hispanic: 31; White: 60. States: S.C; Students with Limited English Proficiency: 16; Students with Disabilities: 16; African-American: 19; Economically Disadvantaged: 22; Hispanic: 24; White: 49. States: S.Dak; Students with Limited English Proficiency: 37; Students with Disabilities: 48; African-American: 56; Economically Disadvantaged: 65; Hispanic: 62; White: 83. States: Tex; Students with Limited English Proficiency: 75; Students with Disabilities: 76; African-American: 75; Economically Disadvantaged: 80; Hispanic: 81; White: 93. States: Utah; Students with Limited English Proficiency: 53; Students with Disabilities: 43; African-American: 56; Economically Disadvantaged: 71; Hispanic: 52; White: 78. States: Vt; Students with Limited English Proficiency: 67; Students with Disabilities: 37; African-American: 51; Economically Disadvantaged: 58; Hispanic: data not available; White: 73. States: Va.[A]; Students with Limited English Proficiency: 84; Students with Disabilities: 74; African-American: 77; Economically Disadvantaged: 79; Hispanic: 84; White: 92. States: Wash; Students with Limited English Proficiency: 27; Students with Disabilities: 29; African-American: 38; Economically Disadvantaged: 44; Hispanic: 39; White: 66. States: W.Va; Students with Limited English Proficiency: 68; Students with Disabilities: 33; African-American: 53; Economically Disadvantaged: 64; Hispanic: 68; White: 70. States: Wisc; Students with Limited English Proficiency: 50; Students with Disabilities: 51; African-American: 45; Economically Disadvantaged: 56; Hispanic: 53; White: 80. States: Wyo; Students with Limited English Proficiency: 15; Students with Disabilities: 21; African-American: 25; Economically Disadvantaged: 29; Hispanic: 24; White: 42. States: Number of states where students with limited English proficiency had lower proficiency levels than this group (bolded numbers); Students with Disabilities: 14; African-American: 28; Economically Disadvantaged: 37; Hispanic: 41; White: 49. Source: Consolidated State Performance Reports, 2003-04 school year. Notes: Bolded numbers indicate when the percentage of students in a group achieving proficient scores is greater than the percentage of students with limited English proficiency achieving proficient scores. Student groups are not mutually exclusive, so that results for a student who is both Hispanic and economically disadvantaged appear in both groups. New York and Tennessee did not provide assessment data in their State Consolidated Performance Reports for the 2003-2004 school year. [A] Most states reported assessment data for students in the fourth grade. States marked with this superscript reported on some other grade at the elementary school level, usually either third grade or fifth grade. [B] The percentages for Arkansas do not include those students with limited English proficiency or students with disabilities who were considered proficient based on the state's portfolio assessment. [End of table] Factors beyond Student Performance Influence Progress Measures Reported by States: Factors beyond student performance can influence the number of states, districts, and schools meeting progress goals for students with limited English proficiency. One factor that can affect a state or district's ability to meet progress goals for this student group is the criteria states use to determine which students are counted as limited English proficient. Some states define limited English proficiency so that students may be more likely to stay in the group for a longer time, giving them more of an opportunity to develop the language skills necessary to demonstrate their academic knowledge on state academic assessments administered in English. On the basis of our review of state accountability plans, we found that some states removed students from the group after they have achieved proficiency on the state's English language proficiency assessment, while other states continued to include these students until they met additional academic requirements, such as achieving proficient scores on the state's language arts assessment. A number of states measured adequate yearly progress for students with limited English proficiency by including test scores for students for a set period of time after they were considered proficient in English, following Education's policy announcement in February 2004 allowing such an approach. How rigorously a state defines the proficient level of academic achievement can also influence the ability of states, districts, and schools to meet their progress goals. States with less rigorous definitions of proficiency are more likely to meet their progress goals for students with limited English proficiency or any other student group than states with more stringent definitions. Comparing the performance of students from different states on a nationally administered assessment suggests that states differ in how rigorously they define proficiency. For example, eighth-grade students in Colorado and Missouri achieved somewhat similar scores on the National Assessment of Educational Progress in mathematics in 2003.[Footnote 9] Specifically, 34 percent of Colorado students scored proficient or above on this national assessment compared to 28 percent of Missouri students. On their own state assessments in 2003, however, 67 percent of Colorado students scored proficient or above, compared to just 21 percent in Missouri.[Footnote 10] These results may reflect, among other things, a difference in the level of rigor in the tests administered by these states. However, they may also be due in part to differences in what the national test measures versus what each of the state tests measure. The likelihood of a state, district, or school meeting its annual progress goals also depends, in part, on the proficiency levels of its students when NCBLA was enacted, as well as how the state sets its annual goals. States vary significantly in the percentage of students scoring at the proficient level on their academic assessments, so that states with lower proficiency levels must, on average, establish larger annual increases in proficiency levels to meet the 2014 goal. Some states planned for large increases every 2 to 3 years, while others set smaller annual increases. States that established smaller annual increases in their initial proficiency goals may be more likely to meet their progress goals at this time, compared with states that set larger annual increases. The use of statistical procedures, such as confidence intervals, can also affect whether a state, district, or school meets its progress goals. Education officials said that states use such procedures to improve the reliability of determinations about the performance of districts. According to some researchers, such methods may improve the validity of results because they help to account for the effect of small group sizes and year-to-year changes in student populations.[Footnote 11] Most states currently use some type of confidence interval to determine if a state or district has met its progress goals, according to the Center on Education Policy.[Footnote 12] A confidence interval establishes a range of proficiency levels around a state's annual progress goal.[Footnote 13] If the percentage of students with limited English proficiency scoring proficient on a state's academic assessments falls within that range, that group has made the annual progress goal. States and Districts We Visited Have Taken Steps to Improve Performance of Students with Limited English Proficiency: To help students with limited English proficiency progress academically, state and district officials in our 5 study states reported using somewhat similar strategies, many of which are also considered good practices for all students. Among the key factors cited by state and district officials for their success in working with this group were: * strong professional development focused on effective teaching strategies for students with limited English proficiency; * school or district leadership that focuses on the needs of these students, such as providing sufficient resources to meet those needs and establishing high academic standards for these students; * "data driven" decisions, such as using data strategically to identify students who are doing well and those who need more help, to identify effective instructional approaches, or to provide effective professional development; and: * efforts to work with parents to support the academic progress of their children. These approaches are similar to those used by "blue ribbon" schools-- schools identified by Education as working successfully with all students to achieve strong academic outcomes. The qualities shared by these blue ribbon schools include professional development related to classroom instruction, strong school leadership and a vision that emphasizes high academic expectations and academic success for all students, using data to target instructional approaches, and parental involvement. While many blue ribbon schools have a high percentage of disadvantaged students, including those with limited English proficiency, their common approaches help them achieve student outcomes that place them among the top 10 percent of all schools in the state or that demonstrate dramatic improvement. Officials in all 5 of our study states stressed the importance of providing teachers with the training they need to work effectively with students with limited English proficiency. For example, state officials in North Carolina told us that they are developing a statewide professional development program to train mainstream teachers to present academic content material so that it is more understandable to students with limited English proficiency and to incorporate language development while teaching subjects such as mathematics and science. In one rural North Carolina school district where students with limited English proficiency have only recently become a large presence, district officials commented that this kind of professional development has helped teachers become more comfortable with these students and given them useful strategies to work more effectively with them. In 4 of our study states, officials emphasized the need for strong school or district leadership that focuses on the needs of students with limited English proficiency. For example, officials in a California school district with a high percentage of students with limited English proficiency told us that these students are a district priority and that significant resources are devoted to programs for them. The district administration has instilled the attitude that students with limited English proficiency can meet high expectations and are the responsibility of all teachers. To help maintain the focus on these students, the district has created an English language development progress profile to help teachers track the progress of each student in acquiring English and meeting the state's English language development standards. In addition, officials in 4 of our study states attributed their success in working with students with limited English proficiency to using data strategically, for example, to identify effective practices and guide instruction. At one California school we visited, officials reviewed test scores to identify areas needing improvement for different classes and different student groups and to identify effective practices. In addition, they reviewed test data for each student to identify areas of weakness. If test data showed that a student was having trouble with vocabulary, the teacher would work in class to build the student's vocabulary. Similarly, officials in a New York school reported that they followed student test scores over 3 years to set goals for different student groups and identify areas in need of improvement. Officials in 3 states we visited also cited the importance of involving parents of students with limited English proficiency in their children's education. In Nebraska, for example, a technical assistance agency implemented a family literacy program to help parents and their children improve their English, and also to involve parents in their children's education. The program showed parents how they can help children with their homework and the importance of reading to their children in their native language to develop their basic language skills. At a New York middle school, officials told us that they use a parent coordinator to establish better communication with families, learn about issues at home that can affect the student's academic performance, and help families obtain support services, if needed. Selected States Considered Language Issues When Developing Academic Assessments, but Validity and Reliability Concerns Remain: For academic assessments in language arts and mathematics, officials in the 5 states we studied reported that they have taken some steps, such as reviewing test items to eliminate unnecessarily complex language, to address the specific challenges associated with assessing students with limited English proficiency. However, Education recently reviewed the assessment documentation of 38 states and noted some concerns related to using these assessments for students with limited English proficiency. Our group of experts also indicated that states are generally not taking the appropriate set of comprehensive steps to create assessments that produce valid and reliable results for students with limited English proficiency. To increase the validity and reliability of assessment results for this population, most states offered accommodations, such as providing extra time to complete the assessment and offering native language assessments. However, offering accommodations may or may not improve the validity of test results, as research on the appropriate use of accommodations for these students is lacking. In addition, native language assessments are not appropriate for all students with limited English proficiency and are difficult and expensive to develop. States Reported Efforts to Improve Validity of Assessment Results for Students with Limited English Proficiency: Officials in the 5 states we studied reported taking some steps to address the specific challenges associated with assessing students with limited English proficiency in language arts and mathematics. Officials in 4 of these states reported following generally accepted test development procedures when developing their academic assessments, while a Nebraska official reported that the state expects districts to follow such procedures when developing their tests. Test development involves a structured process with specific steps; however, additional steps and special attention to language issues are required when developing a test that includes students with limited English proficiency to ensure that the results are valid and reliable for these students. As the Standards for Educational and Psychological Testing notes, for example, the test instructions or the response format may need to be modified to ensure that the test provides valid information about the skills of students with limited English proficiency. Officials in 2 states and at several testing companies mentioned that they have been focusing more on the needs of these students in recent years. Officials in California, New York, North Carolina, and Texas told us that they try to implement the principles of universal design, which support making assessments accessible to the widest possible range of students. This is done by ensuring, among other things, that instructions, forms, and questions are clear and not more linguistically complex than necessary. In addition, officials in all 5 states we studied told us they included students with limited English proficiency in the field testing of assessments. North Carolina officials reported that they oversample for students with limited English proficiency to ensure that these students are adequately represented in the field tests. Another step officials in some states reported taking is assembling panels or committees to review test items for bias and testing data for bias related to a student's English proficient status. For example, Texas and North Carolina officials reported creating review committees to ensure that test items are accessible to students with limited English proficiency. Specifically, when developing mathematics items, these states try to make the language as clear as possible to ensure that the item is measuring primarily mathematical concepts and to minimize the extent to which it is measuring language proficiency. A mathematics word problem involving subtraction, for example, might refer to fish rather than barracuda. Officials in 4 of our study states told us they used a statistical approach to evaluate test items for bias against specific student groups, and three of these reported using it to detect bias related to students with limited English proficiency. However, this type of analysis can only be used when a relatively large number of students in the specific group is taking the test. Members of our expert group recommended the use of this technique for states with a large enough number of students with limited English proficiency; however, one member noted that this technique may not be appropriate if a state's population of students with limited English proficiency is diverse but is treated as homogenous in the analyses. Some of our study states also reported including experts on limited English proficiency or English as a second language (ESL) issues in the development and review of test items, although only 1 reported involving them in all aspects of test development. In North Carolina, for example, officials told us that ESL teachers and supervisors are involved in reviewing all aspects of the test development process, including item writing, field testing, and operational testing. Some state officials also told us that they included education staff involved with students with limited English proficiency in the development of assessments. Both Education's Peer Reviews and Our Group of Experts Raised Concerns Regarding State Efforts to Ensure Valid and Reliable Results: Education's recent NCLBA peer reviews of 38 states[Footnote 14] found that 25 did not provide sufficient evidence on the validity or reliability of results for students with limited English proficiency, although states have been required to include these students in their assessments since 1994.[Footnote 15] For example, peer reviewers found that Alabama's documentation did not include sufficient evidence on the selection process for committee members to review test items for bias, noting that no evidence was provided on whether representatives for students with limited English proficiency were included. In Idaho, peer reviewers commented that the state did not report reliability data for student groups, including students with limited English proficiency. See table 5 for further examples. Table 5: Examples of Issues Relating to Assessing Students with Limited English Proficiency Raised in Education's Peer Review Reports: Validity: The state's item development and review procedures are inadequate to ensure that the assessments do not reflect unfair irrelevant constructs. For example, there is no empirical evidence that the item review process is successful in removing barriers due to overly complex language. Further, a statistical process to determine bias is evaluated only for gender and race/ethnicity; the state should consider using it to evaluate geographic and demographic diversity, such as students with limited English proficiency; Reliability: The state did not provide any reliability data for each reported subpopulation; Accommodations: The state provides a reasonable list of accommodations, but does not distinguish among those that are allowable for students with disabilities, and those which are allowable for students with limited English proficiency. The state may wish to provide separate lists of accommodations to support the selection of appropriate accommodations that are aligned with instructional approaches for individual students. Further, although it appears that the state has a system in place for monitoring the selection of accommodations for students with disabilities, it does not for students with limited English proficiency. Validity: The state should conduct a bias review, especially for the common items and the alternate assessments for students with limited English proficiency and students with disabilities; Reliability: The state does not routinely report subgroup reliability data; Accommodations: The state did not provide evidence to support the assertion that accommodations for students with limited English proficiency allow for valid inferences about these students' knowledge and skills. It does not appear that the state monitors availability of accommodations during test administration. The use of accommodations should be validated on the state's own student population. Source: GAO review of Education's peer review notes for 38 states. Note: This table includes examples from the categories used in Education's peer review process that determine if a state's assessments are valid, reliable, fair and accessible, and use appropriate accommodations. [End of table] Our group of experts indicated that states are generally not taking the appropriate set of comprehensive steps to create assessments that produce valid and reliable results for students with limited English proficiency and identified essential steps that should be taken. The group noted that no state has implemented an assessment program for students with limited English proficiency that is consistent with the Standards for Educational and Psychological Testing and other technical standards. Specifically, the group said that students with limited English proficiency are not defined consistently within and across states, which is a crucial first step to ensuring the reliability of test results. A reliable test should produce consistent results, so that students achieve similar scores if tested repeatedly. If the language proficiency levels of students with limited English proficiency are classified inconsistently, an assessment may produce results that appear inconsistent because of the variable classifications rather than actual differences in skill levels. One expert noted, for example, that some studies have shown that a student's language proficiency plays a small role in determining whether a student is classified as limited English proficient. Inconsistency in defining these students may be due to variation in how school districts apply state definitions. For example, according to a 2005 study on students with limited English proficiency in California, state board of education guidelines suggest that districts consider a student's performance on the state's English language proficiency assessment and on the state's language arts test, a teacher evaluation of the student's academic performance, and parental recommendations when determining if a student should or should not continue to be considered limited English proficient. However, the study noted that districts interpreted and applied these factors differently.[Footnote 16] Further, it appears that many state assessment programs do not conduct separate analyses for different groups of limited English proficient students. Our group of experts indicated that the reliability of a test may be different for heterogeneous groups of students with limited English proficiency, such as students who are literate in their native language and those who are not. Our group of experts also noted that states are not always explicit about whether an assessment is attempting to measure skills only (such as mathematics) or mathematics skills as expressed in English. According to the group, a fundamental issue affecting the validity of a test is the definition of what is being measured. Members of the group emphasized that approaches to ensure valid test results should vary based on which of these is being measured. For example, North Carolina officials stated that the state did not offer native language assessments because the state has explicitly chosen to measure student knowledge in English. The expert group emphasized that determining the validity and reliability of academic assessments for students with limited English proficiency is complicated and requires a comprehensive collection of evidence rather than a single analysis or review. As one expert noted, "you can't just do one thing and assume things are valid." In addition, the appropriate combination of analyses will vary from state to state, depending on the characteristics of the student population and the type of assessment. For example, because reliability of test results can vary based on a student's English proficiency status or a student's native language, states with more diverse groups of limited English proficient students may need to conduct additional analyses to ensure sufficient reliability. The group indicated that states are not universally using all the appropriate analyses to evaluate the validity and reliability of test results for students with limited English proficiency. Instead, our experts noted that states vary in terms of the particular techniques they use for this purpose, and in the extent to which they collect valid background data. Members indicated that some states may need assistance to conduct appropriate analyses that will offer useful information about the validity of their academic assessments for these students. Finally, our group of experts indicated that reducing language complexity is essential to developing valid assessments for these students, but expressed concern that some states and test developers do not have a strong understanding of universal design principles or how to use them to develop assessments from the beginning to eliminate language that is not relevant to measuring a student's knowledge of, for example, mathematics. Members believed that some states may need more information on how to implement these principles to develop assessments that produce valid results for students with limited English proficiency. Accommodations Can Increase Validity of Assessment Results, but Research on Appropriate Use Is Limited: The majority of states offered some accommodations to try to increase the validity and reliability of assessment results for students with limited English proficiency. These accommodations are intended to permit students with limited English proficiency to demonstrate their academic knowledge, despite their limited language ability. Our review of state Web sites found available documentation on accommodations for 42 states. The number of accommodations offered varied considerably among states. One state, for example, offered students with limited English proficiency the use of a bilingual dictionary and a side-by- side English-Spanish version of its grade 10 mathematics test. Another state listed over 40 acceptable accommodations, including clarifying test directions in English or the student's native language, offering extra time, and providing responses (written or oral) in the student's native language. Our review found that the most common accommodations offered by these states were allowing the use of a bilingual dictionary and reading test items aloud in English (see table 6). In addition, they offered other accommodations intended to create a less distracting environment for students, such as administering the assessment to the student in a small group or individually. Some states also gave students with limited English proficiency extra time to complete a test to account for their slower reading speed and information processing time in English. The 5 states we studied varied in how they established and offered accommodations to students. For example, Texas officials reported working with its limited English proficiency focus group to develop a list of allowable accommodations, which may be offered on a test when they are routinely used by students in their classrooms. In addition, each school district has a committee to select particular accommodations based on the needs of individual students. California officials told us the state provides guidance to districts on the appropriate use of accommodations. However, they said that districts might not provide approved accommodations because of high administrator turnover. Table 6: Most Frequently Cited Accommodations in 42 States: Accommodation: Bilingual dictionary; Number of states: 32. Accommodation: Reading items aloud in English; Number of states: 32. Accommodation: Small group administration; Number of states: 29. Accommodation: Extra time; Number of states: 27. Accommodation: Individual administration; Number of states: 27. Accommodation: Separate location; Number of states: 25. Accommodation: Extra breaks; Number of states: 25. Accommodation: Directions in student's native language; Number of states: 24. Source: GAO review of state documentation. [End of table] According to our expert group and our review of the relevant literature, research is lacking on what specific accommodations are appropriate for students with limited English proficiency, as well as their effectiveness in improving the validity of assessment results. A 2004 review of state policies found that few studies focus on accommodations intended to address the linguistic needs of students with limited English proficiency or on how accommodations affect the performance of students with limited English proficiency.[Footnote 17] In contrast, significantly more research has been conducted on accommodations for students with disabilities, much of it funded by Education. Because of this research disparity, our group of experts reported that some states offer accommodations to students with limited English proficiency based on those they offer to students with disabilities, without determining their appropriateness for individual students. Our experts noted the importance of considering individual student characteristics to ensure that an accommodation appropriately addresses the needs of the student. Other researchers have raised similar issues about the use of accommodations by states. Education's peer reviews of state academic assessments identified issues related to accommodations for students with limited English proficiency in all 38 states reviewed. For example, the reviewers noted that South Dakota does not clearly indicate whether students with limited English proficiency were provided accommodations that they do not regularly use in the classroom. If an accommodation is not used regularly in the classroom, it may not improve the validity of test results because the student may not be comfortable with a new procedure. In addition, they noted that South Dakota does not appear to be monitoring the use of accommodations and suggested that the state study accommodations to ensure that they are instructionally appropriate and that they improve the validity and reliability of the results. In Texas, the reviewers noted that the state needs to provide information regarding the quality and consistency of accommodations for students with limited English proficiency--specifically whether the state routinely monitors the use of accommodations for these students. In North Carolina, they noted a lack of evidence that the state has performed research on accommodations. Although conducting such research could provide useful information on the validity of accommodated tests, having each state individually study accommodations could be financially burdensome for them. While research on accommodations for this population would be useful, it does not have to be conducted directly by states to be applicable to a state's student population. Further, such research could involve short-term studies, rather than large-scale, longitudinal efforts. Native Language and Alternate Assessments May Improve the Validity of Results but Are Challenging to Implement: In our survey, 16 states reported that they offered statewide native language assessments in language arts or mathematics in some grades for certain students with limited English proficiency in the 2004-2005 school year. For example, New York translated its statewide mathematics assessments into Spanish, Chinese, Russian, Korean, and Haitian-Creole. In addition, 3 states were developing or planning to develop a native language assessment, and several states allowed school districts to translate state assessments or offer their own native language assessments. Our group of experts told us that this type of assessment is difficult and costly to develop. An assessment provided in a student's native language is intended to remove language barriers students face in demonstrating their content knowledge and thereby improve the validity of test results. Of the 16 states that offered statewide native language assessments, 4 were able to provide complete data on the number of students taking native language assessments. These data indicated that relatively few students took these assessments. Our group of experts and some state officials also described the challenges of developing native language assessments that produce valid results. Members of our expert group and other experts told us that native language assessments are generally an effective accommodation only for students in specific circumstances, such as students who are instructed in their native language or are literate in their native language. In addition, our experts emphasized that developing valid native language assessments is challenging, time-consuming, and expensive. Development of a valid native language assessment involves more than a simple translation of the original test; in most situations, a process of test development and validation similar to that of the nontranslated test is recommended to ensure the validity of the test. In addition, the administration of native language assessments may not be practicable, for example, when only a small percentage of limited English students in the state speak a particular language or when a state's student population has many languages. Thirteen states offered statewide alternate assessments (such as reviewing a student's classroom work portfolio) in 2005 for certain students with limited English proficiency, based on our review of accountability plans for all states and the District of Columbia as of March 2006. We also found that 4 additional states allowed school districts to offer alternate assessments, while 7 states and the District of Columbia planned to offer alternate assessments. An official in Wisconsin told us that the state administers an alternate assessment because developing a native language assessment for its relatively small Spanish-speaking population would be impractical and the state does not have bilingual programs in the second most common language, Hmong (a language that is native to Southeast Asia). However, our group of experts noted that alternate assessments are difficult and expensive to develop, and may not be feasible because of the amount of time required for such an assessment. Members of the group also expressed concern about the extent to which these assessments are objective and comparable and can be aggregated with regular assessments. See figure 4 for information on which states offered native language or alternate assessments for students with limited English proficiency. Figure 4: Use of Native Language and Alternate Assessments for Students with Limited English Proficiency: [See PDF for image] Source: GAO survey and state accountability workbooks. [End of figure] Most States Implemented New English Language Proficiency Assessments but Faced Challenges Establishing Their Validity and Reliability: With respect to English language proficiency assessments, many states implemented new tests to address NCLBA requirements, and are working to align them with newly required state English language proficiency standards. State and consortia officials reported that states are using assessments or test items developed by state consortia, customized assessments developed by testing companies, state-developed assessments, and off-the-shelf assessments. While a few states already had the required English language proficiency assessments in place, many states are implementing them for the first time in spring 2006; as a result, evidence on their validity and reliability may not be fully developed. States Are Working with Consortia and Test Developers and Individually to Develop New English Language Proficiency Assessments: Many states implemented new English language proficiency assessments for the 2005-2006 school year to meet Education's requirement for states to administer English language proficiency tests that meet NCLBA requirements by the spring of 2006.[Footnote 18] These assessments must allow states to track student progress in learning English; in addition, Education requires that these assessments be aligned to a state's English language proficiency standards. According to Education and test development officials, prior to NCLBA, most states used off- the-shelf English language proficiency assessments to determine the placement of students in language instruction programs, but these assessments did not have to be aligned with standards. Education officials said that because many states did not have tests that met NCLBA requirements, the agency funded four state consortia to develop new assessments that were to be aligned with state standards and measure student progress. Officials in some states told us they have chosen to use these consortium-developed tests, while officials in other states reported developing their own tests or continuing to use off-the-shelf tests. Some states had only recently determined what test they are going to administer this year, while others may administer a new test in the 2006-2007 school year. Education officials noted that states' decisions on these tests have been in flux during this transition year. In the 2005-2006 school year, 22 states used assessments or test items developed by one of four state consortia, making this the most common approach taken by states to develop new English language proficiency assessments. Each of the four consortia varied somewhat in its development approach.[Footnote 19] For example, officials in two consortia reported that they examined all their member states' English language proficiency standards and reached consensus on core standards for use on the English language proficiency assessments. They also planned to continue working with member states in implementing their assessments. For example, one consortium plans to provide ongoing professional development to help educators understand the consortium's standards. In contrast, officials in the other two consortia reported that the consortia disbanded after developing their assessments. One state official told us that the state hired a contractor to customize the consortium-developed assessment to more closely align with state standards. In addition, officials in other states, such as New Mexico, told us they are using a combination of consortium-developed test items, along with items developed by another test developer. Fifteen states participated in one of the consortia, but officials in these states told us they chose not to use the assessments developed by the consortia in the 2005-2006 school year for a variety of reasons, including lack of alignment with state standards, the length of the assessment, and the cost of implementation. For example, Kentucky chose not to use the consortium assessment because of cost effectiveness concerns and lack of alignment with state standards. Another state decided not to use the consortium-developed assessment, as officials were concerned about its cumbersome nature and associated cost. Officials in some states told us they plan to use consortium-developed assessments in the future. For example, Florida officials reported that the state will use a consortium assessment in the 2006-2007 school year. Appendix V shows the states that participated in the consortia and which used consortia-developed assessments in the 2005-2006 school year. Officials in states that did not use consortia assessments told us that they used other approaches to develop their English language proficiency assessments. Eight states worked with test developers to augment off-the-shelf English language proficiency assessments to incorporate state standards. For example, Mississippi, South Dakota, and Wyoming are using versions of an English language proficiency assessment that has been augmented to align to their respective state standards. Officials in 14 states indicated that they are administering off-the-shelf assessments. These officials indicated varying degrees of alignment between the off-the-shelf tests being used and their state's English language proficiency standards; in 11 of these states, the assessment has not been fully aligned with state standards.[Footnote 20] Seven states, including Texas, Minnesota, and Kansas, created their own English language proficiency assessments. Officials in these states said they typically worked with a test developer or research organization to create the assessments. See figure 5 and appendix VI for more detailed information on the English language proficiency assessments used by each state. Figure 5: Type of English Language Proficiency Assessment Administered in 2005-2006 School Year: [See PDF for image] Source: Interviews with consortia and state officials. [End of figure] Some officials in our 5 study states and 28 additional states we contacted to determine what English language proficiency assessment they planned to use in 2006 pointed to some challenges involving their English language proficiency assessments. Some of these state officials expressed concerns about using both their old and new English language proficiency assessments to measure student progress in learning English. NCLBA required states to begin tracking student progress in the 2002-2003 school year, before most states had implemented their new English language proficiency assessments. In May 2006, Education officials told us that states must rely on baseline results from their old tests and determine how results from their old tests relate to results from their new tests in order to track student progress since 2003, as required by NCLBA. They noted that states may change their English language proficiency goals based on results from their new assessments, but they cannot change the initial baseline established with their old test. In its technical comments on this report, Education noted that it allows states to make such determinations in a variety of ways, as long as annual progress is reported. Officials in some states want to rely solely on data from their new tests to track student progress. They stated that, unlike their old tests, their new tests provide more accurate data on student progress because they are aligned to their English language proficiency standards and were designed to measure student progress. Officials from other states questioned the usefulness of conducting studies to determine the relationship between their old and new tests, especially in states that had previously used multiple English language proficiency assessments. Officials in a few of our study states also expressed concern about the appropriateness of NCLBA's requirement to assess students with limited English proficiency in kindergarten and the first and second grades. For example, Texas officials told us traditional tests do not produce good test results for students this young in part because of their limited attention spans. In addition, officials in Texas and North Carolina noted that English proficient students in these grades are not required to be assessed in the same way. Many States Are Still in the Process of Establishing the Validity and Reliability of English Language Proficiency Assessments: Officials in our study states and test developers we interviewed reported that they commonly apply generally accepted test development procedures in the development of English language proficiency assessments, but some are still in the process of documenting the validity and reliability of these assessments. For example, some evidence needed to confirm the validity and reliability of the test can be obtained only after the assessment has been fully administered. One consortium contracted with a research organization to assess the validity and reliability testing of its English language proficiency assessment. According to a consortium official, the research organization performed all of the standard steps that are taken to ensure high-quality assessments. These included piloting and field testing the assessment and conducting statistical modeling. An official from another consortium said that its test vendor is conducting basic psychometric research and analyzing field test data for evidence of reliability. California officials noted that the process for developing and ensuring the validity and reliability of its English language proficiency assessment is similar to that used for its state academic assessments. Although states have taken steps toward determining validity, documenting the validity and reliability of a new assessment is an ongoing process. A 2005 review of the documentation of 17 English language proficiency assessments used by 33 states in the 2005-2006 school year found that the evidence presented on validity and reliability was generally insufficient.[Footnote 21] The report, which was funded by Education, reviewed documentation for consortium- developed assessments, off-the-shelf assessments, and custom-developed assessments for evidence of validity, reliability, and freedom from test bias, among other things. It found that the technical adequacy of English language proficiency assessments is undeveloped compared to the adequacy of assessments for general education. The study noted that none of the assessments contained "sufficient technical evidence to support the high-stakes accountability information and conclusions of student readiness they are meant to provide." In addition, many states are in the process of aligning these assessments to state English language proficiency standards, which in turn must be aligned to state content standards. These steps are needed to comply with NCLBA requirements. Alignment, which refers to the degree to which an assessment's items measure the content they are intended to measure, is critical in assuring the validity of an assessment. Officials in some states[Footnote 22] have expressed uncertainty about how to align their English language proficiency test with their standards for academic subjects, such as mathematics and science.[Footnote 23] Officials in 2 states told us that their English language proficiency assessments are aligned to state language arts standards but are not aligned to state mathematics standards, meaning that the assessment may not measure the language needed to succeed in a mathematics class. Findings from Education's Title III monitoring reviews of 13 states indicated that 8 states had not yet fully completed alignment; of these, 5 had not yet linked their English language proficiency and academic content standards, while 5 had not yet aligned their English language proficiency assessments with their English language proficiency standards[Footnote 24]. Education Has Provided Assistance, but States Reported Need for Additional Guidance and Flexibility: Education has offered states a variety of technical assistance to help them appropriately assess students with limited English proficiency, such as providing training and expert reviews of their assessment systems, as well as flexibility in assessing these students. However, Education has issued little written guidance on how states are expected to assess and track the English proficiency of these students, leaving state officials unclear about Education's expectations. To support states' efforts to incorporate these students into their accountability systems, Education has offered states some flexibilities in how they track progress goals for these students. However, many of the state and district officials we interviewed told us that the current flexibilities do not fully account for some characteristics of certain students in this student group, such as their lack of previous schooling. These officials indicated that additional flexibility is needed to ensure that the federal progress measures accurately track the academic progress of these students. Education Has Provided a Variety of Support on Assessment Issues but Little Written Guidance on Assessing Students with Limited English Proficiency: Education offers support in a variety of ways to help states meet NCLBA's requirements for assessing students with limited English proficiency for both their language proficiency and their academic knowledge. Some of these efforts focus specifically on students with limited English proficiency, while others, such as the Title I monitoring visits, focus on all student groups and on broader compliance issues but review some assessment issues related to students with limited English proficiency as part of their broader purposes. The agency's primary technical assistance efforts have included the following: * Title I peer reviews of states' academic standards and assessment systems: Education is currently conducting peer reviews of the academic assessments that states use in measuring adequate yearly progress. During these reviews, three independent experts review evidence provided by the state about the validity and reliability of these assessments (including whether the results are valid and reliable for students with limited English proficiency) and make recommendations to Education about whether the state's assessment system is technically sufficient and meets all legal requirements. Education shares information from the peer review to help states address issues identified during the review. Education has imposed a deadline requiring that states receive peer review approval by June 30, 2006, but only 10 states have had their assessment systems fully approved by Education as of that date.[Footnote 25] * Title III monitoring visits: Education began conducting site visits to review state compliance with Title III requirements in 2005 and has visited 15 states. Education officials reported that they plan to visit 11 more states in 2006. As part of these visits, the agency reviews the state's progress in developing English language proficiency assessments that meet NCLBA requirements. * Comprehensive centers: Education has contracted with 16 regional comprehensive centers to build state capacity to help districts that are not meeting their adequate yearly progress goals. The grants for these centers were awarded in September 2005, and the centers provide a broad range of assistance, focusing on the specific needs of individual states. At least 3 of these centers plan to assist individual states in developing appropriate goals for student progress in learning English. In 2005, Education also funded an assessment and accountability comprehensive center, which provides technical assistance to the regional comprehensive centers on issues related to the assessment of students, including those with limited English proficiency. * Ongoing technical assistance for English language proficiency assessments: Education has provided information and ongoing technical assistance to states using a variety of tools and has focused specifically on the development of the English language proficiency standards and assessments required by NCLBA. These include: * a semiannual review of reports states submit to Education and phone calls to state officials focused on state progress in developing their English language proficiency assessments; * on-site technical assistance to states regarding their English language proficiency assessments; * an annual conference focused on students with limited English proficiency that includes sessions on assessment issues, such as aligning English language proficiency and academic content standards; * videoconference training sessions for state officials on developing English language proficiency assessments; * providing guidance on issues related to students with limited English proficiency on its Web site; * distributing information through an electronic bulletin board and a weekly electronic newsletter focused on students with limited English proficiency; * disseminating information through the National Clearinghouse for English Language Acquisition and Language Instruction Educational Programs; * semiannual meetings and training sessions with state Title III directors; and: * responding to questions from individual states as needed. * Enhanced Assessment Grants: Since 2003, Education has awarded these grants, authorized by NCLBA, to support state activities designed to improve the validity and reliability of state assessments. According to an Education official, most of the grants up to now have funded the English language proficiency consortia, although some grants have been used to conduct research on accommodations. For grants to be awarded in 2006, Education will give preference to projects involving accommodations and alternate assessments intended to increase the validity of assessments for students with limited English proficiency and students with disabilities. * Title I monitoring visits: As part of its monitoring visits to review state compliance with Title I requirements, Education reviews some aspects of the academic assessments administered by states, but in less detail than during its peer reviews. During these visits, for example, states may receive some feedback on how the state administers academic assessments to students with limited English proficiency and the appropriateness of accommodations offered to these students. Education staff also reported that they respond to questions about Title I requirements from individual states as needed. While providing states with a broad range of technical assistance and guidance through informal channels, Education has issued little written guidance on developing English language proficiency assessments that meet NCLBA's requirements and on tracking the progress of students in acquiring English. Education issued some limited nonregulatory guidance on NCLBA's basic requirements for English language proficiency standards and assessments in February 2003. However, officials in about one-third of the 33 states we visited or directly contacted expressed uncertainty about implementing these requirements. They told us that they would like more specific guidance from Education to help them develop tests that meet NCLBA requirements, generally focusing on two issues. First, some officials said they were unsure about how to align English language proficiency standards with content standards for language arts, mathematics, and science, as required by NCLBA. An official in 1 state said the state needed specific guidance on what Education wants from these assessments, such as how to integrate content vocabulary on the English language proficiency assessment without creating an excessively long test. In another state, officials explained that the state was developing its English language proficiency test by using an off-the-shelf test and incorporating additional items to align the test with the state's English language proficiency and academic standards. However, the state discovered that it had not correctly augmented the test and will consequently have to revise the test. Officials in this state noted that they have had to develop this test without a lot of guidance from Education. Second, some officials reported that they did not know how to use the different scores from their old and new English language proficiency assessments to track student progress. For example, an official in 1 state said that she would like guidance from Education on how to measure student progress in English language proficiency using different tests over time. Another official was unsure if Education required a formal study to correlate the results from their old and new English language proficiency assessments, noting that more specific guidance would help them better understand Education's requirements. Without guidance and specific examples on both of these issues, some of these officials were concerned that they will spend time and resources developing an assessment that may not meet Education's requirements. Education officials told us that they are currently developing additional nonregulatory guidance on these issues, but it has not been finalized. They also pointed out that they have provided extensive technical assistance on developing English language proficiency standards and assessments, and have clearly explained the requirements to state officials at different meetings on multiple occasions. An Education official acknowledged that states were looking for more guidance on the degree of alignment required between their English language proficiency assessments and standards, noting that Education is still considering the issue. She stated that the issue would be addressed in the guidance it plans to issue in the future. With respect to academic content assessments, our group of experts reported that some states could use more assistance in creating valid academic assessments for students with limited English proficiency. While 4 of the 5 states we studied in depth had significant experience in, and multiple staff devoted to, developing language arts and mathematics assessments, some members of our expert group pointed out that the assessment departments in other states have limited resources and expertise, as well as high turnover. As a result, these states need help to conduct appropriate analyses that will offer useful information about the validity and reliability of their academic assessments for students with limited English proficiency. An Education official told us that the agency recently began offering technical assistance to states that need help addressing issues raised during their peer reviews. Our group of experts suggested several areas where states could benefit from additional assistance and guidance in developing academic assessments for students with limited English proficiency. Several members noted the lack of good research on what kinds of accommodations can help mitigate language barriers for students with limited English proficiency. Several experts also believed that some states need more information on how to implement universal design principles to develop assessments that produce valid results for students with limited English proficiency. In addition, some group members pointed out that developing equivalent assessments in other languages (that is, assessments that measure the same thing and are of equivalent difficulty) is challenging and that states need more information about how to develop such assessments, as well as examples. Education Has Offered Different Accountability Options for Students with Limited English Proficiency, but State Officials Reported Additional Flexibility Is Needed: Education has offered states several flexibilities in tracking academic progress goals for students with limited English proficiency to support their efforts to develop appropriate accountability systems for these students. In a February 2004 notice, Education recognized the existence of language barriers that hinder the assessment of students who have been in the country for a short time and provided some testing flexibility for these students. Specifically, Education does not require students with limited English proficiency to participate in a state's language arts assessment during their first year in U.S. schools. In addition, while these students must take a state's mathematics assessment during their first year in U.S. schools, a state may exclude their scores in determining whether it met its progress goals.[Footnote 26] Education offered additional flexibility in its February 2004 notice, recognizing that limited English proficiency is a more transient quality than having a disability or being of a particular race. Unlike the other NCLBA student groups, students who achieve English proficiency leave the group at the point when they are more prepared to demonstrate their academic knowledge in English, while new students with lower English proficiency are constantly entering the group (see fig. 6). Given the group's continually changing composition, meeting progress goals may be more difficult than doing so for other student groups, especially in districts serving large numbers of students with limited English proficiency. To compensate for this, Education allowed states to include, for up to 2 years, the scores of students who were formerly classified as limited English proficient when determining whether a state met its progress goals for students with limited English proficiency. In addition, Education has approved requests from several states to permit students who have been redesignated as English proficient to remain in the group of students with limited English proficiency until they have achieved the proficient level on the state's language arts assessment for 1 or more years. Figure 6: Movement of Students In and Out of Limited English Proficient Student Group and Other Student Groups: [See PDF for image] Source: GAO analysis and Art Explosion images. [End of figure] Several state and local officials in our study states told us that additional flexibility would be helpful to ensure that the annual progress measures provide meaningful information about the performance of students with limited English proficiency. Officials in 4 of the states we studied suggested that certain students with limited English proficiency should be exempt for longer periods from taking academic content assessments or that their test results should be excluded from a state's annual progress determination for a longer period than is currently allowed. Several officials voiced concern that some of these students have such poor English skills or so little previous school experience that the assessment results do not provide any meaningful information. Instead, some of these officials stated that students with limited English proficiency should not be included in academic assessments until they demonstrate appropriate English skills on the state's English language proficiency assessment. However, the National Council of La Raza, an Hispanic advocacy organization, has voiced concern that excluding too many students with limited English proficiency from a state's annual progress measures will allow some states and districts to overlook the needs of these students. Education officials reported that they are developing a regulation with regard to how test scores for this student group are included in a state's annual progress measures, but it has not yet been finalized. With respect to including the scores of students previously classified as limited English proficient in a state's progress measures for this group for up to 2 years, officials in 2 of our 5 study states, as well as one member of our expert group, thought it would be more appropriate for these students to be counted in the limited English proficient group throughout their school careers--but only for accountability purposes. They pointed out that by keeping students formerly classified as limited English proficient in the group, districts that work well with these students would see increases in the percentage who score at the proficient level in language arts and mathematics. An Education official explained that the agency does not want to label these students as limited English proficient any longer than necessary and considered including test results for these students for 2 years after they have achieved English proficiency to be the right balance. Education officials also noted that including all students who were formerly limited English proficient would inflate the achievement measures for the student group. District officials in 4 of the states we studied argued that tracking the progress of individual students in this group is a better measure of how well these students are progressing academically. Officials in one district pointed to a high school with a large percentage of students with limited English proficiency that had made tremendous progress with these students, doubling the percentage of students achieving academic proficiency. The school missed the annual progress target for this group by a few percentage points, but school officials said that the school would be considered successful if it was measured by how much individual students had improved in their test scores. A district official in another state explained that many students with limited English proficiency initially have very low test scores, but demonstrate tremendous improvement in these scores over time. In response to educators and policymakers who believe such an approach should be used for all students, Education initiated a pilot project in November 2005, allowing a limited number of states to incorporate measures of student progress over time in determining whether districts and schools met their annual progress goals. Even using this approach, however, states must still establish annual goals that lead to all students achieving proficient scores by 2014.[Footnote 27] Conclusions: NCLBA has focused attention on the academic performance of all students, especially those who have historically not performed as well as the general student population, such as students with limited English proficiency. NCLBA requires states to include these students in their language arts and mathematics assessments and to assess them in a valid and reliable manner, and states are in various stages of doing so. Although Education has provided some technical assistance to states, our group of experts and others have noted the complexity of developing academic assessments for these students and have raised concerns about the technical expertise of states to ensure the validity and reliability of assessment results. Using assessment results that are not a good measure of student knowledge is likely to lead to poor measures of state and district progress, thereby undermining NCLBA's purpose to hold schools accountable for student progress. Further, although most states offered these students accommodations, research on their appropriateness is limited. National research on accommodations has informed states' practices in assessing students with disabilities. Without similar research efforts, accommodations offered to students with limited English proficiency may not improve the validity of their test results. While Education has provided some support and training to states, officials in a number of states are still uncertain about how to comply with some of the more technical requirements of the new English language proficiency assessments required by NCLBA. State officials reported that they need more guidance from Education to develop these assessments. States have had to develop many new assessments under NCLBA for both English language proficiency and academic content, and some states may lack the technical expertise to develop assessments that produce valid results for students with limited English proficiency. Without more specific guidance outlining Education's requirements, states may spend time developing English language proficiency assessments that do not adequately track student progress in learning English or otherwise meet NCLBA's requirements. Including students with limited English proficiency in NCLBA's accountability framework presents unique challenges. For example, students who have little formal schooling may make significant progress in learning academic skills, but may not achieve proficiency on state academic assessments for several years. The movement of students into and out of the group also makes it more difficult for the group to meet state progress goals, even when these students are making academic progress. Education has addressed some of the unique characteristics of this student group and provided some flexibility in how states and districts are held accountable for the progress of these students. However, these current flexibilities may not fully account for the characteristics of certain students with limited English proficiency, such as those who have little previous formal schooling. Recommendations for Executive Action: We recommend that the Secretary of Education: 1. Support additional research on appropriate accommodations for students with limited English proficiency and disseminate information on research-based accommodations to states. 2. Determine what additional technical assistance states need with respect to assessing the academic knowledge of students with limited English proficiency and to improve the validity and reliability of their assessment results (such as consultations with assessment experts and examples of assessments targeted to these students) and provide such additional assistance. 3. Publish additional guidance with more specific information on the requirements for assessing English language proficiency and tracking the progress of students with limited English proficiency in learning English. 4. Explore ways to provide additional flexibilities to states in terms of holding states accountable for students with limited English proficiency. For example, among the flexibilities that could be considered are: * allowing states to include the assessment scores for all students formerly considered to have limited English proficiency in a state's annual progress results for the group of students with limited English proficiency, * extending the period during which the assessment scores for some or all students with limited English proficiency would not be included in a state's annual progress results, and: * adjusting how states account for recent immigrants with little formal schooling in their annual progress results. Agency Comments: We provided a draft of this report to Education for review and comment. The agency provided comments, which are reproduced in appendix VII. Education also provided technical clarifications, which we incorporated when appropriate. Education agreed with our first three recommendations. The department noted that it has conducted some research on the effectiveness of accommodations and is currently working with its National Research and Development Center for Assessment and Accountability to synthesize the existing research literature on the assessment of students with limited English proficiency. Education also explained that it has begun the process of identifying the additional technical assistance needs of states with respect to academic assessments; specifically, it will have its Assessment and Accountability Comprehensive Center conduct a needs assessment this fall to determine specific areas in which states need assistance and will provide technical assistance to address those areas. In addition, the department stated that it is exploring ways to help states assess English language proficiency. Education did not explicitly agree or disagree with our fourth recommendation. Instead, the agency commented that it has explored and already provided various types of flexibility regarding the inclusion of students with limited English proficiency in accountability systems. Further, Education noted that it is in the process of completing a regulation on flexibility for these students. However, the department also emphasized that all students with limited English proficiency must be included in school accountability systems to improve both instruction and achievement outcomes. Through our recommendation, we encourage the department to continue its efforts. We are sending copies of this report to the Secretary of Education, relevant congressional committees, and other interested parties. We will make copies available to others upon request. In addition, the report will be available at no charge on GAO's Web site at [Hyperlink, http://www.gao.gov]. If you or your staff have any questions or wish to discuss this report further, please contact me at (202) 512-7215 or at shaulm@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. Other contacts and major contributors are listed in appendix VIII. Signed by: Marnie S. Shaul: Director, Education, Workforce, and Income Security Issues: [End of section] Appendix I: GAO's Group of Experts on Assessing the Academic Knowledge of Students with Limited English Proficiency: On January 20, 2006, GAO, with the assistance of the National Academy of Sciences, convened a group of experts in Davis, California, to discuss issues related to assessing the academic knowledge of students with limited English proficiency. Specifically, we asked the group to discuss the following questions: * To meet the requirements of the No Child Left Behind Act (NCLBA), what steps should states take to ensure the validity and reliability of language arts and mathematics assessments for students with limited English proficiency? * What steps should states take to ensure that students with limited English proficiency receive appropriate accommodations on language arts and mathematics assessments? * Given NCLBA's accountability framework, what is the most appropriate way to hold schools and districts accountable for the performance of students with limited English proficiency? * How can the U.S. Department of Education assist states in their efforts to meet NCLBA's assessment and accountability requirements for students with limited English proficiency? Group members who were selected had significant technical and research expertise in assessments issues. Some members had technical expertise on general assessment issues, while others had specifically conducted assessment research focused on students with limited English proficiency. The members of our expert group are listed below: Dr. Jamal Abedi: University of California, Davis: Dr. Stephen Dunbar: The University of Iowa: Dr. Richard Durán: University of California, Santa Barbara: Dr. Steven Ferrara: American Institutes for Research: Dr. Patricia Gándara: University of California, Davis: Dr. Edward Haertel: Stanford University: Dr. Rebecca Kopriva: University of Maryland: Dr. Stanley Rabinowitz: WestEd: Dr. Charlene Rivera: The George Washington University: Dr. Rebecca Zwick: University of California, Santa Barbara: [End of section] Appendix II: Determining Adequate Yearly Progress for Student Groups: NCLBA requires states to report adequate yearly progress (AYP) results at the state level for each of the required student groups, including students with limited English proficiency. The law also requires Education, starting in the 2004-2005 school year, to make an annual determination about whether states have made adequate yearly progress for each student group.[Footnote 28] Education has issued some general regulations regarding state-level adequate yearly progress. However, Education has not yet collected any such state-level adequate yearly progress results and has not issued any guidance on how states should determine whether a student group has made adequate yearly progress. As a result, some states have not yet made adequate yearly progress determinations for student groups at the state level. In order for a student group, such as students with limited English proficiency, to make adequate yearly progress, it must make a number of different goals. Specifically: * At least 95 percent of students in the group must take the state's language arts and mathematics assessments, and: * The student group must meet the progress goals established by the state for both language arts and mathematics proficiency or: * The percentage of students who did not achieve proficient scores must have decreased by at least 10 percent from the previous year, and the student group must also meet the progress goals established by the state for its other academic indicator (graduation rate for high schools and usually attendance rate for other schools). Figure 7 illustrates the basic decision process for determining adequate yearly progress for a student group. Figure 7: Process for Determining Adequate Yearly Progress for a Student Group: [See PDF for image] Source: GAO analysis. [End of figure] Because states have different assessment systems, they use different methods for determining adequate yearly progress. A state can have an assessment system that allows it to create the same progress goal for mathematics and language arts for all grades, despite using different tests in each grade. In this case, the state could review data for all students in a student group across the state to determine if the group met its annual progress goals. A state can also establish different progress goals for different grades or groups of grades, depending on the particular test being used. In this case, according to an Education official, a state would have to meet all the proficiency and participation goals for all the different grades or groups of grades in order to make adequate yearly progress. [End of section] Appendix III: Percentage of Districts Making AYP Goals for Mathematics for Students with Limited English Proficiency: [See PDF for Image] Source: GAO analysis of district report cards and district data provided by state officials. Notes: Data are for school year 2003-2004. We requested district-level achievement data from 20 states, and 18 states responded to our request. When districts reported proficiency data for different grades or groups of grades, we determined that the percentage of students with limited English proficiency met a state's mathematics progress goal if the student group met the goal for all grades reported. Results from charter schools are included when a charter school is its own school district or part of a larger school district. Hawaii only has one school district. [End of figure] [End of section] Appendix IV: Proficiency Scores on Mathematics Tests for All Students and Students with Limited English Proficiency: [See PDF for Image] Source: Consolidated performance reports for the 2003-2004 school year. Notes: New York and Tennessee did not provide assessment data in their state consolidated performance reports for the 2003-2004 school year. The results for Arkansas do not include those students with limited English proficiency who were considered proficient based on the state's portfolio assessment. The total student population includes students with limited English proficiency. [A] Most states reported assessment data for students in the fourth grade. States marked with this superscript reported on some other grade at the elementary school level, usually either third grade or fifth grade. [End of figure] [End of section] Appendix V: Enhanced Assessment Consortia Participation: Assessment; World-Class Instructional Design and Assessment (WIDA) Consortium: Assessing Comprehension and Communication in English State- to-State for English Language Learners (ACCESS for ELLs); State Collaborative on Assessment and Student Standards (SCASS) Consortium: English Language Development Assessment (ELDA); Mountain West Assessment Consortium (MWAC): MWAC; Pennsylvania Enhanced Assessment Grant (PA EAG): Comprehensive English Language Learning Assessment (CELLA). Consortia states: Using assessment in 2005-2006 school year; World-Class Instructional Design and Assessment (WIDA) Consortium: Alabama Delaware District of Columbia Georgia Illinois Maine New Hampshire New Jersey Oklahoma Rhode Island Vermont Wisconsin; State Collaborative on Assessment and Student Standards (SCASS) Consortium: Iowa Louisiana Nebraska Ohio South Carolina West Virginia; Mountain West Assessment Consortium (MWAC): Idaho[A] New Mexico[A] Michigan [A, B]; Pennsylvania Enhanced Assessment Grant (PA EAG): Tennessee. Consortia states: Not using assessment in 2005-2006 school year; World-Class Instructional Design and Assessment (WIDA) Consortium: none; State Collaborative on Assessment and Student Standards (SCASS) Consortium: Kentucky Nevada North Carolina Texas; Mountain West Assessment Consortium (MWAC): Alaska Colorado Montana Nevada North Dakota Utah Wyoming; Pennsylvania Enhanced Assessment Grant (PA EAG): Florida Maryland Michigan[B] Pennsylvania. Source: Interviews with consortia and state officials. Note: This table reflects states that participated in the consortia prior to or during the 2005-2006 school year. Some states are no longer consortia members. [A] Using test items from consortium-developed assessment. [B] Participated in more than one consortium. [End of table] [End of section] Appendix VI: English Language Proficiency Assessments Used in the 2005- 2006 School Year, by State: State: Alabama; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Alaska; English Language Proficiency Assessment: IDEA Proficiency Test. State: Arizona; English Language Proficiency Assessment: Stanford English Language Proficiency Test. State: Arkansas; English Language Proficiency Assessment: MAC II (Maculaitis Assessment of Competencies) Test of English Language Proficiency. State: California; English Language Proficiency Assessment: California English Language Development Test. State: Colorado; English Language Proficiency Assessment: Colorado English Language Assessment. State: Connecticut; English Language Proficiency Assessment: LAS (Language Assessment System) Links. State: Delaware; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: District of Columbia; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Florida; English Language Proficiency Assessment: Various off- the-shelf tests[A]. State: Georgia; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Hawaii; English Language Proficiency Assessment: LAS (Language Assessment System) Links. State: Idaho; English Language Proficiency Assessment: Mountain West Assessment Consortium test items. State: Illinois; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Indiana; English Language Proficiency Assessment: LAS (Language Assessment System) Links. State: Iowa; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: Kansas; English Language Proficiency Assessment: Kansas English Language Proficiency Assessment. State: Kentucky; English Language Proficiency Assessment: 2004 IDEA Proficiency Test or Language Assessment Scales (LAS)[A]. State: Louisiana; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: Maine; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Maryland; English Language Proficiency Assessment: LAS (Language Assessment System) Links. State: Massachusetts; English Language Proficiency Assessment: Massachusetts English Proficiency Assessment. State: Michigan; English Language Proficiency Assessment: English Language Proficiency Assessment[B] (includes Mountain West Consortium test items). State: Minnesota; English Language Proficiency Assessment: Test of Emerging Academic English, Minnesota Student Oral Language Observation Matrix, and checklist for reading and writing for K-2 students. State: Mississippi; English Language Proficiency Assessment: Stanford English Language Proficiency Test. State: Missouri; English Language Proficiency Assessment: MAC II (Maculaitis Assessment of Competencies) Test of English Language Proficiency. State: Montana; English Language Proficiency Assessment: Iowa Test of Basic Skills, Woodcock-Muñoz Language Survey (English), or other state- approved test[A]. State: Nebraska; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: Nevada; English Language Proficiency Assessment: LAS (Language Assessment System) Links. State: New Hampshire; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: New Jersey; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: New Mexico; English Language Proficiency Assessment: New Mexico English Language Proficiency Assessment (includes Mountain West Consortium test items). State: New York; English Language Proficiency Assessment: New York State English as a Second Language Achievement Test. State: North Carolina; English Language Proficiency Assessment: IDEA Proficiency Test. State: North Dakota; English Language Proficiency Assessment: 2004 IDEA Proficiency Test, Woodcock-Muñoz Language Survey (English), and Language Assessment Scales (LAS)[A]. State: Ohio; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: Oklahoma; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Oregon; English Language Proficiency Assessment: English Language Proficiency Assessment[B]. State: Pennsylvania; English Language Proficiency Assessment: Stanford English Language Proficiency Test. State: Rhode Island; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: South Carolina; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: South Dakota; English Language Proficiency Assessment: Dakota English Language Proficiency assessment. State: Tennessee; English Language Proficiency Assessment: Comprehensive English Language Learning Assessment (PA EAG). State: Texas; English Language Proficiency Assessment: Texas English Language Proficiency Assessment System; consists of Reading Proficiency Tests in English and Texas Observation Protocols. State: Utah; English Language Proficiency Assessment: 2004 IDEA Proficiency Test. State: Vermont; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Virginia; English Language Proficiency Assessment: Stanford English Language Proficiency Test. State: Washington; English Language Proficiency Assessment: Washington Language Proficiency Test. State: West Virginia; English Language Proficiency Assessment: English Language Development Assessment (SCASS). State: Wisconsin; English Language Proficiency Assessment: Assessing Comprehension and Communication in English State-to-State for English Language Learners (WIDA). State: Wyoming; English Language Proficiency Assessment: Wyoming English Language Learner Assessment. Source: Interviews with consortia and state officials. [A] State allows school districts to individually choose tests. [B] Assessments are not the same; Oregon's is a state developed assessment, while Michigan's is a combination of an augmented version of the Stanford English Language Proficiency Test and test items from the Mountain West Assessment Consortium. [End of table] [End of section] Appendix VII: Comments from the Department of Education: United States Department Of Education: Office Of Elementary And Secondary Education: The Assistant Secretary: July 20, 2006: Ms. Marnie S. Shaul: Director, Education, Workforce and Income Security Issues: Government Accountability Office: 441 G Street, NW: Washington, DC 20548: Dear Ms. Shaul: I am writing in response to your request for comments on the Government Accountability Office (GAO) draft report (GAO-06-815), dated August 2006, and entitled "No Child Left Behind Act: Assistance from Education Could Help States Better Measure Progress of Students with Limited English Proficiency." I appreciate the opportunity to comment on the draft report and provide insight on our activities to support states in serving limited English proficient (LEP) students. The following are responses to the recommendations in the report to assist states better measure the progress of students with limited English proficiency: Recommendation 1. Support additional research on appropriate accommodations for students with limited English proficiency and disseminate information on research-based accommodations to states. The Department agrees with acquiring additional information on better serving LEP students. The Department's Institute of Education Sciences (IES), through its National Assessment of Educational Progress (NAEP), has sponsored studies of the effectiveness of various accommodations for LEP students, such as simplified English and use of translations, glossaries, and bilingual dictionaries, and we have applied the findings to NAEP. The Department has also shared the results of the studies at various meetings and national conferences. NAEP has also conducted studies on the impact on validity of test results of providing accommodations to LEP students, and on the impact of excluding such students from the assessment. In addition, IES is working with its National Research and Development Center for Assessment and Accountability --known as the Center for Research on Evaluation, Standards, and Student Testing (CRESST) --to synthesize the existing research literature on the assessment of LEP students. The synthesis addresses both accessibility and validity. After disseminating the information to states, CRESST will work with states to help them review and improve their assessment instruments and procedures, especially with regard to LEP students. Recommendation 2. Determine what additional technical assistance states need with respect to assessing the academic knowledge of students with limited English proficiency and to improve the validity and reliability of the assessment results (such as consultations with assessment experts and examples of assessments targeted to these students) and provide such additional assistance. The Department has begun the process of identifying the additional technical assistance needs of states related to assessing the academic knowledge of limited English proficient students. For this process, the Department is using the resources of our Assessment and Accountability Comprehensive Center (AACC) and the 16 Regional Comprehensive Centers (RCCs) funded by the Department. Specifically, the AACC has developed a plan approved by the Department to provide information and resources to the RCCs, and through the RCCs to State education agencies (SEAS), regarding the assessment and accountability of their special student populations (LEP students and students with disabilities). During 2006- 07 and subsequent years, the AACC will continue to address aspects of assessment and accountability as they relate to LEP students. This fall, the AACC will conduct a needs assessment of the RCCs and SEAS to determine specific areas in which SEAS need assistance and, subsequently, the AACC will use that information to develop resource materials and to provide specific technical assistance. We anticipate that the AACC, in coordination with the Department, will be able to provide the following resources to the RCCs and SEAS during the 2006-07 school year: 1. A summary evaluation of evidence related to the technical aspects of content or language proficiency assessments for LEP students (including the reliability and validity of widely used instruments), and a synopsis of information on existing available resources and on those being developed. The evaluation will yield results that are intended to assist consumers of assessments (e.g., SEAs, LEAs) for LEP students. It will focus on the technical adequacy (i.e., validity, reliability, freedom from bias) of evidence related to assessments used to meet relevant Title I and Title III requirements under the Elementary and Secondary Education Act, as amended by the No Child Left Behind Act (ESEA) (e.g., assessments developed specifically for a state, consortium-developed assessments, assessments published for wider use). Technical evidence associated with assessments for LEP students will be evaluated against a comprehensive set of validated criteria (Rabinowitz & Sato, 2005). Using the results of this evaluation, RCCs and SEAS should be better able to gauge the technical adequacy of the assessments they are using or have available for use with their LEP students and to identify appropriate and necessary next steps for ensuring the assessments' validity and, ultimately, the defensibility of their assessment systems and results. 2. Guidelines for the interpretation of regulations regarding the implementation of assessments for LEP students (under Titles I and III of ESEA), including other pertinent statutes and regulations. These guidelines are expected to address compliance with NCLB regulations for special student populations such as LEP students and will provide information to RCCs as they work with their states in gauging whether they are meeting federal requirements; will focus attention on priority issues related to implementing practices and systems that are in compliance with federal regulations; and will select specific implementation strategies, given the particular needs and conditions of the state. Selected strategies will have evidence of effectiveness in a state or states and will be examples of methods for enhancing and supporting federal guidance. 3. Web-posting and dissemination of evidence-based products and services for easy access to the most current information. The AACC, headed by a national expert on assessment technical quality and staffed by a team of academicians and former practitioners, is well suited to conduct the needs assessment necessary to develop resources and provide technical assistance. The Department will monitor the development of the needs assessment and the subsequent activities of the AACC to ensure that the information being developed is of consistently high quality and relevance. Recommendation 3. Publish additional guidance with more specific information on the requirements for assessing English language proficiency and tracking the progress of students with limited English proficiency in learning English. The Department agrees with this recommendation and is exploring ways - -through guidance to states, the Office of English Language Acquisition's annual Title III Summit, and other means --to help states appropriately assess and track the progress of students with limited proficiency in English. Recommendation 4. Explore ways to provide additional flexibilities to states in terms of holding states accountable for students with limited English proficiency. Such flexibilities could include, for example. * Allowing states to include the assessment scores for all students formerly considered to have limited English proficiency in a state's annual progress results for the group of students with limited English proficiency, * Extending the period during which the assessment scores for some or all students with limited English proficiency would not be included in a state's annual progress results, and: * Adjusting how states account for recent immigrants with little formal schooling in their annual progress results. As your report points out, the Department has explored and has already provided various types of flexibility to states regarding LEP students and how they are considered in states' annual progress results. The Department has engaged in discussions with researchers, practitioners, and educators to explore appropriate means of providing appropriate flexibility in holding schools accountable for the academic progress of children who have not grown up speaking English. NCLB gives states some flexibility in defining the students who constitute the LEP subgroup. States also have some flexibility in determining how to assess their LEP students. States can offer a menu of accommodations (e.g., use of bilingual dictionaries, extra time, use of translators) or develop alternative assessments (e.g., a native- language version or simplified English version of its assessment). The law also allows states three years to test LEP students in language arts using a native language assessment, with an additional two years if needed on a case-by-case basis. The Department recognizes that LEP students new to the United States often have some challenges in participating in state assessments due to language barriers and the challenge of adjusting to U.S. schools. In addition, because students exit the LEP subgroup once they attain English language proficiency, states may have difficulty demonstrating improvements on state assessments for these students. To address these two issues, the Department has received public comments on a notice of proposed rulemaking and is in the process of completing a regulation on flexibility for "recent arrivals" and "formerly LEP" students. In the meantime, the Department is implementing a transitional policy with regard to these policy issues. Under that transitional policy, recently arrived LEP students during the first year of enrollment in U.S. schools can be exempted from taking reading/language arts assessments and the scores of recently arrived LEP students on state math assessments, as well as reading/ language arts assessments if taken, can be excluded from adequate yearly progress (AYP) calculations for that first year in U.S. schools as well. In addition, the flexibility allows states to include, within the LEP subgroup for AYP purposes, for up to two years, the scores on state assessments for "formerly LEP" students who have since attained English proficiency. While the Department has explored and implemented flexibility policies related to LEP students, we also know that regular and high-quality student assessment, school accountability, and clear and accessible data and information about student and school performance are absolutely essential to improving instruction for LEP students. The No Child Left Behind Act shines a bright light on the needs of students who have so often been left behind in our Nation's schools - including students with limited English proficiency. Therefore, the Department believes that all LEP students need to be included and visible in our school accountability systems to improve both instruction and achievement outcomes. The ability to better measure performance and analyze student data is an important vehicle for improving instruction and closing the achievement gap for LEP students and reaching the goal of the No Child Left Behind Act of 100 percent of students being able to master grade level work by 2013-14. We appreciate the opportunity to share our comments, accomplishments and plans. Please let me know if you need additional information regarding activities underway at the Department to assist States better measure the progress of students with limited English proficiency. Sincerely, Signed by: Henry L. Johnson: Enclosure: [End of section] Appendix VIII: GAO Contacts and Acknowledgments: GAO Contacts: Marnie Shaul, (202) 512-7215, shaulm@gao.gov and Cornelia Ashby (202) 512-7215, ashbyc@gao.gov: Staff Acknowledgments: Harriet Ganson (Assistant Director) and Michelle St. Pierre (Analyst- in-Charge) managed all aspects of this assignment. Shannon Groff, Eileen Harrity, and Krista Loose made significant contributions to this report. Katie Brillantes contributed to the initial design of the assignment. Carolyn Boyce, John Mingus, and Lynn Musser provided key technical support, James Rebbe provided legal support, and Scott Heacock assisted in message and report development. [End of section] Related GAO Products: No Child Left Behind Act:, States Face Challenges Measuring Academic Growth That Education's Initiatives May Help Address. GAO-06-661. Washington, D.C.: July 17, 2006. No Child Left Behind Act: Improved Accessibility to Education's Information Could Help States Further Implement Teacher Qualification Requirements. GAO-06-25. Washington, D.C.: November 21, 2005. No Child Left Behind Act: Education Could Do More to Help States Better Define Graduation Rates and Improve Knowledge about Intervention Strategies. GAO-05-879. Washington, D.C.: September 20, 2005. No Child Left Behind Act: Most Students with Disabilities Participated in Statewide Assessments, but Inclusion Options Could Be Improved. GAO- 05-618. Washington, D.C.: July 20, 2005. Head Start: Further Development Could Allow Results of New Test to Be Used for Decision Making. GAO-05-343. Washington, D.C.: May 17, 2005. No Child Left Behind Act: Education Needs to Provide Additional Technical Assistance and Conduct Implementation Studies for School Choice Provision. GAO-05-7. Washington, D.C.: December 10, 2004. No Child Left Behind Act: Improvements Needed in Education's Process for Tracking States' Implementation of Key Provisions. GAO-04-734. Washington, D.C.: September 30, 2004. FOOTNOTES [1] For more information on differences in state approaches to meeting NCLBA requirements, see GAO, No Child Left Behind Act: Improvements Needed in Education's Process for Tracking States' Implementation of Key Provisions, GAO-04-734 (Washington, D.C.: Sept. 30, 2004). [2] These standards are sponsored and published jointly by the American Educational Research Association, the American Psychological Association, and the National Council on Measurement in Education. [3] We included the District of Columbia in our 48-state total. [4] In 7 of the 17 states, students with limited English proficiency met a state's adequate yearly progress goals through NCLBA's safe harbor provision--that is, by decreasing the percentage of students scoring nonproficient by 10 percent or more and showing progress on another academic indicator. [5] We requested district data from 20 states, and 18 states responded to our request. [6] The number of districts reporting results for these students may increase in the future, since states were required to assess students in more grades beginning with the 2005-2006 school year. Testing in more grades will likely increase the number of districts meeting the minimum number of limited English proficient students that will be required to separately report proficiency results. [7] States determine the minimum number of students in a student group, usually between 25 and 45 students. [8] Student groups are not mutually exclusive, with each of the ethnic and racial categories probably including some number of students with limited English proficiency. For example, the results for a student who is both white and limited English proficient would be included in both groups. [9] Under NCLBA, states are required to participate in the National Assessment of Educational Progress for reading and mathematics assessments in grades four and eight. The results from these assessments provide a national measure of student achievement and serve as confirmatory evidence about student achievement on state tests. [10] Example taken from Robert L. Linn, "CRESST Policy Brief 8: Fixing the NCLB Accountability System," Policy Brief of the National Center for Research on Evaluation, Standards, and Student Testing, Summer 2005. [11] Theodore Coladarci, Gallup Goes to School: The Importance of Confidence Intervals for Evaluating "Adequate Yearly Progress" in Small Schools, the Rural School and Community Trust Policy Brief, Oct. 2003, and Thomas J. Kane and Douglas O. Steiger, "Volatility in School Test Scores: Implications for Test-Based Accountability Systems," in Diane Ravitch, ed., Brookings Papers on Education Policy 2002, pp. 235-283. Washington, D.C.: Brookings Institution. [12] Center on Education Policy, "From the Capital to the Classroom: Year 4 of the No Child Left Behind Act," March 2006. [13] When using confidence intervals, upper and lower limits around a district's or state's percentage of proficient students are calculated, creating a range of values within which there is "confidence" the true value lies. For example, instead of saying that 72 percent of students scored at the proficient level or above on a test, a confidence interval may show that percentage to be between 66 and 78 percent, with 95 percent confidence. [14] As of July 2006, Education had conducted peer reviews of 50 states and the District of Columbia. However, detailed peer review notes were available from only 38 states at the time of our review. [15] States were given several years to meet the requirements of the Improving America's School Act of 1994. [16] Christopher Jepsen and Shelley de Alth, "English Learners in California Schools," (San Francisco, California: Public Policy Institute of California, 2005). [17] Charlene Rivera and Eric Collum. An Analysis of State Assessment Policies Addressing the Accommodation of English Language Learners. The George Washington University Center for Equity and Excellence in Education, Arlington, Virginia: (January 2004). [18] Education officials told us that the agency has approved an extension of this deadline for 1 state and is currently considering extension requests from 2 other states. [19] The four consortia include the World-Class Instructional Design and Assessment (WIDA) Consortium, State Collaborative on Assessment and Student Standards (SCASS) Consortium, Mountain West Assessment Consortium (MWAC), and Pennsylvania Enhanced Assessment Grant (PA EAG) Consortium. [20] Although these assessments are not fully aligned to state standards, Education officials noted that they have not yet provided states guidance on what level of alignment the agency expects. Thus, some of these may ultimately meet Education's requirements. [21] Stanley Rabinowitz and Edynn Sato, "Evidence-Based Plan: Technical Adequacy of Assessments for Alternate Student Populations: A Technical Review of High-Stakes Assessments for English Language Learners," WestEd (December 2005). [22] The states providing these comments represent more than our 5 case study states. We also contacted officials in 28 additional states to determine what English language proficiency assessment they planned to use in 2006. [23] Education does not require state English language proficiency tests to be aligned to state academic standards. However, states' English language proficiency tests and academic standards are connected, in that Education requires that state's English language proficiency tests be aligned to state English language proficiency standards and NCLBA requires that state English language proficiency standards be aligned with state academic standards. [24] We reviewed reports from Title III monitoring visits of 13 states conducted between April and October 2005 that were available from Education as of March 31, 2006. [25] Education has sent letters to the remaining states outlining the issues that need to be resolved in order for their assessment systems to be approved. [26] On June 24, 2004, Education issued proposed regulations on these flexibilities for students with limited English proficiency for public comment, but the regulations have not been finalized as of June 2006. [27] See GAO, No Child Left Behind Act: States Face Challenges Measuring Academic Growth That Education's Initiatives May Help Address, GAO-06-661 (Washington, D.C.: July 17, 2006) for further information on Education's pilot project. [28] NCLBA also requires Education to annually determine whether states have met their Title III goals related to increases in students making progress in learning English and attaining English proficiency. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548: