This is the accessible text file for GAO report number GAO-10-18 entitled 'Student Achievement: Schools Use Multiple Strategies to Help Students Meet Academic Standards, Especially Schools with Higher Proportions of Low-Income and Minority Students' which was released on November 16, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Committees: United States Government Accountability Office: GAO: November 2009: Student Achievement: Schools Use Multiple Strategies to Help Students Meet Academic Standards, Especially Schools with Higher Proportions of Low-Income and Minority Students: GAO-10-18: GAO Highlights: Highlights of GAO-10-18, a report to congressional committees. Why GAO Did This Study: The federal government has invested billions of dollars to improve student academic performance, and many schools, teachers, and researchers are trying to determine the most effective instructional practices with which to accomplish this. The Conference Report for the Consolidated Appropriations Act for Fiscal Year 2008 directed GAO to study strategies used to prepare students to meet state academic achievement standards. To do this, GAO answered: (1) What types of instructional practices are schools and teachers most frequently using to help students achieve state academic standards, and do those instructional practices differ by school characteristics? (2) What is known about how standards-based accountability systems have affected instructional practices? (3) What is known about instructional practices that are effective in improving student achievement? GAO analyzed data from a 2006-2007 national survey of principals and 2005- 2006 survey of teachers in three states, conducted a literature review of the impact of standards-based accountability systems on instructional practices and of practices that are effective in improving student achievement, and interviewed experts. What GAO Found: Nationwide, most principals focused on multiple strategies to help students meet academic standards, such as using student data to inform instruction and increasing professional development for teachers, according to our analysis of data from a U.S. Department of Education survey. Many of these strategies were used more often at high-poverty schools—those where 75 percent or more of the students were eligible for the free and reduced-price lunch program—and high-minority schools— those where 75 percent or more of students were identified as part of a minority population, than at lower poverty and minority schools. Likewise, math teachers in California, Georgia, and Pennsylvania increased their use of certain instructional practices in response to their state tests, such as focusing more on topics emphasized on assessments and searching for more effective teaching methods, and teachers at high-poverty and high-minority schools were more likely than teachers at lower-poverty schools and lower-minority schools to have made these changes, according to GAO’s analysis of survey data collected by the RAND Corporation. Some researchers suggested that differences exist in the use of these practices because schools with lower poverty or lower minority student populations might generally be meeting accountability requirements and therefore would need to try these strategies less frequently. Research shows that standards-based accountability systems can influence instructional practices in both positive and negative ways. For example, some research notes that using a standards-based curriculum that is aligned with corresponding instructional guidelines can facilitate the development of higher order thinking skills in students. But, in some cases, teacher practices did not always reflect the principles of standards-based instruction, and the difficulties in aligning practice with standards were attributed, in part, to current accountability requirements. Other research noted that assessments can be powerful tools for improving the learning process and evaluating student achievement, but assessments can also have some unintended negative consequences on instruction, including narrowing the curriculum to only material that is tested. Many experts stated that methodological issues constrain knowing more definitively the specific instructional practices that improve student learning and achievement. Nevertheless, some studies and experts pointed to instructional practices that are considered to be effective in raising student achievement, such as differentiated instruction. Professional development for teachers was also highlighted as important for giving teachers the skills and knowledge necessary to implement effective teaching practices. What GAO Recommends: GAO makes no recommendations in this report. Education provided comments about issues pertaining to the study’s approach that it believes should be considered. GAO clarified the report as appropriate. View [hyperlink, http://www.gao.gov/products/GAO-10-18] or key components. For more information, contact Cornelia Ashby at (202) 512- 7215 or AshbyC@gao.gov. [End of section] Contents: Letter: Background: Principals and Teachers Used a Variety of Instructional Practices to Help Students Meet Standards, and Many of These Practices Were Used More Frequently at Schools with Higher Proportions of Low-Income and Minority Students: Research Shows That Standards-based Accountability Systems Can Influence Instructional Practices through Standards and Assessments in Both Positive and Negative Ways: Research Highlights Some Potentially Successful Practices for Improving Student Achievement, although Experts Contend That Methodological Issues Constrain Reaching Definitive Conclusions about What Works: Agency Comments and Our Evaluation: Appendix I: Scope and Methodology: Appendix II: Analyses of the Relationship between School Characteristics and Principals' Focus on School Improvement Strategies: Appendix III: List of Education Researchers: Appendix IV: Studies Meeting GAO's Criteria for Methodological Quality: Appendix V: Comments from the Department of Education: Appendix VI: GAO Contact and Staff Acknowledgments: Table: Table 1: Odds Ratios Indicating the Difference in Likelihood of Principals to Make School Improvement Strategies a Moderate or Major Focus after Controlling for Different Factors: Figures: Figure 1: Principals' Responses Indicating That a School Improvement Strategy Was a Major or Moderate Focus of the School Improvement Efforts: Figure 2: Percent of Elementary and Middle School Math Teachers Who Reported Increasing Their Use of Certain Instructional Practices as a Result of State Test: Figure 3: How Survey Responses Differed between Math Teachers at High- Poverty and Low-Poverty Schools in Three States: Abbreviations: AYP: adequate yearly progress: ESEA: Elementary and Secondary Education Act of 1965: IASA: Improving America's Schools Act of 1994: NCLBA: No Child Left Behind Act of 2001: NLS-NCLB: National Longitudinal Study of No Child Left Behind: NSF: National Science Foundation: RAND: The RAND Corporation: [End of section] United States Government Accountability Office: Washington, DC 20548: November 16, 2009: Congressional Committees: The federal government has invested billions of dollars to help schools meet requirements of the No Child Left Behind Act of 2001 (NCLBA) to improve student academic performance in reading, math, and science. [Footnote 1] To this end, many schools, teachers, and researchers are trying to determine the most effective instructional practices to improve student achievement. Instructional practices refer to school or district-level improvement strategies, such as aligning curriculum with academic standards, restructuring the school day, or providing additional professional development to teachers.[Footnote 2] Instructional practices can also refer to classroom teaching practices like assigning more homework or searching for more effective teaching methods. Little is known about the extent to which instructional practices have changed in response to NCLBA's accountability requirements, whether these practices vary by type of school, and the extent to which some practices have proven to be more effective than others. Under NCLBA, states are required to develop challenging student academic achievement standards, administer tests based on those standards (standards-based assessments) to measure student proficiency, and develop targets for performance on these tests. Specifically, NCLBA requires states to develop a plan to ensure that their students are making adequate yearly progress (AYP) toward proficiency in reading, math, and science by 2014 for students collectively and in key student subgroups, including low-income and minority students. While NCLBA creates requirements for student proficiency, it generally allows states to determine how best to meet those requirements. The Conference Report accompanying the Consolidated Appropriations Act for Fiscal Year 2008 directed that GAO conduct a study of strategies used to prepare students to meet state academic achievement standards. In response, we agreed with the Senate and House Appropriations Committees, the Senate Committee on Health, Education, Labor and Pensions, and the House Committee on Education and Labor to address the following questions: 1. What types of instructional practices are schools and teachers most frequently using to help students achieve state academic standards, and do those instructional practices differ by school characteristics? 2. What is known about how standards-based accountability systems such as that in NCLBA have affected instructional practices? 3. What is known about instructional practices that are effective in improving student achievement? To answer these questions, we analyzed data from two recent surveys of principals and teachers that were conducted by the RAND Corporation (RAND). The first survey, the nationally representative National Longitudinal Study of No Child Left Behind (NLS-NCLB), was sponsored by the U.S. Department of Education (Education) and asked principals the extent to which their schools were focusing on certain strategies in their school improvement efforts.[Footnote 3] We conducted an analysis of the school year 2006-2007 survey data on school improvement strategies by controlling for school characteristic variables, such as the percentage of a school's students receiving free or reduced price lunch (poverty); the percentage of students who are a racial minority (minority); whether the school is in an urban, urban fringe or large town, or rural area (school location); and the school's AYP performance status. The second survey, a three-state survey sponsored by the National Science Foundation (NSF), asked elementary and middle school teachers in California, Georgia, and Pennsylvania how their classroom teaching strategies differed due to a state math test.[Footnote 4] RAND selected these states to represent a range of approaches to standards- based accountability and to provide some geographic and demographic diversity. Using school year 2005-2006 data from the three-state survey, which is representative only of those three states individually, we measured associations between the teacher responses and the school characteristic variables. As part of these survey analyses, we reviewed documentation and performed electronic testing of the data obtained through the surveys and conducted interviews with the primary RAND researchers responsible for the data collection and analysis. We determined the survey data were sufficiently reliable for the purposes of our study. To answer questions two and three, we conducted a literature review and synthesis.[Footnote 5] We supplemented our synthesis by interviewing prominent education researchers identified in frequently cited articles and through discussions with other knowledgeable individuals[Footnote 6]. We also reviewed relevant federal laws and regulations. We conducted our work from July 2008 to November 2009 in accordance with all sections of GAO's Quality Assurance Framework that are relevant to our objectives. The framework requires that we plan and perform the engagement to obtain sufficient and appropriate evidence to meet our stated objectives and to discuss any limitations in our work. We believe that the information and data obtained, and the analysis conducted, provide a reasonable basis for any findings and conclusions in this product. Background: NCLBA reauthorized the Elementary and Secondary Education Act of 1965 (ESEA)[Footnote 7] and built upon accountability requirements created under a previous reauthorization, the Improving America's Schools Act of 1994 (IASA).[Footnote 8] Under ESEA, as amended, Congress sought to improve student learning by incorporating academic standards and assessments in the requirements placed on states. Academic standards, which describe what students should know and be able to do at different grade levels in different subjects, help guide school systems in their choice of curriculum and help teachers plan for classroom instruction. Assessments, which states use to measure student progress in achieving the standards, are required to be administered by states. NCLBA further strengthened some of the accountability requirements contained in ESEA, as amended. Specifically, NCLBA's accountability provisions require states to develop education plans that establish academic standards and performance goals for schools to meet AYP and lead to 100 percent of their students being proficient in reading, math, and science by 2014. This proficiency must be assessed annually in reading and math in grades 3 through 8 and periodically in science, whereas assessments were required less frequently under the IASA. [Footnote 9] Under NCLBA, schools' assessment data generally must be disaggregated to assess progress toward state proficiency targets for students in certain designated groups, including low-income students, minority students, students with disabilities, and those with limited English proficiency. Each of these groups must make AYP in order for the school to make AYP. Schools that fail to make AYP for 2 or more consecutive years are required to implement various improvement measures identified in NCLBA, and these measures are more extensive than those required under IASA. Education, which has responsibility for general oversight of NCLBA, reviews and approves state plans for meeting AYP requirements. As we have previously reported, Education had approved all states' plans--fully or conditionally--by June 2003. [Footnote 10] NCLBA also recognizes the role of teachers in providing a quality education by requiring states to ensure that all teachers in core academic subjects are "highly qualified." Under this requirement, teachers generally must have a bachelor's degree, be fully certified, and demonstrate their knowledge of the subjects they teach. Previously, there were no specific requirements regarding teacher quality under ESEA, as amended.[Footnote 11] Principals and Teachers Used a Variety of Instructional Practices to Help Students Meet Standards, and Many of These Practices Were Used More Frequently at Schools with Higher Proportions of Low-Income and Minority Students: According to our analysis of NLS-NCLB data from Education, most principals reported their schools focused on multiple instructional practices in their voluntary school improvement efforts.[Footnote 12] These strategies were used more often at schools with higher proportions of low-income students ("high-poverty schools") and schools with higher proportions of minority students ("high-minority schools") than at schools with lower proportions of low-income students ("low- poverty schools") and schools with lower proportions of minority students ("low-minority schools").[Footnote 13] Likewise, the survey of math teachers in California, Georgia, and Pennsylvania indicates teachers were using many different instructional practices in response to their state tests, and teachers at high-poverty and high-minority schools were more likely than teachers at low-poverty and low-minority schools to have been increasing their use of some of these practices. Some researchers we spoke with suggested that differences in the use of these instructional practices exist because schools with low-poverty or low-minority student populations might generally be meeting accountability standards and, therefore, would need to try these strategies less frequently. Principals at High-Poverty and High-Minority Schools Emphasized Certain School Improvement Strategies More Than Principals at Other Schools: According to nationally representative data from Education's NLS-NCLB, in school year 2006-2007 most principals focused on multiple strategies in their school improvement efforts. The survey asked principals the extent to which their schools were focusing on ten different strategies in their voluntary school improvement initiatives. The three most common strategies were: (1) using student achievement data to inform instruction and school improvement; (2) providing additional instruction to low-achieving students; and (3) aligning curriculum and instruction with standards and/or assessments. (See figure 1.) Nearly all school principals placed a major or moderate focus on three or more surveyed strategies in their school improvement efforts, and over 80 percent of principals placed a major or moderate focus on six or more strategies. However, as Education's report on the survey data cautioned, the number of improvement strategies emphasized was not necessarily an indication of the intensity or quality of the improvement efforts. Figure 1: Principals' Responses Indicating That a School Improvement Strategy Was a Major or Moderate Focus of the School Improvement Efforts: [Refer to PDF for image: illustration] School improvement strategies: Using student achievement data to inform instruction and school improvement; Percent saying major or moderate focus: 96%. School improvement strategies: Providing additional instruction to low- achieving students; Percent saying major or moderate focus: 92%. School improvement strategies: Aligning curriculum and instruction with standards and/or assessments; Percent saying major or moderate focus: 91%. School improvement strategies: Increasing the intensity, focus, and effectiveness of professional development; Percent saying major or moderate focus: 85%. School improvement strategies: Implementing new instructional approaches or curricula in reading; Percent saying major or moderate focus: 84%. School improvement strategies: Implementing new instructional approaches or curricula in mathematics; Percent saying major or moderate focus: 80%. School improvement strategies: Providing extended-time instructional programs (e.g., before-school, after-school, or weekend instructional programs); Percent saying major or moderate focus: 64%. School improvement strategies: Restructuring the school day to teach core content areas in greater depth (e.g., establishing a literacy block); Percent saying major or moderate focus: 61%. School improvement strategies: Implementing strategies for increasing parents’ involvement in their children’s education; Percent saying major or moderate focus: 54%. School improvement strategies: Increasing instructional time for all students (e.g., by lengthening the school day or year, shortening recess); Percent saying major or moderate focus: 32%. Sources: GAO analysis of school year 2006-2007 NLS-NCLB survey data, Art Explosion (images). Note: Some of the voluntary school improvement strategies identified above are similar to the corrective actions and restructuring options schools identified for improvement under NCLBA are required to choose from in preparing their school improvement plan. For example, implementing a new curriculum and extending the school day are both voluntary improvement strategies and possible strategies for improvement under the law. [End of figure] While nearly all principals responded that they used multiple improvement strategies, there were statistically significant differences in principals' responses across a range of school characteristics, including percentage of the school's students receiving free or reduced price lunch (poverty), percentage of minority students, the school's location, and AYP status.[Footnote 14] For example, when comparing schools across poverty levels, we found that principals at high-poverty schools were two to three times more likely than principals at low-poverty schools to focus on five particular strategies in their school improvement efforts: * Restructuring the school day to teach core content areas in greater depth;[Footnote 15] * Increasing instructional time for all students (e.g., by lengthening the school day or year, shortening recess); * Providing extended-time instructional programs (e.g., before-school, after-school, or weekend instructional programs); * Implementing strategies for increasing parents' involvement in their children's education; and: * Increasing the intensity, focus, and effectiveness of professional development.[Footnote 16] Likewise, when comparing schools across minority levels, we found that principals at high-and moderate-minority schools were approximately two to three times more likely than principals at low-minority schools to make six particular school improvement strategies a major or moderate focus of their school improvement efforts.[Footnote 17] For instance, principals at schools with a high percentage of minority students were more than three times as likely as principals at schools with a low percentage of minority students to provide extended-time instruction such as after-school programs. A school's location was associated with differences in principals' responses about the strategies they used as well: principals at rural schools were only about one-third to one-half as likely as central city schools to make five of these school improvement strategies a moderate or major focus of their school improvement efforts.[Footnote 18] When we compared principal responses based on AYP status, there was some evidence of a statistically significant association between AYP status and the extent to which principals focused these strategies in their school improvement efforts, but it was limited when the other variables such as poverty and minority were taken into account. AYP status had some correlation with the demographic characteristics of poverty and minority, and those characteristics explained the patterns of principals' responses more fully than the AYP characteristic. However, our analysis generally showed that schools that had not made AYP were more likely to make six of these school improvement strategies a moderate or major focus of their school improvement plan than schools that had made AYP. Additionally, Education reported that schools identified for improvement under NCLBA--that is, schools that have not made AYP for two or more consecutive years--were engaged in a greater number of improvement efforts than non-identified schools. Therefore, principals of the non-identified schools may have been less likely than principals of identified schools to view specific strategies as a major or moderate focus. We spoke with several researchers about the results of our analysis of the principals' responses, especially at high-poverty and high-minority schools. While the researchers could not say with certainty the reasons for the patterns, they noted that high-poverty and high-minority schools tend to be most at risk of not meeting their states' standards, so that principals at those schools might be more willing to try different approaches. Conversely, the researchers noted that principals at schools meeting standards would not have the same incentives to adopt as many school improvement strategies. Most Math Teachers in Three Surveyed States Have Increased Their Use of Certain Instructional Practices in Response to State Tests, especially in High-Poverty and High-Minority Schools: The RAND survey of elementary and middle school math teachers in California, Georgia and Pennsylvania showed that in each of the three states at least half of the teachers reported increasing their use of certain instructional practices in at least five areas as a result of the statewide math test (see figure 2). For example, most teachers in Pennsylvania responded that due to the state math test they: (1) focused more on standards, (2) emphasized assessment styles and formats, (3) focused more on subjects tested, (4) searched for more effective teaching methods, and (5) spent more time teaching content. Figure 2: Percent of Elementary and Middle School Math Teachers Who Reported Increasing Their Use of Certain Instructional Practices as a Result of State Test: [Refer to PDF for image: illustration] Change in instructional practices: Focus more on standards; Percent change in California: 72%; Percent change in Georgia: 75%; Percent change in Pennsylvania: 75%. Change in instructional practices: Focus more on topics emphasized in assessment; Percent change in California: 63%; Percent change in Georgia: 72%; Percent change in Pennsylvania: 71%. Change in instructional practices: Emphasize assessment styles and formats of problems; Percent change in California: 53%; Percent change in Georgia: 76%; Percent change in Pennsylvania: 72%. Change in instructional practices: Search for more effective teaching methods; Percent change in California: 65%; Percent change in Georgia: 72%; Percent change in Pennsylvania: 59%. Change in instructional practices: Spend more time teaching content; Percent change in California: 50%; Percent change in Georgia: 56%; Percent change in Pennsylvania: 52%. Change in instructional practices: Spend more time teaching test-taking strategies; Percent change in California: 51%; Percent change in Georgia: 53%; Percent change in Pennsylvania: 49%. Change in instructional practices: Focus more on students who are close to proficient; Percent change in California: 34%; Percent change in Georgia: 37%; Percent change in Pennsylvania: 28%. Change in instructional practices: Assign more homework; Percent change in California: 40%; Percent change in Georgia: 30%; Percent change in Pennsylvania: 28%. Change in instructional practices: Rely more heavily on open-ended tests; Percent change in California: 19%; Percent change in Georgia: 24%; Percent change in Pennsylvania: 47%. Change in instructional practices: Offer more assistance outside of school for students who are not proficient; Percent change in California: 29%; Percent change in Georgia: 36%; Percent change in Pennsylvania: 21%. Change in instructional practices: Rely more heavily on multiple-choice tests; Percent change in California: 24%; Percent change in Georgia: 37%; Percent change in Pennsylvania: 17%. Sources: GAO analysis of 2005 survey data from Standards-Based Accountability Under No Child Left Behind: Experiences of Teachers and Administrators in Three States. Hamilton et al. Art Explosion (images). [End of figure] As we did with the survey responses of principals, we analyzed the teacher survey data to determine whether math teachers' responses differed by school characteristics for poverty, minority, location, and AYP status. As with the principals' responses, we found that elementary and middle school math teachers in high-poverty and high-minority schools were more likely than teachers in low-poverty and low-minority schools to report increasing their use of certain instructional practices, and this pattern was consistent across the three states (see figure 3). For example, 69 percent of math teachers at high-poverty schools in California indicated they spent more time teaching test- taking strategies as opposed to 38 percent of math teachers in low- poverty schools. In Georgia, 50 percent of math teachers in high- poverty schools reported offering more outside assistance to non- proficient students in contrast to 26 percent of math teachers in low- poverty schools. Fifty-one percent of math teachers at high-poverty schools in Pennsylvania reported focusing more attention on students close to proficiency compared to 23 percent of math teachers doing so in low poverty schools.[Footnote 19] Figure 3: How Survey Responses Differed between Math Teachers at High- Poverty and Low-Poverty Schools in Three States: [Refer to PDF for image: vertical bar graph] Percentage of teachers changing instructional practices as a result of state math test: Way instruction changed: Assign more or more difficult homework: California*: Low poverty: 28%; High poverty: 39%; Georgia: Low poverty: 21%; High poverty: 44%; Pennsylvania*: Low poverty: 28%; High poverty: 34%. Way instruction changed: Search for more effective teaching methods: California: Low poverty: 48%; High poverty: 75%; Georgia: Low poverty: 58%; High poverty: 81%; Pennsylvania: Low poverty: 54%; High poverty: 75%. Way instruction changed: Focus more on state standards: California: Low poverty: 58%; High poverty: 79%; Georgia: Low poverty: 62%; High poverty: 87%; Pennsylvania: Low poverty: 72%; High poverty: 91%. Way instruction changed: Focus more on state test topics: California: Low poverty: 55%; High poverty: 76%; Georgia: Low poverty: 51%; High poverty: 78%; Pennsylvania: Low poverty: 64%; High poverty: 86%. Way instruction changed: Emphasize state test problem formats: California: Low poverty: 42%; High poverty: 73%; Georgia: Low poverty: 59%; High poverty: 85%; Pennsylvania: Low poverty: 75%; High poverty: 85%. Way instruction changed: More time teaching general test-taking strategies; California: Low poverty: 38%; High poverty: 69%; Georgia: Low poverty: 37%; High poverty: 70%; Pennsylvania: Low poverty: 46%; High poverty: 69%. Way instruction changed: More time teaching math content; California: Low poverty: 38%; High poverty: 67%; Georgia: Low poverty: 42%; High poverty: 70%; Pennsylvania: Low poverty: 50%; High poverty: 67%. Way instruction changed: Focus more on students close to proficient; California: Low poverty: 25%; High poverty: 48%; Georgia: Low poverty: 34%; High poverty: 46%; Pennsylvania: Low poverty: 23%; High poverty: 51%. Way instruction changed: Offer outside assistance to non-proficient students; California: Low poverty: 18%; High poverty: 41%; Georgia: Low poverty: 26%; High poverty: 50%; Pennsylvania: Low poverty: 16%; High poverty: 43%. Way instruction changed: Rely more on multiple-choice tests; California: Low poverty: 19%; High poverty: 46%; Georgia: Low poverty: 26%; High poverty: 47%; Pennsylvania: Low poverty: 15%; High poverty: 38%. Way instruction changed: Rely more on open-ended questions on tests; California: Low poverty: 10%; High poverty: 27%; Georgia*: Low poverty: 23%; High poverty: 30%; Pennsylvania: Low poverty: 42%; High poverty: 60%. * = Not statistically significant with a 95 percent level of confidence for the difference between high-poverty and low-poverty schools. Sources: GAO analysis of school year 2005-2006 survey data from Standards-Based Accountability Under No Child Left Behind:Experiences of Teachers and Administrators in Three States. Hamilton et al. [End of figure] Similar to what our poverty analysis showed, survey responses provided some evidence that math teachers in high-minority schools were more likely than those in low-minority schools to change their instructional practices. Math teachers at high-minority schools in each of the three states, as compared to those at low-minority schools, were more likely to: * rely on open-ended tests in their own classroom assessments; * increase the amount of time spent teaching mathematics by replacing non-instructional activities with mathematics instruction; * focus on topics emphasized in the state math test; and: * teach general test-taking strategies. We also analyzed the RAND data with regard to school location and a school's AYP status, but results from these characteristics were not significant for as many instructional practices.[Footnote 20] As we did regarding the survey responses of principals, we spoke to several researchers, including the authors of the three-state teacher study, regarding possible reasons for the patterns we saw in the teacher survey data. The researchers we spoke with provided similar possible reasons for the patterns in the teacher survey as they did for patterns in the principal survey. For instance, the researchers noted that high-poverty and high-minority schools are more likely to be at risk of failing to meet the state standards, which might prompt teachers to try different approaches. On the other hand, the researchers stated that teachers at those schools meeting the standards would not have the same incentives to change their instructional practices. Research Shows That Standards-based Accountability Systems Can Influence Instructional Practices through Standards and Assessments in Both Positive and Negative Ways: Research shows that using a standards-based curriculum that is aligned with corresponding instructional guidelines can positively influence teaching practices. Specifically, some studies reported changes by teachers who facilitated their students developing higher-order thinking skills, such as interpreting meaning, understanding implied reasoning, and developing conceptual knowledge, through practices such as multiple answer problem solving, less lecture and more small group work. Additionally, a few researchers we interviewed stated that a positive effect of NCLBA's accountability provisions has been a renewed focus on standards and curriculum.[Footnote 21] However, some studies indicated that teachers' practices did not always reflect the principles of standards-based instruction and that current accountability policies help contribute to the difficulty in aligning practice with standards. Some research shows that, while teachers may be changing their instructional practices in response to standards- based reform, these changes may not be fully aligned with the principles of the reform. That research also notes that the reliability in implementing standards in the classroom varied in accordance with teachers' different beliefs in and support for standards-based reform as well as the limitations in their instructional capabilities. For example, one observational study of math teachers showed that, while teachers implemented practices envisioned by standards-based reform, such as getting students to work in small groups or using manipulatives (e.g., cubes or tiles), their approaches did not go far enough in that students were not engaged in conversations about mathematical or scientific concepts and ideas.[Footnote 22] To overcome these challenges, studies point to the need for teachers to have opportunities to learn, practice, and reflect on instructional practices that incorporate the standards, and then to observe their effects on student learning. However, some researchers have raised concerns that current accountability systems' focus on test scores and mandated timelines for achieving proficiency levels for students do not give teachers enough time to learn, practice, and reflect on instructional practices and may discourage some teachers from trying ambitious teaching practices envisioned by standards-based reform. Another key element of a standards-based accountability system is assessments, which help measure the extent to which schools are improving student learning through assessing student performance against the standards. Some researchers note that assessments are powerful tools for managing and improving the learning process by providing information for monitoring student progress, making instructional decisions, evaluating student achievement, and evaluating programs. In addition, assessments can also influence instructional content and help teachers use or adjust specific classroom practices. As one synthesis concluded, assessments can influence whether teachers broaden or narrow the curriculum, focus on concepts and problem solving--or emphasize test preparation over subject matter content. [Footnote 23] In contrast, some of the research and a few experts we interviewed raised concerns about testing formats that do not encourage challenging teaching practices and instructional practices that narrow the curriculum as a result of current assessment practices.[Footnote 24] For example, depending on the test used, research has shown that teachers may be influenced to use teaching approaches that reflect the skills and knowledge to be tested. Multiple choice tests tend to focus on recognizing facts and information while open-ended formats are more likely to require students to apply critical thinking skills. Conclusions from a literature synthesis conducted by the Department of Education stated that " teachers respond to assessment formats used, so testing programs must be designed and administered with this influence in mind. Tests that emphasize inquiry, provide extended writing opportunities, and use open-ended response formats or a portfolio approach tend to influence instruction in ways quite different from tests that use closed-ended response formats and which emphasize procedures."[Footnote 25] We recently reported that states have most often chosen multiple choice items over other item types of assessments because they are cost effective and can be scored within tight time frames. While multiple choice tests provide cost and time saving benefits to states, the use of multiple choice items make it difficult, if not impossible, to measure highly complex content.[Footnote 26] Other research has raised concerns that, to avoid potential consequences from low-scoring assessment results under NCLBA, teachers are narrowing the curriculum being taught--sometimes referred to as "teaching to the test"--either by spending more classroom time on tested subjects at the expense of other non-tested subjects, restricting the breadth of content covered to focus only on the content covered by the test, or focusing more time on test-taking strategies than on subject content.[Footnote 27] Research Highlights Some Potentially Successful Practices for Improving Student Achievement, although Experts Contend That Methodological Issues Constrain Reaching Definitive Conclusions about What Works: Our literature review found some studies that pointed to instructional practices that appear to be effective in raising student achievement. But, in discussing the broader implications of these studies with the experts that we interviewed, many commented that, taken overall, the research is not conclusive about which specific instructional practices improve student learning and achievement. Some researchers stated that this was due to methodological issues in conducting the research. For example, one researcher explained that, while smaller research studies on very specific strategies in reading and math have sometimes shown powerful relationships between the strategy used and positive changes in student achievement, results from meta-analyses of smaller studies have been inconclusive in pointing to similar patterns in the aggregate. A few other researchers stated that the lack of empirical data about how instruction unfolds in the classroom hampers the understanding about what works in raising student performance. A few researchers also noted that conducting research in a way that would yield more conclusive results is difficult. One of the main difficulties, as explained by one researcher, is the number of variables a study may need to examine or control for in order to understand the effectiveness of a particular strategy, especially given the number of interactions these variables could have with each other. One researcher mentioned cost as a challenge when attempting to gather empirical data at the classroom level, stating "teaching takes place in the classroom, but the expense of conducting classroom-specific evaluations is a serious barrier to collecting this type of data." Finally, even when research supports the efficacy of a strategy, it may not work with different students or under varying conditions. In raising this point, one researcher stated that "educating a child is not like making a car" whereby a production process is developed and can simply be repeated again and again. Each child learns differently, creating a challenge for teachers in determining the instructional practices that will work best for each student. Some of the practices identified by both the studies and a few experts as those with potential for improving student achievement were: * Differentiated instruction. In this type of instruction, teaching practices and plans are adjusted to accommodate each student's skill level for the task at hand. Differentiated instruction requires teachers to be flexible in their teaching approach by adjusting the curriculum and presentation of information for students, thereby providing multiple options for students to take in and process information. As one researcher described it, effective teachers understand the strategies and practices that work for each student and in this way can move all students forward in their learning and achievement. * More guiding, less telling. Researchers have identified two general approaches to teaching: didactic and interactive. Didactic instruction relies more on lecturing and demonstrations, asking short answer questions, and assessing whether answers are correct. Interactive instruction focuses more on listening and guiding students, asking questions with more than one correct answer, and giving students choices during learning. As one researcher explained, both teaching approaches are important, but some research has shown that giving students more guidance and less direction helps students become critical and independent thinkers, learn how to work independently, and assess several potential solutions and apply the best one. These kinds of learning processes are important for higher-order thinking. However, implementing "less instruction" techniques requires a high level of skill and creativity on the part of the teacher.[Footnote 28] * Promoting effective discourse. An important corollary to the teacher practice of guiding students versus directing them is effective classroom discussion. Research highlights the importance of developing students' understanding not only of the basic concepts of a subject, but higher-order thinking and skills as well. To help students achieve understanding, it is necessary to have effective classroom discussion in which students test and revise their ideas, and elaborate on and clarify their thinking. In guiding students to an effective classroom discussion, teachers must ask engaging and challenging questions, be able to get all students to participate, and know when to provide information or allow students to discover it for themselves. Additionally, one synthesis of several experimental studies examining practices in elementary math classrooms identified two instructional approaches that showed positive effects on student learning. The first was cooperative learning in which students work in pairs or small teams and are rewarded based on how well the group learns. The other approach included programs that helped teachers introduce math concepts and improve skills in classroom management, time management, and motivation. This analysis also found that using computer-assisted instruction had moderate to substantial effects on student learning, although this type of instruction was always supplementary to other approaches or programs being used. We found through our literature review and interviews with researchers that the issue of effective instructional practices is intertwined with professional development. To enable all students to achieve the high standards of learning envisioned by standards-based accountability systems, teachers need extensive skills and knowledge in order to use effective teaching practices in the classroom. Given this, professional development is critical to supporting teachers' learning of new skills and their application. Specifically, the research concludes that professional development will more likely have positive impacts on both teacher learning and student achievement if it: * Focuses on a content area with direct links to the curriculum; * Challenges teachers intellectually through reflection and critical problem solving; * Aligns with goals and standards for student learning; * Lasts long enough so that teachers can practice and revise their techniques; * Occurs collaboratively within a teacher learning community--ongoing teams of teachers that meet regularly for the purposes of learning, joint lesson planning, and problem solving; * Involves all the teachers within a school or department; * Provides active learning opportunities with direct applications to the classroom; and: * Is based on teachers' input regarding their learning needs. Some researchers have raised concerns about the quality and intensity of professional development currently received by many teachers nationwide. One researcher summarized these issues by stating that professional development training for teachers is often too short, provides no classroom follow up, and models more "telling than guiding" practices. Given the decentralized nature of the U.S. education system, the support and opportunity for professional development services for teachers varies among states and school districts, and there are notable examples of states that have focused resources on various aspects of professional development. Nevertheless, shortcomings in teachers' professional development experiences overall are especially evident when compared to professional development requirements for teachers in countries whose students perform well on international tests, such as the Trends in International Mathematics and Science Study and the Program for International Student Assessment. For example, one study showed that fewer than 10 percent of U.S. math teachers in school year 2003-04 experienced more than 24 hours of professional development in mathematics content or pedagogy during the year; conversely, teachers in Sweden, Singapore, and the Netherlands are required to complete 100 hours of professional development per year.[Footnote 29] Agency Comments and Our Evaluation: We provided a copy of our draft report to the Secretary of Education for review and comment. Education's written comments, which are contained in appendix V, expressed support for the important questions that the report addresses and noted that the American Recovery and Reinvestment Act of 2009 included $250 million to improve assessment and accountability systems. The department specifically stated that the money is for statewide data systems to provide information on individual student outcomes that could help enable schools to strengthen instructional practices and improve student achievement. However, the department raised several issues about the report's approach. Specifically, the department commented that we (1) did not provide the specific research citations throughout the report for each of our findings or clearly explain how we selected our studies; (2) mixed the opinions of education experts with our findings gleaned from the review of the literature; (3) did not present data on the extent to which test formats had changed or on the relationship between test format and teaching practices when discussing our assessment findings; and (4) did not provide complete information from an Education survey regarding increases and decreases in instructional time. As stated in the beginning of our report, the list of studies we reviewed and used for our findings are contained in appendix IV. We provide a description in appendix I of our criteria, the types of databases searched, the types of studies examined (e.g., experimental and nonexperimental) and the process by which we evaluated them. We relied heavily on two literature syntheses conducted by the Department of Education--Standards in Classroom Practice: Research Synthesis and The Influence of Standards on K-12 Teaching and Student Learning: A Research Synthesis, which are included in the list. These two syntheses covered, in a more comprehensive way than many of the other studies that we reviewed, the breadth of the topics that we were interested in and included numerous research studies in their reviews. Many of the findings in this report about the research are taken from the conclusions reached in these syntheses. However, to make this fact clearer and more prominent, we added this explanation to our abbreviated scope and methodology section on page 5 of the report. Regarding the use of expert opinion, we determined that obtaining the views of experts about the research we were reviewing would be critical to our understanding its broader implications. This was particularly important given the breadth and scope of our objectives. The experts we interviewed, whose names and affiliations are listed in appendix III, are prominent researchers who conduct, review, and reflect on the current research in the field, and whose work is included in some of the studies we reviewed, including the two literature syntheses written by the Department of Education and used by us in this study. We did not consider their opinions "conjecture" but grounded in and informed by their many years of respected work on the topic. We have been clear in the report as to when we are citing expert opinion, the research studies, or both. Regarding the report section discussing the research on assessments, it was our intent to highlight that, according to the research, assessments have both positive and negative influences on classroom teaching practices, not to conclude that NCLBA was the cause of either. Our findings in this section of the report are, in large part, based on conclusions from the department's syntheses mentioned earlier. For example, The Influence of Standards on K-12 Teaching and Student Learning: A Research Synthesis states "… tests matter--the content covered, the format used, and the application of their results--all influence teacher behavior." Furthermore, we previously reported that states most often have chosen multiple choice assessments over other types because they can be scored inexpensively and their scores can be released prior to the next school year as required by NCLBA.[Footnote 30] That report also notes that state officials and alignment experts said that multiple choice assessments have limited the content of what can be tested, stating that highly complex content is "difficult if not impossible to include with multiple choice items." However, we have revised this paragraph to clarify our point and provide additional information. Concerning the topic of narrowing the curriculum, we agree with the Department of Education that this report should include a fuller description of the data results from the cited Education survey in order to help the reader put the data in an appropriate context. Hence, we have added information to that section of the report. However, one limitation of the survey data we cite is that it covers changes in instructional time for a short time period--from school year 2004-05 to 2006-07. In the its technical comments, the Department refers to its recent report, Title I Implementation: Update on Recent Evaluation Findings for a fuller discussion of this issue. The Title I report, while noting that most elementary teachers reported no change from 2004- 05 to 2006-07 in the amount of instructional time that they spent on various subjects, also provides data over a longer, albeit earlier period time period, from 1987-88 to 2003-04, from the National Center on Education Statistics Schools and Staffing Survey. In analyzing this data, the report states that elementary teachers had increased instructional time on reading and mathematics and decreased the amount of time spent on science and social studies during this period. We have added this information as well. Taken together, we believe these data further reinforce our point that assessments under current accountability systems can have, in addition to positive influences on teaching, some negative ones as well, such as the curriculum changes noted in the report, even if the extent of these changes is not fully known. Education also provided technical comments that we incorporated as appropriate. We are sending copies of this report to the Secretary of Education, relevant congressional committees, and other interested parties. The report also is available at no charge on the GAO Web site at [hyperlink, http://www.gao.gov]. If you or your staff have any questions about this report, please contact me at (202) 512-7215 or ashbyc@gao.gov. Contact points for our Office of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made major contributions to this report are listed in appendix VI. Signed by: Cornelia M. Ashby: Director, Education, Workforce, and Income Security Issues: List of Congressional Committees: The Honorable Tom Harkin: Chairman: The Honorable Thad Cochran: Ranking Member: Subcommittee on Labor, Health and Human Services, Education and Related Agencies: Committee on Appropriations: United States Senate: The Honorable Dave Obey: Chairman: The Honorable Todd Tiahrt: Ranking Member: Subcommittee on Labor, Health and Human Services, Education and Related Agencies: Committee on Appropriations: House of Representatives: [End of section] Appendix I: Scope and Methodology: To address the objectives of this study, we used a variety of methods. To determine the types of instructional practices schools and teachers are using to help students achieve state academic standards and whether those practices differ by school characteristics, we used two recent surveys of principals and teachers. The first survey, a nationally- representative survey from the Department of Education's (Education) National Longitudinal Study of No Child Left Behind (NLS-NCLB) conducted by the RAND Corporation (RAND), asked principals the extent to which their schools were focusing on certain strategies in their voluntary school improvement efforts. Education's State and Local Implementation of the No Child Left Behind Act Volume III-- Accountability Under NCLB: Interim Report included information about the strategies emphasized by principals as a whole, and we obtained from Education the NLS-NCLB database to determine the extent to which principals' responses differed by school characteristic variables. We conducted this analysis on school year 2006-2007 data by controlling for four school characteristic variables: (1) the percentage of a school's students receiving free or reduced price lunch (poverty); (2) the percentage of students who are a racial minority (minority); (3) whether the school is in an urban, urban fringe (suburban), or rural area (school location); and (4) the school's adequate yearly performance (AYP) status. We analyzed data from a second RAND survey, which was a three-state survey sponsored by the National Science Foundation that asked math teachers in California, Georgia, and Pennsylvania how their classroom teaching strategies differed due to a state math test.[Footnote 31] RAND selected these states to represent a range of approaches to standards-based accountability and to provide some geographic and demographic diversity; the survey data is representative only for those three states individually. RAND's report on the three-state survey data included information about how teachers within each of the three states had changed their teaching practices due to a state accountability test.[Footnote 32] RAND provided us with descriptive data tables based on its school year 2005-2006 survey data; we analyzed the data to measure associations between the strategies used and the school characteristic variables.[Footnote 33] We requested tables that showed this information for teachers in all schools, and separately for teachers in different categories of schools (elementary and middle schools) and by the school characteristics of poverty, minority, school location and AYP status. We obtained from RAND standard error information associated with the estimates from the different types of schools and thus were able to test the statistical significance of differences in likelihood between what teachers from different types of schools reported. As part of our analyses for both surveys, we reviewed documentation and performed electronic testing of the data obtained through the surveys. We also conducted several interviews with several researchers responsible for the data collection and analyses and obtained information about the measures they took to ensure data reliability. On the basis of our efforts to determine the reliability of the data, we determined the data from each of these surveys were sufficiently reliable for the purposes of our study. We reviewed existing literature to determine what researchers have found regarding the effect of standards-based accountability systems on instructional practices, and practices that work in raising student achievement. To identify existing studies, we conducted searches of various databases, such as the Education Resources Information Center, Proquest, Dialog EDUCAT, and Education Abstracts. We also asked all of the education researchers that we interviewed to recommend additional studies. From these sources, we identified 251 studies that were relevant to our study objectives about the effect of standards-based accountability systems on instructional practices and instructional practices there are effective in raising student achievement. We selected them according to the following criteria: covered the years 2001 through 2008 and were either experimental or quasi-experimental studies, literature syntheses, or studied multiple sites.[Footnote 34] We selected the studies for our review based on their methodological strength, given the limitations of the methods used, and not necessarily on whether the results could be generalized. We performed our searches from August 2008 to January 2009. To assess the methodological quality of the selected studies, we developed a data collection instrument to obtain information systematically about each study being evaluated and about the features of the evaluation methodology. We based our data collection and assessments on generally accepted social science standards. We examined factors related to the use of comparison and control groups; the appropriateness of sampling and data collection methods; and for syntheses, the process and criteria used to identify studies. A senior social scientist with training and experience in evaluation research and methodology read and coded the methodological discussion for each evaluation. A second senior social scientist reviewed each completed data collection instrument and the relevant documentation to verify the accuracy of every coded item. This review identified 20 selected studies that met GAO's criteria for methodological quality. We supplemented our synthesis by interviewing prominent education researchers identified in frequently cited articles and through discussions with knowledgeable individuals. We also conducted interviews with officials at the U.S. Department of Education, including the Center on Innovation and Improvement, and the Institute on Education Sciences' National Center for Education Evaluation and Regional Assistance, as well as other educational organizations. We also reviewed relevant federal laws and regulations. [End of section] Appendix II: Analyses of the Relationship between School Characteristics and Principals' Focus on School Improvement Strategies: In order to analyze the National Longitudinal Study of No Child Left Behind (NLS-NCLB) principal survey conducted by the RAND Corporation, we analyzed strategies on which principals most often focused, taking into account the percentage of a school's students receiving free or reduced price lunch (poverty), the percentage of students who are a racial minority (minority), whether the school is in an urban, suburban, or rural area (school location), and the school's adequate yearly performance (AYP) status (see table 1).[Footnote 35] Our analyses used "odds ratios," generally defined as the ratio of the odds of an event occurring in one group compared to the odds of it occurring in another group, to express differences in the likelihoods of schools with different characteristics using these strategies. We used odds ratios rather than percentages because they are more appropriate for statistical modeling and multivariate analysis. Odds ratios indicate how much higher (when they are greater than 1.0) or lower (when they are less than 1.0) the odds were that principals would respond that a given strategy was a major or moderate focus. We included a reference category for the school characteristics (low minority, low poverty, and central city) in the top row of table 1, and put comparison groups beneath those reference categories, as indicated by the column heading in the second row (high-minority, high-poverty, or rural schools). As an example, the third cell in the "high-minority schools" column indicates that principals in high-minority schools were 2.65 times more likely to make "implementing new instructional approaches or curricula in reading/language arts/English" a focus of their school improvement efforts. In another example, the odds that principals would "restructure the school day to teach core content areas in greater depth (e.g., establishing a literacy block)" were 2.8 times higher for high-poverty schools than low poverty schools, as seen in the sixth cell under "high-poverty schools." Those cells with an asterisk indicate statistically significant results; that is, we have a high degree of confidence that the differences we see are not just due to chance but show an actual difference in the survey responses. See appendix I for further explanation of our methodology. Table 1: Odds Ratios Indicating the Difference in Likelihood of Principals to Make School Improvement Strategies a Moderate or Major Focus after Controlling for Different Factors: School demographic School Improvement Strategy: Using student achievement data to inform instruction and school improvement; (Compared to low-minority schools): High-minority schools: 1.24; (Compared to low-minority schools): Middle minority schools: 3.01*; (Compared to low-poverty schools): High-poverty schools: 2.51; (Compared to low-poverty schools): Middle poverty schools: 1.34; (Compared to central city schools): Rural schools: 0.46; (Compared to central city schools): Suburban/fringe schools: 0.98. School demographic School Improvement Strategy: Aligning curriculum and instruction with standards and/or assessments; (Compared to low-minority schools): High-minority schools: 1.24; (Compared to low-minority schools): Middle minority schools: 2.09*; (Compared to low-poverty schools): High-poverty schools: 1.81; (Compared to low-poverty schools): Middle poverty schools: 0.92; (Compared to central city schools): Rural schools: 0.58; (Compared to central city schools): Suburban/fringe schools: 0.79. School demographic School Improvement Strategy: Implementing new instructional approaches or curricula in reading/language arts/ English; (Compared to low-minority schools): High-minority schools: 2.65*; (Compared to low-minority schools): Middle minority schools: 1.66; (Compared to low-poverty schools): High-poverty schools: 0.99; (Compared to low-poverty schools): Middle poverty schools: 1.24; (Compared to central city schools): Rural schools: 0.80; (Compared to central city schools): Suburban/fringe schools: 0.97. School demographic School Improvement Strategy: Implementing new instructional approaches or curricula in mathematics; (Compared to low-minority schools): High-minority schools: 1.78; (Compared to low-minority schools): Middle minority schools: 1.79*; (Compared to low-poverty schools): High-poverty schools: 1.68; (Compared to low-poverty schools): Middle poverty schools: 1.39; (Compared to central city schools): Rural schools: 0.56*; (Compared to central city schools): Suburban/fringe schools: 0.85. School demographic School Improvement Strategy: Providing additional instruction to low-achieving students; (Compared to low-minority schools): High-minority schools: 2.39*; (Compared to low-minority schools): Middle minority schools: 3.46*; (Compared to low-poverty schools): High-poverty schools: 1.00; (Compared to low-poverty schools): Middle poverty schools: 0.48*; (Compared to central city schools): Rural schools: 0.31*; (Compared to central city schools): Suburban/fringe schools: 0.83. School demographic School Improvement Strategy: Restructuring the school day to teach core content areas in greater depth (e.g., establishing a literacy block); (Compared to low-minority schools): High-minority schools: 1.85*; (Compared to low-minority schools): Middle minority schools: 1.29; (Compared to low-poverty schools): High-poverty schools: 2.84*; (Compared to low-poverty schools): Middle poverty schools: 1.66*; (Compared to central city schools): Rural schools: 0.55*; (Compared to central city schools): Suburban/fringe schools: 1.18. School demographic School Improvement Strategy: Increasing instructional time for all students (e.g., by lengthening the school day or year, shortening recess); (Compared to low-minority schools): High-minority schools: 1.86*; (Compared to low-minority schools): Middle minority schools: 1.22; (Compared to low-poverty schools): High-poverty schools: 2.48*; (Compared to low-poverty schools): Middle poverty schools: 1.77*; (Compared to central city schools): Rural schools: 0.53; (Compared to central city schools): Suburban/fringe schools: 0.99. School demographic School Improvement Strategy: Providing extended- time instructional programs (e.g., before-school, after-school or weekend instructional programs); (Compared to low-minority schools): High-minority schools: 3.54*; (Compared to low-minority schools): Middle minority schools: 2.11*; (Compared to low-poverty schools): High-poverty schools: 2.51*; (Compared to low-poverty schools): Middle poverty schools: 2.49*; (Compared to central city schools): Rural schools: 0.46*; (Compared to central city schools): Suburban/fringe schools: 1.12. School demographic School Improvement Strategy: Implementing strategies for increasing parents' involvement in their children's education; (Compared to low-minority schools): High-minority schools: 1.86*; (Compared to low-minority schools): Middle minority schools: 2.19*; (Compared to low-poverty schools): High-poverty schools: 2.33*; (Compared to low-poverty schools): Middle poverty schools: 1.33; (Compared to central city schools): Rural schools: 0.76; (Compared to central city schools): Suburban/fringe schools: 0.98. School demographic School Improvement Strategy: Increasing the intensity, focus, and effectiveness of professional development; (Compared to low-minority schools): High-minority schools: 1.61; (Compared to low-minority schools): Middle minority schools: 1.39; (Compared to low-poverty schools): High-poverty schools: 2.38*; (Compared to low-poverty schools): Middle poverty schools: 1.3; (Compared to central city schools): Rural schools: 0.54*; (Compared to central city schools): Suburban/fringe schools: 1.00. * = Statistically significant at the 95% confidence level. Source: GAO analysis of NLS-NCLB data. [End of table] [End of section] Appendix III: List of Education Researchers: Name: Dr. David K. Cohen; Affiliation: John Dewey Collegiate Professor of Education Walter H. Annenberg Professor of Education Policy University of Michigan. Name: Dr. Linda Darling-Hammond; Affiliation: Charles Ducommon Professor of Education Stanford University. Name: Dr. Richard Elmore; Affiliation: Gregory R. Anrig Professor of Educational Leadership Director, Consortium for Policy Research in Education Harvard University. Name: Dr. David Figlio; Affiliation: Institute for Policy Research Northwestern University National Bureau of Economic Research. Name: Dr. William A. Firestone; Affiliation: Director, Center for Educational Policy Analysis; Principal Investigator, New Jersey Math Science Partnership; Professor Rutgers University. Name: Dr. Susan Fuhrman; Affiliation: President, Teachers College Columbia University. Name: Dr. Margaret Goertz; Affiliation: Professor Co-Director, Consortium for Policy Research in Education University of Pennsylvania. Name: Dr. Laura Hamilton; Affiliation: Senior Behavioral/Social Scientist RAND. Name: Dr. Jane Hannaway; Affiliation: Director of Education Policy Urban Institute. Name: Dr. Richard Murnane; Affiliation: Juliana W. and William Foss Thompson Professor of Education and Society Harvard University. Name: Dr. William Sanders; Affiliation: Senior Research Fellow University of North Carolina. Name: Dr. Brian Stecher; Affiliation: Senior Social Scientist RAND. Source: GAO. [End of table] [End of section] Appendix IV: Studies Meeting GAO's Criteria for Methodological Quality: Title: Accountability and Teaching Practices: School Level Actions and Teacher Responses; Author: Laura. S Hamilton; Brian M Stecher; Jennifer Linn Russell; Julie A. Marsh; Jeremy Miles; Source: "Strong States, Weak Schools: The Benefits and Dilemmas of Centralized Accountability"; Research in Sociology of Education, vol. 16, 2008; Method: Case studies of three states; representative surveys for these states. Title: Catching Up Impact of the Talent Development Ninth Grade Instructional Interventions in Reading and Mathematics in High-Poverty High Schools; Author: Robert Belfanz; Nettie Legters; Will Jordan; Source: Report 69 April 2004 The Johns Hopkins University Center for Research on the Education of Students Placed at Risk; Method: Quasi-experimental design with matched groups; multiple regressions used with data. Limitations: Two school districts (around Baltimore); small percentage of all those enrolled in the 9th grade. Title: Differentiated Curriculum Enhancement in Inclusive Middle School Science: Effects on Classroom and High-Stakes Tests; Author: Margo A. Mastropieri; Thomas E. Scruggs; Jennifer J. Norland; Sheri Berkeley; Kimberly McDuffie; Elizabeth Halloran Tornquist; Nicole Connors; Source: The Journal of Special Education vol. 40, no. 3. 2006, 130-137; Method: Quasi-experimental design; 13 classes matched by teacher, and randomly assigned to treatment or control group. Limitations: some external validity issues. Title: Effective Programs in Elementary Mathematics: A Best-Evidence Synthesis; Author: Robert E. Slavin; Cynthia Lake; Source: Review of Educational Research. Washington: September 2008. vol. 78, issue 3. 427; Method: Literature review using a best-evidence synthesis (related to a meta-analysis). Title: Feeling the Florida Heat? How Low-Performing Schools Respond to Voucher and Accountability Pressure; Author: Cecilia Elena Rouse; Jane Hannaway; Dan Goldhaber; David Figlio; Source: Calder/Urban Institute National Center for Analysis of Longitudinal Data in Education Research Working Paper November 2007; Method: Administrative data used to develop comparison groups of schools; regression discontinuity design; results apply to Florida schools only. Title: Formulating Secondary-Level Reading Interventions; Author: Debra M. Kamps; Charles R. Greenwood; Source: Journal of Learning Disabilities, vol. 38, no. 6. November/December 200, 500-509; Method: Quasi-experimental; random assignment of schools, but not students; Limitations: cannot be generalized beyond the 8 schools involved in the study. Title: Helping At-Risk Students Meet Standards A Synthesis of Evidence- Based Classroom Practices; Author: Zoe Barley; Patricia A. Lauer; Sheila A. Arens; Helen S. Apthorp; Kelly S. Englert; David Snow; Motoko Akiba; Source: Regional Education Laboratory; Office of Educational Research and Improvement; U.S. Department of Education; Mid-continent Research for Education and Learning October 2002 corrected 12/02; Method: Literature review; in some cases a meta-analysis was conducted; effect sizes were computed for meta-analysis when available; some studies were outside the time frames of our search criteria. Title: High Poverty Schools and the Distribution of Teachers and Principals; Author: Charles Clotfelter; Helen F. Ladd; Jacob Vigdor; Justin Wheeler; Source: Sanford Working Paper Series SAN06-08 December 2006; Method: Time series analysis using administrative data for all schools in North Carolina. Limitation: applies to North Carolina only. Title: High Stakes Testing and Curricular Control: A Qualitative Metasynthesis; Author: Wayne Au; Source: Educational Researcher; vol. 36, no. 5, June/Jul 2007; 258-267; Method: Meta-synthesis of qualitative studies; Limitations: Results for Chicago only; some coding issues. Title: Instructional Policy and Classroom Performance: The Mathematics Reform in CA; Author: David K. Cohen; Heather C. Hill; Source: Teachers College Record, vol. 102, no. 2. February 2000, 294- 343; Method: Regression analysis of data from teacher surveys and administrative data. Limitations: results based on a 1994 survey; response rate was 61 percent. Title: Instructional Time in Elementary Schools A Closer Look at Changes for Specific Subjects; Author: Center on Education Policy; Source: From the Capital to the Classroom: Year of the No Child Left Behind Act; Center on Education Policy February 2008; Method: Survey of school districts and states, qualitative interviews; Limitation: high non-response rate from school districts in large urban areas. Title: Standards in Classroom Practice: Research Synthesis; Author: Helen S. Apthorp; Ceri B. Dean; Judy E. Florian; Patricia A. Lauder; Robert Reichardt; Nancy M. Sanders; Ravay Snow-Renner; Source: Regional Education Laboratory; Office of Educational Research and Improvement; U.S. Department of Education; Mid-continent Research for Education and Learning October 31, 2001; Method: Literature review; no meta-analysis conducted; some studies outside our time frame. Title: Standards-Based Reform in Practice: Evidence on State Policy and Classroom Instruction from the NAEP State Assessments; Author: Christopher B. Swanson; David Lee Stevenson; Source: Educational Evaluation and Policy Analysis, vol. 24, no. 1. Spring 2002, 1-27; Method: Hierarchical linear modeling on survey data from the National Assessment of Educational Progress (NAEP); limitation is that only 30 of the original 40 states are included, with some of the largest of the states missing. Title: Studying Large-Scale Reforms of Instructional Practice: An Example from Mathematics and Science; Author: Laura S. Hamilton; Daniel F. McCaffrey; Brian Stecher; Stephen P. Klein; Abby Robyn; Delia Bugliari; Source: Educational Evaluation and Policy Analysis, vol. 25, no. 1. Spring 2003, 1-29; Method: Regression analysis; Limited to 11 sites; results small and positive, but not statistically significant. Title: Supporting Literacy Across the Sunshine State: A Study of Florida Middle School Reading Coaches; Author: Julie A. Marsh; Jennifer Sloan McCombs; J.R. Lockwood; Francisco Martorell; Daniel Gershwin; Scott Naftel; Vi-Nhuan Le; Molly Shea; Heather Barney; Al Crego; Source: RAND Corporation 2008; Method: Case study of Florida; longitudinal data analysis of data from 1997-1998 to 2006-2007 based on a survey of teachers, principals, and students in 8 middle schools. Title: Teaching Methods for Secondary Algebra: A Meta-Analysis of Findings; Author: Matthew Haas; Source: National Association of Secondary School Principals. NASSP Bulletin, March 2005, 89, 642; Research Library 24; Method: Meta-analysis of 35 studies. Title: Test Preparation in New Jersey: inquiry-oriented and didactic responses; Author: William A. Firestone; Lora Monfils; Roberta Y. Schorr; Source: Assessment in Education: Principles, Policy & Practice, vol. 11, no.1. March 2004, 67-88; Method: Survey, exploratory factor analysis, and hierarchical linear modeling time series; results limited to New Jersey. Title: The Influence of Standards on K-12 Teaching and Student Learning: A Research Synthesis; Author: Patricia A. Lauer; David Snow; Mya Martin-Glenn; Rebecca J. Van Buhler; Kristen Stoutemyer; Ravay Snow-Renner; Source: Regional Education Laboratory, August 19, 2005; Method: Literature review; no meta-analysis; both quantitative and qualitative studies used; comprehensive selection process. Title: The New Accountability, Student Failure, and Teachers' Work in Urban High Schools; Author: Dorothea Anagnostopoulous; Source: Educational Policy, vol. 17, no. 3. July 2003, 291-316; Method: Case study of two high schools; findings are suggestive. Title: Value-Added Assessment in Practice: Lessons from Pennsylvania Value-Added Assessment System Pilot Project; Author: Daniel F. McCaffrey; Laura S. Hamilton; Source: RAND Corporation 2007; Method: Quasi-experimental design for 93 non-random study districts in Pennsylvania; not generalizable to the nation or the state. Source: GAO analysis. [End of table] [End of section] Appendix V: Comments from the Department of Education: United States Department Of Education: Office Of Planning, Evaluation And Policy Development: 400 Maryland Ave, SW: Washington, DC 20202: October 2, 2009: Ms. Cornelia M. Ashby: Director: Education, Workforce, and Income Security Issues: U.S. Government Accountability Office: Washington, DC 20548: Dear Ms. Ashby: Thank you for the opportunity to comment on the draft GAO report, Student Achievement: Schools Use Multiple Strategies to Help Students Meet Academic Standards, Especially Schools with Higher Proportions of Low Income and Minority Students. GAO's report asks important questions about the effects of standards- based accountability on instructional practices and the effectiveness of specific instructional practices in improving student achievement, and seeks to answer these questions through a literature review and interviews with prominent education researchers. The report also examines data on the types of instructional practices that schools and teachers arc using to help students achieve to state academic standards, in part based on surveys conducted for the Department's National Longitudinal Study of No Child Left Behind. While the report addresses important policy questions, there are some issues pertaining to the study's approach that we recommend be taken into consideration. First, the draft report does not clearly explain how GAO selected the 20 studies included in its literature review or the methods used in the studies that were selected. Moreover, in discussing specific findings from the literature review, the report frequently does not indicate which studies are being relied on as evidence for each finding. Second, the report mixes findings that may be based on rigorous research with findings that appear to be based on conjecture and on what "some researchers believe," and does not always present a complete and balanced summary of the relevant research. For example, the report states that "difficulties in aligning practice with standards were attributed, in part, to current accountability requirements," but appears to rely only on expert opinion for this causal conclusion. Similarly, the report states that "a few researchers as well as some of the literature we reviewed report some unintended negative consequences on instruction as a result of assessment practices," including the reported consequences of "multiple choice tests that do not encourage more challenging teacher practices" and "instructional practices that narrow the curriculum." These statements may accurately report the opinions of the individuals interviewed, but the report provides weak empirical evidence to support these conclusions and does not include all of the available evidence. With respect to the assertion that the assessment provisions in the Elementary and Secondary Education Act, as amended by the No Child Left Behind Act (NCLB), have resulted in multiple choice tests that do not encourage more challenging teaching practices, the report does not present any data on the extent to which test formats have changed or on the relationship between test format and teaching practices. Instead the report notes that some researchers believe that states are increasingly using multiple-choice testing formats, and hypothesizes that teachers "may be influenced" to change their teaching approaches because of the tests. Any conclusions about what the "research shows" should he supported by specific references to rigorous research that used appropriate methods for measuring impacts. In the discussion of whether there has been narrowing of the curriculum, the report notes that a Department survey found that 18 to 22 percent of elementary teachers reported increasing instructional time for mathematics and reading, respectively, and concludes that this is occurring "at the expense of other non-tested subjects." However, the report does not mention the finding from the same survey that most elementary teachers reported no change from 2004-05 to 2006-07 in the amount of instructional time that they spent on other subjects. The report also notes that some research has raised concern that teachers may be restricting the breadth of content covered within a particular subject, but does not acknowledge the converse concern, based on research conducted for the Third International Math and Science Study (TIMSS), that curricula in American schools may be "a mile wide and an inch deep" and thus some refocusing of curricula may be beneficial. The Department recognizes that improvements in assessment and accountability systems could help enable schools to strengthen instructional practices and improve student achievement. As one step toward that goal, the American Recovery and Reinvestment Act included $250 million for Statewide Data Systems to help ensure that states and school districts have the robust data systems they need to provide information on individual student outcomes that educators and policymakers can use to drive educational improvement. More research is needed to better understand what instructional practices and policy changes could be most effective in closing achievement gaps and improving educational outcomes. Attached are technical comments provided by Department staff on the text of the report. If you have any questions, we would he glad to discuss our comments with your research team. Sincerely, Signed by: Alan Ginsburg: Director: Policy and Program Studies Service: Enclosure: [End of section] Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: Cornelia M. Ashby (202) 512-7215 or ashbyc@gao.gov. Staff Acknowledgments: Janet Mascia (Assistant Director), Bryon Gordon (Assistant Director), and Andrew Nelson (Analyst-in-Charge) managed all aspects of the assignment. Linda Stokes and Caitlin Tobin made significant contributions to this report in all aspects of the work. Kate van Gelder contributed to writing this report, and Ashley McCall contributed to research for the report. Luann Moy, Justin Fisher, Cathy Hurley, Douglas Sloane, and John Smale Jr. provided key technical support, and Doreen Feldman and Sheila R. McCoy provided legal support. Mimi Nguyen developed the graphics for the report. [End of section] Footnotes: [1] Pub. L. No. 107-110. [2] We use the phrase "instructional practices" to include tools for improving classroom teaching practices, such as providing additional professional development. [3] State and Local Implementation of the No Child Left Behind Act Volume III--Accountability under NCLB: Interim Report. A report from the National Longitudinal Study of No Child Left Behind (NLS-NCLB) and the Study of State Implementation of Accountability and Teacher Quality under No Child Left Behind (SSI-NCLB) Kerstin Carlson Le Floch, AIR, Felipe Martinez, RAND, Jennifer O'Day, AIR, Brian Stecher, RAND, James Taylor, AIR, Andrea Cook, AIR. Prepared for: U.S. Department of Education Office of Planning, Evaluation and Policy Development Policy and Program Studies Service (2007). [4] Laura S. Hamilton, Brian M. Stecher, Julie A. Marsh, Jennifer Sloan McCombs, Abby Robyn, Jennifer Lin Russell, Scott Naftel, and Heather Barney. "Standards-Based Accountability under No Child Left Behind: Experiences of Teachers and Administrators in Three States." Sponsored by the National Science Foundation. RAND 2007. The survey also asked about reported changes in strategies for science instruction as a result of the state science test, but we are only reporting on math instruction. [5] Of the 20 studies we used that met our criteria for methodological quality, we relied heavily on two literature syntheses conducted by the Department of Education because of the large number of studies they included and the breadth of the topics they covered. For a list of these and the other studies meeting our criteria for methodological quality, see appendix IV. Additionally, a few other studies are cited in footnotes throughout the report but not included in the list of studies that we formally reviewed. Those cited in the footnotes were used because they provided more details or supplementary information about points that the experts made during our interviews. [6] For a list of knowledgeable individuals with whom we spoke, see appendix III. [7] Pub. L. No. 89-10. [8] Pub. L. No. 103-382. [9] Assessments in science, which were first required under NCLBA in school year 2007-2008, are required at least once in grades 3 to 5, grades 6 to 9, and grades 10 to 12. High school students are required only to be assessed once in math and reading or language arts. In addition to annual assessments, high schools must include students' graduation rate, and elementary and middle schools must include one other academic indicator determined by the state to assess whether they made AYP. [10] GAO, No Child Left Behind Act: Improvements Needed in Education's Process for Tracking States' Implementation of Key Provisions, [hyperlink, http://www.gao.gov/products/GAO-04-734] (Washington, D.C.: Sept. 30, 2004). [11] For more information on teacher quality, see GAO, Teacher Quality: Sustained Coordination among Key Federal Education Programs Could Enhance State Efforts to Improve Teacher Quality, [hyperlink, http://www.gao.gov/products/GAO-09-593] (Washington, D.C.: July 2009). [12] For purposes of this report, we use the term "school improvement" to refer to the voluntary strategies used by school administrators and teachers to address various challenges within a school. By way of contrast, under NCLBA, schools that are identified for "school improvement" are those that have failed to make AYP for 2 or more consecutive years. These schools must implement certain activities identified in NCLBA that are meant to improve student academic achievement. [13] Education classified schools as having "high--75 percent or more," "moderate--35 to less than 75," or "low--35 percent or less" percentages of low-income students using the number of students at the school that were eligible for the free and reduced-price lunch program. Schools were classified as having "high--75 percent or more," "moderate--25 to less than 75," or "low--25 percent or less" percentages of minority students, based on the school population that principals reported to be American Indian/Alaskan Native, Asian, Black or African-American, Hispanic or Latino, and Native Hawaiian or other Pacific Islander. Schools also were classified as central city (urban), urban fringe/large town (suburban), or small/fringe town (rural). [14] See appendix II for additional information about how principals' responses differed across school characteristics. [15] Core content areas include those subjects for which testing is required under NCLBA--specifically, reading, math, and science. [16] For the last three of these five strategies and one other- providing additional instruction to low-achieving students--there were also significant differences between moderate-poverty and low-poverty schools. [17] See appendix II for a table that indicates which six strategies differed by school minority level. [18] Urban fringe or large town schools were no different from the central city schools with respect to making these strategies a major or moderate focus. In the 2003-2004 school year, about 30 percent of all U.S. elementary and secondary public schools were located in rural areas and approximately 20 percent of public school students were enrolled in rural schools. See S. Provasnik, A. KewalRamani, M. M. Coleman, L. Gilbertson, W. Herring, and Q. Xie, Status of Education in Rural America (NCES 2007-040). National Center for Education Statistics, Institute of Education Sciences, U.S. Department of Education (Washington, D.C.: 2007). See appendix II for a table that indicates which five strategies differed by school geographic type. [19] When we compared moderate-poverty schools to high-poverty and low- poverty schools, we saw fewer statistically significant differences than in our high-poverty and low-poverty school comparison. [20] For the three state data, we conducted a simple analysis that did not control for multiple factors, since we had access only to RAND's bi- variate analyses of the data rather than the data itself. Because of this, we could not perform a multivariate analysis, which would allow us to control for other factors. [21] The National Governors' Association and the Council of Chief State School Officers are coordinating a committee of experts to develop common academic standards for math and language arts skills. As of June 2009, 46 states had signed onto this effort to adopt the common standards once they were completed. [22] W. Firestone, R. Schorr, and L. Monfils, editors. Ambiguity of Teaching to the Test: Standards, Assessments, and Educational Reform, 160-161 (2004). [23] Helen S. Apthorp, et al., "Standards in Classroom Practice Research Synthesis," Mid-Continent Research for Education and Learning (October 2001). [24] NCLBA added to the assessment requirements included in IASA. For example, NCLBA requires states to implement annual assessments for all students in every grade for grades 3-8 in reading and math; IASA required assessments at least once in each of three grade spans: 3-5, 6- 9, and 10-12. Additionally, unlike IASA, NCLBA sets a uniform timeline for when all students must meet state proficiency targets. [25] P. A. Lauer, D. Snow, M. Martin-Glenn, R.J.Van Buhler, K. Stoutemyer, R. Snow-Renner, The Influence of Standards on K-12 Teaching and Student Learning: A Research Synthesis, Regional Education Laboratory, August 19, 2005, p. 91. [26] GAO, No Child Left Behind Act: Enhancements in the Department of Education's Review Process Could Improve State Academic Assessments, [hyperlink, http://www.gao.gov/products/GAO-09-911], (Washington, D.C.: September 2009). [27] For example, according to data from Education's national survey, about 18 percent of elementary school teachers reported that instruction time for math increased from school years 2004-2005 to 2006- 2007, and about 22 percent of elementary school teachers reported that instruction time for reading/language arts increased over the same period. However, approximately three-quarters of teachers reported no change in instructional time in these two subjects. GAO, Access to Arts Education: Inclusion of Additional Questions in Education's Planned Research Would Help Explain Why Instruction Time Has Decreased for Some Students, [hyperlink, http://www.gao.gov/products/GAO-09-286] (Washington, D.C.: February 2009). In addition, a report by the Department of Education states that from 1987-1988 to 2003-2004, teacher survey results from the Schools and Staffing Survey conducted by the National Center for Education Statistics indicate that elementary teachers had increased instructional time on reading and mathematics and decreased the amount of time spent on science and social studies during this period. See U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy and Program Studies Service, Title I Implementation--Update on Recent Evaluation Findings (Washington, D.C.: 2009). [28] The final report of the National Mathematics Advisory Panel takes a slightly different position regarding this practice stating that "All- encompassing recommendations that instruction should be entirely 'student centered' or 'teacher directed' are not supported by research ...High-quality research does not support the exclusive use of either approach." National Mathematics Advisory Panel. Foundations for Success: The Final Report of the National Mathematics Advisory Panel, U.S. Department of Education (Washington, D.C.: 2008). [29] L. Darling-Hammond, R. Wei, A. Andree, N. Richardson, and S. Orphanos, Professional Learning in the Learning Profession: A Status Report on Teacher Development in the United States and Abroad, Technical Report (National Staff Development Council and The School Redesign Network at Stanford University: February 2009) 18 and 22. [30] GAO, No Child Left Behind Act: Enhancements in the Department of Education's Review Process Could Improve State Academic Assessments, [hyperlink, http://www.gao.gov/products/GAO-09-911] (Washington, D.C.: September 2009). [31] Several education experts we spoke to said the list of practices was fairly complete, but one expert noted that professional development is also an important instructional practice. [32] Laura S. Hamilton, Brian M. Stecher, Julie A. Marsh, Jennifer Sloan McCombs, Abby Robyn, Jennifer Lin Russell, Scott Naftel, and Heather Barney, "Standards-Based Accountability under No Child Left Behind: Experiences of Teachers and Administrators in Three States" (Sponsored by the National Science Foundation. RAND 2007). [33] Scott Naftel, Laura S. Hamilton, and Brian M. Stecher, "Working Paper Supplemental Analyses of ISBA Survey Responses" (WR-628-EDU. RAND. November 2008). [34] Some research, including the syntheses that we reviewed, included some studies outside these date parameters. Additionally, the syntheses used to support some of the findings were not meta-analyses but literature reviews, although both qualitative and quantitative studies were included in the syntheses. [35] Table 1 does not include AYP status, because we found that the demographic characteristics of poverty and minority explained the patterns of principals' responses more fully than AYP status. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO’s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO’s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548: