This is the accessible text file for GAO report number GAO-04-38 entitled 'Results-Oriented Government: GPRA Has Established a Solid Foundation for Achieving Greater Results' which was released on March 10, 2004. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: March 2004: RESULTS-ORIENTED GOVERNMENT: GPRA Has Established a Solid Foundation for Achieving Greater Results: GAO-04-38: GAO Highlights: Highlights of GAO-04-38, a report to congressional requesters Why GAO Did This Study: Now that the Government Performance and Results Act (GPRA) has been in effect for 10 years, GAO was asked to address (1) the effect of GPRA in creating a governmentwide focus on results and the government’s ability to deliver results to the American public, (2) the challenges agencies face in measuring performance and using performance information in management decisions, and (3) how the federal government can continue to shift toward a more results-oriented focus. What GAO Found: GPRA’s requirements have established a solid foundation of results- oriented performance planning, measurement, and reporting in the federal government. Federal managers surveyed by GAO reported having significantly more of the types of performance measures called for by GPRA (see figure below). GPRA has also begun to facilitate the linking of resources to results, although much remains to be done in this area to increase the use of performance information to make decisions about resources. We also found agency strategic and annual performance plans and reports we reviewed have improved over initial efforts. Although a foundation has been established, numerous significant challenges to GPRA implementation still exist. Inconsistent top leadership commitment to achieving results within agencies and OMB can hinder the development of results-oriented cultures in agencies. Furthermore, in certain areas, federal managers continue to have difficulty setting outcome-oriented goals, collecting useful data on results, and linking institutional, program, unit, and individual performance measurement and reward systems. Finally, there is an inadequate focus on addressing issues that cut across federal agencies. OMB, as the focal point for management in the federal government, is responsible for overall leadership and direction in addressing these challenges. OMB has clearly placed greater emphasis on management issues during the past several years. However, it has showed less commitment to GPRA implementation in its guidance to agencies and in using the governmentwide performance plan requirement of GPRA to develop an integrated approach to crosscutting issues. In our view, governmentwide strategic planning could better facilitate the integration of federal activities to achieve national goals. What GAO Recommends: GAO recommends that the Office of Management and Budget (OMB) improve its guidance and oversight of GPRA implementation, as well as develop a governmentwide performance plan. GAO also believes Congress should consider amending GPRA to require that (1) agencies update their strategic plans at least once every four years, consult with congressional stakeholders at least once every new Congress, and make interim updates to strategic and performance plans as appropriate; and (2) the President develop a governmentwide strategic plan. OMB generally agreed with our recommendations, but stated that the President’s Budget can serve as both a governmentwide strategic and annual plan. However, we believe the budget provides neither a long- term nor an integrated perspective on the federal government’s performance. www.gao.gov/cgi-bin/getrpt?GAO-04-38. To view the full product, including the scope and methodology, click on the link above. For more information, contact Patricia A. Dalton at (202) 512-6806 or daltonp@gao.gov. [End of section] Contents: Transmittal Letter: Executive Summary: Purpose: Background: Results in Brief: Principal Findings: Recommendations for Executive Action: Matters for Congressional Consideration: Agency Comments: Chapter 1: Introduction: Impact of Emerging Trends and Fiscal Challenges: GPRA Background: Scope and Methodology: Chapter 2: GPRA Established the Foundation for a More Results-Oriented Federal Government: GPRA Statutory Requirements Laid a Foundation for Agencywide Results- Oriented Management: Chapter 3: Agencies Have Addressed Many Critical Performance Planning and Reporting Challenges, but Weaknesses Persist: Quality of Selected Strategic Plans Reflects Improvements over Initial Drafts: Fiscal Year 2004 Annual Performance Plans Addressed Some Weaknesses of Earlier Plans, but Still Have Room for Significant Improvement: Strengths and Weaknesses of Selected Agencies' Fiscal Year 2002 Annual Performance Reports: Chapter 4: Challenges to GPRA Implementation Persist: Top Leadership Does Not Consistently Show Commitment to Achieving Results: Managers Report Mixed Results in Use of Performance Information: Managers Continue to Confront a Range of Human Capital Management Challenges: Persistent Challenges in Setting Outcome-Oriented Goals, Measuring Performance, and Collecting Useful Data: Crosscutting Issues Hinder Successful GPRA Implementation: Managers View Congress' Use of Performance Information as Limited: Chapter 5: Conclusions and Recommendations: Agenda for Achieving a Sustainable, Governmentwide Focus on Results: Recommendations for Executive Action: Matters for Congressional Consideration: Agency Comments: Appendixes: Appendix I: Objectives, Scope, and Methodology: Methodology for Governmentwide Survey: Methodology for Focus Groups: Methodology for Interviews with Political Appointees: Methodology for Selecting Agencies to Review for Changes in the Quality of Their Strategic Plans, Annual Performance Plans, and Annual Performance Reports: Appendix II: Focus Group Participants Agreed GPRA Provides a Framework for Federal Agencies to Become More Results Oriented: GPRA Accomplishments: Views on Delivering Results to the American Public Were Mixed: Alternate Views on GPRA's Effect: Challenges in Implementing and Overseeing GPRA Activities: Suggestions to Address GPRA Challenges: Appendix III: Observations on Agencies' Strategic Plans: Required Elements of Agency Strategic Plans: Observations on Changes in the Quality of Education's Strategic Plan: Observations on Changes in the Quality of DOE's Strategic Plan: Observations on Changes in the Quality of HUD's Strategic Plan: Observations on Changes in the Quality of SBA's Strategic Plan: Observations on Changes in the Quality of SSA's Strategic Plan: Observations on Changes in the Quality of DOT's Strategic Plan: Appendix IV: Observations on Agencies' Annual Performance Plans: Key Elements of Information for Annual Performance Plans: Observations on Changes in the Quality of Education's Annual Performance Plan: Observations on Changes in the Quality of DOE's Annual Performance Plan: Observations on Changes in the Quality of HUD's Annual Performance Plan: Observations on Changes in the Quality of SBA's Annual Performance Plan: Observations on Changes in the Quality of SSA's Annual Performance Plan: Observations on Changes in the Quality of DOT's Annual Performance Plan: Appendix V: Observations on Agencies' Annual Performance and Accountability Reports: Observations on the Quality of Education's Fiscal Year 2002 Performance and Accountability Report: Observations on the Quality of DOE's Fiscal Year 2002 Annual Performance and Accountability Report: Observations on the Quality of HUD's Fiscal Year 2002 Annual Performance and Accountability Report: Observations on the Quality of SBA's Fiscal Year 2002 Annual Performance and Accountability Report: Observations on the Quality of SSA's Fiscal Year 2002 Annual Performance and Accountability Report: Observations on the Quality of DOT's Fiscal Year 2002 Annual Performance and Accountability Report: Appendix VI: GAO Federal Managers' Survey Data: Appendix VII: Agencies Subject to the Chief Financial Officers Act: Appendix VIII: Comments from the Office of Management and Budget: Appendix IX: Comments from the Department of Energy: GAO Comments: Appendix X: Comments from the Department of Housing and Urban Development: GAO Comments: Appendix XI: Comments from the Social Security Administration: GAO Comments: Appendix XII: GAO Contact and Staff Acknowledgments: GAO Contact: Acknowledgments: Related GAO Products: GPRA/Managing for Results: Strategic Human Capital Management: Linking Resources to Results: Measuring Performance: Data Credibility: Using Performance Information: Tables: Table 1: Agencies' Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 2: Characterizations of Agencies' Fiscal Year 1999 and 2004 Annual Performance Plans: Table 3: Characterizations of Agencies' 2002 Annual Performance Reports: Table 4: Summary of Characteristics of Agencies Selected for Review of Strategic Plans, Annual Performance Plans, and Annual Performance Reports: Table 5: Agencies' Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 6: Education's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 7: DOE's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 8: HUD's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 9: SBA's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 10: SSA's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 11: DOT's Progress in Addressing Required Elements of Strategic Planning under GPRA: Table 12: Characterizations of Agencies' Annual Performance Plans: Table 13: Characterizations of Agencies' Fiscal Year 2002 Annual Performance and Accountability Reports: Figures: Figure 1: Composition of Spending as a Share of GDP Assuming Discretionary Spending Grows with GDP after 2003 and All Expiring Tax Provisions Are Extended: Figure 2: Percentage of Federal Managers Who Reported That There Were Performance Measures for the Programs with Which They Were Involved: Figure 3: Percentage of Federal Managers Who Reported Having Specific Types of Performance Measures to a Great or Very Great Extent: Figure 4: Percentage of Federal Managers Who Reported Their Awareness of GPRA: Figure 5: Percentage of Federal Managers Who Reported Hindrances to Measuring Performance or Using the Performance Information to a Great or Very Great Extent: Figure 6: Percentage of Federal Managers and SES Managers Who Reported That OMB Paid Attention to Their Agency's Efforts under GPRA to a Great or Very Great Extent: Figure 7: Percentage of Federal Managers Who Reported They Considered Strategic Goals to a Great or Very Great Extent When Allocating Resources: Figure 8: Percentage of Federal Managers Who Reported They Considered Performance Information to a Great or Very Great Extent When Allocating Resources: Figure 9: Percentage of Federal Managers Who Reported That Funding Decisions Were Based on Results or Outcome-Oriented Performance Information to a Great or Very Great Extent: Figure 10: Percentage of Federal Managers Who Reported to a Great or Very Great Extent Their Top Leadership Has a Strong Commitment to Achieving Results: Figure 11: Percentage of SES and Non-SES Managers Who Reported to a Great or Very Great Extent Their Agency Top Leadership Demonstrated Strong Commitment to Achieving Results: Figure 12: Percentage of Federal Managers Who Reported Using Information Obtained from Performance Measurement to a Great or Very Great Extent for Various Management Activities: Figure 13: Percentage of Federal Managers Responding "Yes" about Being Involved in the Following Activities: Figure 14: Percentage of Federal Managers Reporting to a Great or Very Great Extent That Managers/Supervisors at Their Levels Had the Decision-Making Authority They Needed to Help the Agency Accomplish Its Strategic Goals: Figure 15: Percentage of Federal Managers, SES, and Non-SES in 2003 Reporting to a Great or Very Great Extent That They Were Held Accountable for the Accomplishment of Agency Strategic Goals: Figure 16: Percentage of Federal Managers in Each Survey Year Who Reported That during the Past 3 Years Their Agencies Provided, Arranged, or Paid for Training That Would Help Them Accomplish Specific Tasks: Figure 17: Percentage of Federal Managers Who Reported to a Great or Very Great Extent That Employees in Their Agencies Received Positive Recognition for Helping Their Agencies Accomplish Their Strategic Goals: Figure 18: Percentage of Federal Managers Reporting to a Great or Very Great Extent That a Lack of Ongoing Congressional Commitment or Support for Using Performance Information in Making Program/Funding Decisions Is a Hindrance: Figure 19: Summary of Education's Performance Indicators for Fiscal Year 2002: Figure 20: Inputs: Allocating Funds for Education's Objective to Ensure That All Students Read on Grade Level by the Third Grade: Figure 21: Summary of DOE's Performance Indicators for Fiscal Year 2002: Figure 22: Summary of HUD's Performance Indicators for Fiscal Year 2002: Figure 23: Summary of SBA's Performance Goals for Fiscal Year 2002: Figure 24: Summary of SSA's Performance Goals for Fiscal Year 2002: Figure 25: Summary of DOT's Performance Indicators for Fiscal Year 2002: Abbreviations: AP: advanced placement: CDBG: Community Development Block Grants: CFO: Chief Financial Officer: CRS: Congressional Research Service: DOE: Department of Energy: DOT: Department of Transportation: EPA: Environmental Protection Agency: FAA: Federal Aviation Administration: FSA: Federal Student Assistance: FTE: full-time employee: GM: general management: GPRA: Government Performance and Results Act of 1993: GS: general schedule: HHS: Department of Health and Human Services: HOME: Home Investment Partnership Program: HUD: Department of Housing and Urban Development: ICH: Interagency Council on the Homeless: IG: Inspector General: IT: information technology: IRS: Internal Revenue Service: JARC: Job Access and Reverse Commute: NAEP: National Assessment for Educational Progress: NASA: National Aeronautics and Space Administration: OASI: Old Age and Survivors Insurance: OMB: Office of Management and Budget: OPM: Office of Personnel Management: PART: Program Assessment Rating Tool: PMA: President’s Management Agenda: SBA: Small Business Administration: SBDC: Small Business Development Centers: SES: Senior Executive Service: SSA: Social Security Administration: SSI: Supplemental Security Income: SSN: Social Security number: VA: Department of Veterans Affairs: Transmittal Letter March 10, 2004: Congressional Requesters: As you requested, we have assessed the effectiveness of the Government Performance and Results Act (GPRA), in light of its 10-year anniversary in 2003. Our review focused on GPRA's accomplishments, challenges to its continued implementation, and an agenda for achieving a sustainable, governmentwide focus on results. Upon issuance, we will send copies to the Director of the Office of Management and Budget and executive branch agencies (see appendix VII for a list). We will also make copies available to others upon request. In addition, this report will be available at no charge on the GAO web site at [Hyperlink, http://www.gao.gov]. If you have any questions concerning this report, please contact Patricia A. Dalton at (202) 512-6806 or daltonp@gao.gov. The major contributors to this report are listed in appendix XII. Signed by: David M. Walker: Comptroller General of the United States: List of Requesters: The Honorable Susan M. Collins: Chairman: The Honorable Joe Lieberman: Ranking Minority Member: Committee on Governmental Affairs: United States Senate: The Honorable George V. Voinovich: Chairman: The Honorable Richard Durbin: Ranking Minority Member: Subcommittee on Oversight of Government Management, the Federal Workforce, and the District of Columbia: Committee on Governmental Affairs: United States Senate: The Honorable Peter G. Fitzgerald: Chairman: The Honorable Daniel K. Akaka: Ranking Minority Member: Subcommittee on Financial Management, the Budget, and International Security: Committee on Governmental Affairs: United States Senate: The Honorable Tom Davis: Chairman: The Honorable Henry A. Waxman: Ranking Minority Member: Committee on Government Reform: House of Representatives: The Honorable Todd Russel Platts: Chairman: The Honorable Edolphus Towns: Ranking Minority Member: Subcommittee on Government Efficiency and Financial Management: Committee on Government Reform: House of Representatives: [End of section] Executive Summary: Purpose: From defending the homeland against terrorists, to preventing the spread of infectious diseases, to providing a reliable stream of social security income to retirees and supporting the transition from welfare to work, the federal government provides funding and services to the American public that can affect their lives in critical ways every day. However, the federal government is in a period of profound transition and faces an array of challenges and opportunities to enhance performance, ensure accountability, and position the nation for the future. A number of overarching trends, such as diffuse security threats and homeland security needs, increasing global interdependency, the shift to knowledge-based economies, and the looming fiscal challenges facing our nation drive the need to reconsider the proper role for the federal government in the 21st century, how the government should do business (including how it should be structured), and in some instances, who should do the government's business. Without effective short-and long-term planning, which takes into account the changing environment and needs of the American public and the challenges they face and establishes goals to be achieved, federal agencies risk delivering programs and services that may or may not meet society's most critical needs. At a cost to taxpayers of over $2 trillion annually, the federal government should be able to demonstrate to the American public that it can anticipate emerging issues, develop sound strategies and plans to address them, and be accountable for the results that have been achieved. Concerned that the federal government was more focused on program activities and processes than the results to be achieved, Congress passed the Government Performance and Results Act of 1993 (GPRA).[Footnote 1] The act required federal agencies to develop strategic plans with long-term, outcome-oriented goals and objectives, annual goals linked to achieving the long-term goals, and annual reports on the results achieved. Now that GPRA has been in effect for 10 years, you asked us to assess the effectiveness of GPRA in creating a focus on results in the federal government. Specifically, this report discusses (1) the effect of GPRA over the last 10 years in creating a governmentwide focus on results and the government's ability to deliver results to the American public, including an assessment of the changes in the overall quality of agencies' strategic plans, annual performance plans, and annual performance reports; (2) the challenges agencies face in measuring performance and using performance information in management decisions; and (3) how the federal government can continue to shift toward a more results-oriented focus. To meet our objectives, we reviewed our extensive prior work on GPRA best practices and implementation and collected governmentwide data to assess the government's overall focus on results. We conducted a random, stratified, governmentwide survey of federal managers comparable to surveys we conducted in 1997 and 2000. We also held eight in-depth focus groups--seven comprised of federal managers from 23 federal agencies and one with GPRA experts. We also interviewed top appointed officials from the current and previous administrations. Finally, we judgmentally selected a sample of six agencies to review for changes in the quality of their strategic plans, performance plans, and performance reports since their initial efforts. The agencies we selected included the Departments of Education (Education), Energy (DOE), Housing and Urban Development (HUD), and Transportation (DOT) and the Small Business (SBA) and Social Security Administrations (SSA). In making this selection, we chose agencies that collectively represented the full range of characteristics in the following four areas: (1) agency size (small, medium, large); (2) primary program type (direct service, research, regulatory, transfer payments, and contracts or grants); (3) quality of fiscal year 2000 performance plans based on our previous review;[Footnote 2] and (4) type of agency (cabinet department and independent agency). Appendix I contains a more detailed discussion of our scope and methodology. We performed our work in Washington, D.C., from January through November 2003 in accordance with generally accepted government auditing standards. Background: GPRA is the centerpiece of a statutory framework that Congress put in place during the 1990s to help resolve the long-standing management problems that have undermined the federal government's efficiency and effectiveness and to provide greater accountability for results. GPRA was intended to address several broad purposes, including strengthening the confidence of the American people in their government; improving federal program effectiveness, accountability, and service delivery; and enhancing congressional decision making by providing more objective information on program performance. GPRA requires executive agencies to complete strategic plans in which they define their missions, establish results-oriented goals, and identify the strategies that will be needed to achieve those goals. GPRA requires agencies to consult with Congress and solicit the input of others as they develop these plans. Through this strategic planning requirement, GPRA has required federal agencies to reassess their missions and long-term goals as well as the strategies and resources they will need to achieve their goals. Agencies developed their first strategic plans in fiscal year 1997, and are required to update the plans every 3 years since then. GPRA also requires executive agencies to prepare annual performance plans that articulate goals for the upcoming fiscal year that are aligned with their long-term strategic goals. These performance plans are to include results-oriented annual goals linked to the program activities displayed in budget presentations as well as the indicators the agency will use to measure performance against the results-oriented goals. Agencies developed their first annual performance plans in fiscal year 1999 and are required to issue plans annually thereafter to correspond with budget submissions to Congress. Finally, GPRA requires agencies to measure performance toward the achievement of the goals in the annual performance plan and report annually on their progress in program performance reports. If a goal was not met, the report is to provide an explanation and present the actions needed to meet any unmet goals in the future. These reports are intended to provide important information to agency managers, policymakers, and the public on what each agency accomplished with the resources it was given. Agencies issued their first annual performance reports on their fiscal year 1999 performance in fiscal year 2000 and are required to issue a report on each subsequent performance plan. The Office of Management and Budget (OMB) plays an important role in the management of federal government performance and specifically GPRA implementation. Part of OMB's overall mission is to ensure that agency plans and reports are consistent with the President's budget and administration policies. OMB is responsible for receiving and reviewing agencies' strategic plans, annual performance plans, and annual performance reports. To improve the quality and consistency of these documents, OMB issues annual guidance to agencies for their preparation, including guidelines on format, required elements, and submission deadlines. GPRA requires OMB to prepare a governmentwide performance plan, based on agencies' annual performance plan submissions. OMB also played an important role in the pilot phase of GPRA implementation by designating agencies for pilot projects in performance measurement, managerial accountability and flexibility, and performance budgeting, and assessing the results of the pilots. Finally, GPRA provides OMB with authority to grant agencies waivers to certain administrative procedures and controls. Recent OMB guidance--OMB Circular A-11, July 2003--requires agencies to submit "performance budgets" in lieu of annual performance plans for their fiscal year 2005 budget submission to OMB and Congress. According to OMB, performance budgets should satisfy all the statutory requirements of GPRA for annual performance plans. In addition, agencies are to include all performance goals used in the assessment of program performance done under OMB's Program Assessment Rating Tool (PART) process.[Footnote 3] Moreover, the guidance states that until all programs have been assessed by PART, the performance budget will also for a time include performance goals for agency programs that have not yet been assessed using PART. The expectation is that agencies are to substitute new or revised performance goals resulting from OMB's review for goals it deemed unacceptable. Results in Brief: Among the purposes of GPRA cited by Congress was to improve federal program effectiveness and service delivery by promoting a new focus on results, service quality, and customer satisfaction by setting program goals measuring performance against goals, and reporting publicly on progress. Furthermore, GPRA was to improve congressional decision making by providing more objective information on achieving objectives, and on the relative effectiveness and efficiency of federal programs and spending. Ten years after enactment, GPRA's requirements have laid a solid foundation of results-oriented agency planning, measurement, and reporting that have begun to address these purposes. Focus group participants and high-level political appointees, as well as OMB officials we interviewed, cited positive effects of GPRA that they generally attributed to GPRA's statutory requirements for planning and reporting. Performance planning and measurement have slowly yet increasingly become a part of agencies' cultures. The results of our stratified, random sample survey of federal managers indicate that since GPRA went into effect governmentwide in 1997, federal managers reported having significantly more of the types of performance measures called for by GPRA--particularly outcome-oriented performance measures. Survey data also suggested that more federal managers, especially at the Senior Executive Service (SES) level, believed that OMB was paying attention to their agencies' efforts under GPRA. GPRA has also begun to facilitate the linking of resources to results, although much remains to be done in this area. Beginning with agencies' initial efforts to develop effective strategic plans in 1997 and annual performance plans and reports for fiscal year 1999, Congress, GAO, and others have commented on the quality of those efforts and provided constructive feedback on how agency plans and reports could be improved. According to our current review of the strategic plans, annual performance plans, and annual performance reports of six selected agencies, these documents reflect much of the feedback that was provided. For example, goals are more quantifiable and results oriented, and agencies are providing more information about goals and strategies to address performance and accountability challenges and the limitations to their performance data. However, certain serious weaknesses persist, such as lack of detail on how annual performance goals relate to strategic goals and how agencies are coordinating with other entities to address common challenges and achieve common objectives. While a great deal of progress has been made in making federal agencies more results oriented, numerous challenges still exist. As we have noted before, top leadership commitment and sustained attention to achieving results, both within the agencies and at OMB, is essential to GPRA implementation. While one might expect an increase in agency leadership commitment since GPRA was implemented governmentwide beginning in fiscal year 1997, federal managers reported that such commitment has not significantly increased. Furthermore, although OMB has recently demonstrated leadership in its review of performance information from a budgetary perspective using the PART tool, it is unclear whether the results of those reviews, such as changes in program performance measures, will complement and be integrated with the long-term, strategic focus of GPRA. OMB provided significantly less guidance on GPRA implementation for the fiscal year 2005 budget, compared to the very detailed guidance provided in prior years. Without consistent guidance from OMB on meeting GPRA requirements and following best practices, it may be difficult to maintain the improvements in agency performance plans and reports or bring about improvements in areas where weaknesses remain. The commitment of top leadership within agencies, OMB, and Congress is critical to the success of strategic planning efforts. However, GPRA specifies time frames for updating strategic plans that do not correspond to presidential or congressional terms. As a result, an agency may be required to update its strategic plan a year before a presidential election and without input from a new Congress. A strategic plan should reflect the policy priorities of an organization's leaders and the input of key stakeholders if it is to be an effective management tool. Managers reported they had more performance measures, but indications that managers are making greater use of this information to improve performance are mixed. Additionally, managers reported several human capital-related challenges that impede results-oriented management, including a lack of authority and training to carry out GPRA requirements, as well as a lack of recognition for completing these tasks. Unfortunately, most existing federal performance appraisal systems are not designed to support a meaningful performance-based pay system in that they fail to link institutional, program, unit, and individual performance measurement and reward systems. Fewer than half of federal managers reported receiving relevant training in critical results-oriented management-related tasks. Managers also reported significant challenges persist in setting outcome-oriented goals, measuring performance, and collecting useful data. In some agencies, particularly those that have a research and development component, managers reported difficulties in establishing meaningful outcome measures. Managers also identified difficulties in distinguishing between the results produced by the federal program and results caused by external factors or nonfederal actors, such as with grant programs. Timely and useful performance information is not always available to federal agencies, making it more difficult to assess and report on progress achieved. Finally, agency officials believe that Congress could make greater use of performance information to conduct oversight and to inform appropriations decisions. GPRA provides a vehicle for Congress to explicitly state its performance expectations in outcome- oriented terms when establishing new programs or in exercising oversight of existing programs that are not achieving desired results. Mission fragmentation and overlap contribute to difficulties in addressing crosscutting issues, particularly when those issues require a national focus, such as homeland security, drug control, and the environment. GPRA requires a governmentwide performance plan, where these issues could be addressed in a centralized fashion, but OMB has not issued a distinct plan since 1999. Most recently, the President's fiscal year 2004 budget focused on describing agencies' progress in addressing the President's Management Agenda (PMA) and the results of PART reviews of agency programs. Such information is important and useful, but is not adequate alone to provide a broader and more integrated perspective of planned performance on governmentwide outcomes. GAO has previously reported on a variety of barriers to interagency cooperation, such as conflicting agency missions, jurisdiction issues, and incompatible procedures, data, and processes. A strategic plan for the federal government, supported by a set of key national indicators to assess the government's performance, position, and progress, could provide an additional tool for governmentwide reexamination of existing programs, as well as proposals for new programs. Such a plan could be of particular value in linking agencies' long-term performance goals and objectives horizontally across the government. In addition, it could provide a basis for integrating, rather than merely coordinating, a wide array of federal activities. To address these challenges, continued and sustained commitment and leadership are needed. OMB, as the primary focal point for overall management in the federal government, can provide this leadership and direction working with the various management councils and work groups of the government. Also, governmentwide planning could better facilitate the integration of federal activities to achieve national goals. GAO recommends that the Director of OMB (1) fully implement GPRA's requirement to develop a governmentwide performance plan; (2) articulate and implement an integrated, complementary relationship between GPRA and PART; (3) provide clearer and consistent guidance to executive branch agencies on how to implement GPRA; (4) continue to maintain a dialogue with agencies about their performance measurement practices with a particular focus on grant-making, research and development, and regulatory functions to identify and replicate successful approaches agencies are using to measure and report on their outcomes, including the use of program evaluation tools; and, work with executive branch agencies to identify the barriers to obtaining timely data to show progress against performance goals and the best ways to report information when there are unavoidable lags in data availability; and (5) work with agencies to ensure they are making adequate investments in training on performance planning and measurement, with a particular emphasis on how to use performance information to improve program performance. We also suggest that Congress consider amending GPRA to require that updates to agency strategic plans be submitted at least once every 4 years, 12-18 months after a new administration begins its term. Additionally, consultations with congressional stakeholders on existing strategic plans should be held at least once every new Congress and revisions should be made as needed. Further, we suggest Congress use these consultations and its oversight role to clarify its performance expectations for agencies. Congress should also consider amending GPRA to require the President to develop a governmentwide strategic plan. In commenting on a draft of this report, OMB generally agreed with our findings and conclusions. OMB agreed to implement most of our recommendations, but stated that the President's Budget represents the executive branch's governmentwide performance plan and could also serve as a governmentwide strategic plan. However, because of the budget's focus on agency-level expenditures for the upcoming fiscal year, we believe that the President's Budget provides neither a long-term nor an integrated perspective on the federal government's performance. OMB's comments appear in appendix VIII. Our response appears in chapter 5. We also provided relevant sections of the draft to the six agencies whose plans and reports we reviewed. DOE, HUD, and SSA disagreed with some of our observations, and we changed or clarified relevant sections of the report, as appropriate. Written comments from DOE, HUD, and SSA are reprinted in appendixes IX, X, and XI, respectively, along with our responses. Principal Findings: GPRA Laid the Foundation for a More Results-Oriented Federal Government: Prior to enactment of GPRA, our 1992 review of the collection and use of performance data by federal agencies revealed that, although many agencies collected performance information at the program level, few agencies had results-oriented performance information to manage or make strategic policy decisions for the agency as a whole.[Footnote 4] GPRA addressed agencies' shortcomings by creating a comprehensive and consistent statutory foundation of required agencywide strategic plans, annual performance plans, and annual performance reports. Participants in eight focus groups comprised of experts on GPRA and federal managers from 23 agencies cited the creation of this statutory foundation as one of the key accomplishments of GPRA. One of the premises of GPRA is that both congressional and executive branch oversight of federal agency performance were seriously hampered by a lack of adequate results- oriented goals and performance information. As noted above, prior to the enactment of GPRA few agencies reported their performance information externally. OMB officials we interviewed as part of our current review suggested that OMB has been a key consumer of the performance information produced under GPRA and that it has provided a foundation for their efforts to oversee agency performance. Federal managers' views of GPRA's effect on the federal government's ability to deliver results to the American public were mixed. When asked about the direct effects of GPRA on the public, 23 percent of the federal managers surveyed agreed to a moderate or greater extent that GPRA improved their agency's ability to deliver results to the American public. High-level political appointees we interviewed cited a number of examples of how the structure of GPRA created a greater focus on results in their agencies. Participants in our focus groups had mixed perceptions of GPRA's effect on their agency's ability to deliver results to the American public. Participants indicated GPRA has had a positive effect by shifting the focus of federal management from program activities and processes to achieving the intended results of those programs. Another major accomplishment of GPRA cited by focus group participants is that GPRA improved the transparency of government results to the American public. Other focus group participants had difficulty attributing the results their agencies achieved directly to GPRA's requirements. Focus group and survey results suggest that performance planning and measurement have slowly, but increasingly, become a part of agencies' cultures. Compared to the results of our 1997 governmentwide survey of federal managers, in our 2003 governmentwide survey more managers reported having performance measures for their programs. When we asked managers who said they had performance measures which of the five types of measures they had to a great or very great extent, they reported increases in all five types of measures between 1997 and 2003,[Footnote 5] all of which were statistically significant. Similarly, focus group participants commented on certain cultural changes that had taken place within their agencies since the passage of GPRA in which the "vocabulary" of performance planning and measurement- -e.g., a greater focus on performance measurement, orientation toward outcomes over inputs and outputs, and an increased focus on program evaluation--had become more pervasive. This perception is partly born out by our survey results. Consistent with our survey results indicating increases in results-oriented performance measures, we also observed a significant decline in the percentage of federal managers who agreed that certain factors hindered measuring performance or using the performance information. Finally, our survey data suggested that more federal managers, especially at the SES level, believed that OMB was paying attention to their agencies' efforts under GPRA, but with no corresponding increase in their concern that OMB would micromanage the programs in their agencies. Agencies have begun to establish a link between results and resources. Our 1998 assessment of fiscal year 1999 performance plans found that agencies generally covered the program activities in their budgets, but most plans did not identify how the funding for those program activities would be allocated to performance goals.[Footnote 6] However, our subsequent reviews of performance plans indicate that agencies have made progress in demonstrating how their performance goals and objectives relate to program activities in the budget. We reviewed a sample of six agencies' strategic plans (Education, DOE, HUD, DOT, SBA, and SSA) and found the quality of the selected plans reflected improvements over these agencies' initial strategic plans. Our 1997 review of agencies' draft strategic plans found that a significant amount of work remained to be done by executive branch agencies if their strategic plans were to fulfill the requirements of GPRA, serve as a basis for guiding agencies, and help congressional and other policymakers make decisions about agency activities and programs.[Footnote 7] The six strategic plans we looked at for this 2003 review reflected many new and continuing strengths as well as improvements over the 1997 initial draft plans, but we continued to find certain persistent weaknesses. Of the six elements required by GPRA, the plans generally discussed all but one--program evaluation, an area in which we have found agencies often lack capacity. Although the strategic plans listed the program evaluations agencies intended to complete over the planning period, they generally did not address how the agencies planned to use their evaluations to establish new or revise existing strategic goals, as envisioned by GPRA. Finally, although not required by GPRA, the strategic plans would have benefited from more complete discussions of how agencies planned to coordinate and collaborate with other entities to address common challenges and achieve common or complementary goals and objectives. The six selected agencies' fiscal year 2004 annual performance plans addressed some weaknesses of earlier plans, but there is still significant room for improvement. During our review of agencies' first annual performance plans, which presented agencies' annual performance goals for fiscal year 1999,[Footnote 8] we found that substantial further development was needed for these plans to be useful in a significant way to congressional and other decision makers. Most of the 2004 plans that we reviewed showed meaningful improvements over the fiscal year 1999 plans by showing a clearer picture of intended performance, providing strategies and resources that were more specifically related to achieving agency goals, and providing a greater level of confidence that performance data would be credible. But these plans also contained a number of serious weaknesses, such as inadequate discussion of coordination and collaboration and inconsistent or limited discussions of procedures used to verify and validate performance data, which limited their quality and undermined their usefulness. Our review of the six agencies' fiscal year 2002 performance reports showed a number of strengths and improvements over their fiscal year 1999 performance reports, as well as areas that needed improvement. As we found in our earlier reviews, the six agencies' fiscal year 2002 reports generally allowed for an assessment of progress made in achieving agency goals. In addition, the majority of agencies discussed the progress achieved in addressing performance and accountability challenges identified by agency inspectors general and GAO. However, as with the fiscal year 1999 reports, many of the weaknesses we identified in the agencies' fiscal year 2002 reports were related to the significant number of performance goals not achieved or for which performance data were unavailable. In addition, the majority of the reports we reviewed did not include other GPRA requirements, such as a summary of the findings from program evaluations. Finally, only one of the six agencies clearly linked its costs to the achievement of performance goals or objectives. Challenges to GPRA Implementation Exist: While a great deal of progress has been made in making federal agencies more results oriented, numerous challenges still exist to effective implementation of GPRA. We observed in our 1997 report that we would expect to see managers' positive perceptions on items, such as the extent to which top leadership is committed to achieving results, become more prevalent and the gap between SES and non-SES managers begin to narrow as GPRA and related reforms are implemented. However, these changes do not appear to be happening to the extent anticipated. The need for strong, committed, and sustained leadership extends to OMB as well. OMB has shown a commitment to improving the management of federal programs, both through its leadership in reviewing agency program performance using the PART tool as well as through the PMA. As part of the President's budget preparation, PART clearly must serve the President's interests. However, it is not well suited to addressing crosscutting (or horizontal) issues or to looking at broad program areas in which several programs address a common goal. GPRA was designed to address the needs of many users of performance information, including (1) Congress to provide oversight and inform funding decisions, (2) agency managers to manage programs and make internal resource decisions, and (3) the public to provide greater accountability. It is not yet clear the extent to which PART performance goals and measures will compete with agencies' long-term, strategic GPRA goals and objectives that were established in consultation with Congress and other stakeholders. We also found that, while the quality of agency plans and reports have improved overall since their initial efforts, they continue to suffer from certain persistent weaknesses as noted above. However, OMB's July 2003 guidance for preparation and submission of annual performance plans is significantly shorter and less detailed than its 2002 guidance. Consistent, more explicit OMB guidance on preparing GPRA documents can help ensure that gains in the quality of GPRA documents are maintained and provide a resource for agency managers to make further improvements in those documents. We also found that timing issues may affect the development of agency strategic plans that are meaningful and useful to top leadership. The commitment and sustained attention of top leadership within agencies, OMB, and Congress is critical to the success of strategic planning efforts. A strategic plan should reflect the policy priorities of an organization's leaders and the input of key stakeholders if it is to be an effective management tool. However, GPRA specifies time frames for updating strategic plans that do not correspond to presidential or congressional terms. As a result, an agency may be required to update its strategic plan a year before a presidential election and without input from a new Congress. If a new president is elected, the updated plan is essentially moot and agencies must spend additional time and effort revising it to reflect new priorities. Our focus group participants, including GPRA experts, strongly agreed that this timing issue should be addressed by adjusting time frames to correspond better with presidential and congressional terms. The benefit of collecting performance information is only fully realized when this information is actually used by managers to bring about desired results. However, federal managers reported mixed results in the use of performance information. Focus group participants and survey respondents noted that although many federal managers understand and use results-oriented management concepts in their day-to-day activities, such as strategic planning and performance measurement, they do not always connect these concepts to the requirements of GPRA. According to our 2003 survey results, the reported use of performance information to a great or very great extent for nine management activities, such as setting program priorities or setting individual job expectations for staff, ranging from 41 to 66 percent, has not changed significantly since our first survey in 1997. One exception was the reported use to a great or very great extent of performance information to adopt new program approaches or change work processes, which was significantly lower than the 1997 results. GPRA's usefulness to agency leaders and managers as a tool for management and accountability was cited as a key accomplishment numerous times by focus group participants. However, a number of alternative views indicated that the usefulness of GPRA as a management tool has been limited. Our survey data also indicate that managers' perceive their participation in activities related to the development and use of performance information has been limited. Federal managers continue to confront a range of important human capital management challenges. These managers report that they are held accountable for program results, but may not have the decision-making authority they need to accomplish agency goals. Moreover, fewer than half of managers reported receiving relevant training. Managers also perceive a lack of positive recognition for helping agencies achieve results. Unfortunately, most existing federal performance appraisal systems are not designed to support a meaningful performance-based pay system in that they fail to link institutional, program, unit, and individual performance measurement and reward systems. In our view, one key need is to modernize performance management systems in executive agencies so that they link to the agency's strategic plan, related goals, and desired outcomes and are therefore capable of adequately supporting more performance-based pay and other personnel decisions. Managers reported persistent challenges in setting outcome-oriented goals, measuring performance, and collecting useful data. Focus group participants and survey respondents noted that outcome-oriented performance measures were especially difficult to establish when the program or line of effort was not easily quantifiable. In some agencies, particularly those that have a research and development component, managers reported difficulties in establishing meaningful outcome measures. Managers also identified difficulties in distinguishing between the results produced by the federal program and results caused by external factors or nonfederal actors, such as with grant programs. Finally, managers reported that timely and useful performance information is not always available. Crosscutting issues continue to be a challenge to GPRA implementation. Our review of six agencies' strategic and annual performance plans showed some improvement in addressing their crosscutting program efforts, but a great deal of improvement is still necessary. We have previously reported and testified that GPRA could provide OMB, agencies, and Congress with a structured framework for addressing crosscutting policy initiatives and program efforts. OMB could use the provision of GPRA that calls for OMB to develop a governmentwide performance plan to integrate expected agency-level performance. It could also be used to more clearly relate and address the contributions of alternative federal strategies. Unfortunately, this provision has not been fully implemented. Instead, OMB has used the President's Budget to present high-level information about agencies and certain program performance issues. The current agency-by-agency focus of the budget does not provide the integrated perspective of government performance envisioned by GPRA. For example, the fiscal year 2004 budget identified budget requests and performance objectives by agency, such as the U.S. Department of Defense, as opposed to crosscutting governmentwide themes. From this presentation, one could assume that the only activities the U.S. government planned to carry out in support of national defense were those listed under the chapter "Department of Defense." However, the chapter on the fiscal year 2004 budget discussing "the Department of State and International Assistance Programs," contains a heading titled, "Countering the Threat from Weapons of Mass Destruction." And while OMB may have a technical reason for not classifying this task as being related to national defense or homeland security, it is unclear that a lay reader could make that distinction. The fiscal year 2005 budget also identified budget requests and performance objectives by agency, not by crosscutting theme. A strategic plan for the federal government could provide an additional tool for governmentwide reexamination of existing programs, as well as proposals for new programs. If fully developed, a governmentwide strategic plan could potentially provide a cohesive perspective on the long-term goals of the federal government and provide a much needed basis for fully integrating, rather than merely coordinating, a wide array of federal activities. Successful strategic planning requires the involvement of key stakeholders. Thus, it could serve as a mechanism for building consensus. Further, it could provide a vehicle for the President to articulate long-term goals and a road map for achieving them. In addition, a strategic plan could provide a more comprehensive framework for considering organizational changes and making resource decisions. The development of a set of key national indicators could be used as a basis to inform the development of governmentwide strategic and annual performance plans. The indicators could also link to and provide information to support outcome-oriented goals and objectives in agency-level strategic and annual performance plans. Focus group members believed that one of the main challenges to GPRA implementation was the reluctance of Congress to use performance information when making decisions, especially appropriations decisions. However, less than one quarter of federal managers in the 2003 survey shared that concern. Further, a recent Congressional Research Service review suggests that Congress uses performance information to some extent, as evidenced by citations in legislation and committee reports. While there is concern regarding Congress' use of performance information, it is important to make sure that this information is useful. In other words, the information presented and its presentation must meet the needs of the user. Regular consultation with Congress about both the content and format of performance plans and reports is critical. As a key user of performance information, Congress also needs to be considered a partner in shaping agency goals at the outset. GPRA provides a vehicle for Congress to explicitly state its performance expectations in outcome-oriented terms when consulting with agencies on their strategic plans or when establishing new programs or exercising oversight of existing programs that are not achieving desired results. This would provide important guidance to agencies that could then be incorporated in agency strategic and annual performance plans. Recommendations for Executive Action: GAO recommends that the Director of OMB implement five suggestions to improve its guidance and oversight of GPRA implementation. To provide a broader perspective and more cohesive picture of the federal government's goals and strategies to address issues that cut across executive branch agencies, we recommend that the Director of OMB fully implement GPRA's requirement to develop a governmentwide performance plan. To achieve the greatest benefit from both GPRA and PART, we recommend that the Director of OMB articulate and implement an integrated and complementary relationship between the two. GPRA is a broad legislative framework that was designed to be consultative with Congress and other stakeholders, and allows for varying uses of performance information. PART looks through a particular lens for a particular use--the executive budget formulation process. To improve the quality of agencies' strategic plans, annual performance plans, and performance reports and help agencies meet the requirements of GPRA, we recommend that the Director of OMB provide clearer and consistent guidance to executive branch agencies on how to implement GPRA. Such guidance should include standards for communicating key performance information in concise as well as longer formats to better meet the needs of external users who lack the time or expertise to analyze lengthy, detailed documents. To help address agencies' performance measurement challenges, we recommend the Director of OMB engage in a continuing dialogue with agencies about their performance measurement practices with a particular focus on grant-making, research and development, and regulatory functions to identify and replicate successful approaches agencies are using to measure and report on their outcomes, including the use of program evaluation tools. Additionally, we recommend that the Director of OMB work with executive branch agencies to identify the barriers to obtaining timely data to show progress against performance goals and the best ways to report information where there are unavoidable lags in data availability. Governmentwide councils, such as the President's Management Council and the Chief Financial Officers Council, may be effective vehicles for working on these issues. To facilitate the transformation of agencies' management cultures to be more results oriented, we recommend that the Director of OMB work with agencies to ensure they are making adequate investments in training on performance planning and measurement, with a particular emphasis on how to use performance information to improve program performance. Matters for Congressional Consideration: GAO also identified two matters for congressional consideration to improve the governmentwide focus on results. To ensure that agency strategic plans more closely align with changes in the federal government leadership, Congress should consider amending GPRA to require that updates to agency strategic plans be submitted at least once every 4 years, 12-18 months after a new administration begins its term. Additionally, consultations with congressional stakeholders should be held at least once every new Congress and interim updates made to strategic and performance plans as warranted. Congress should consider using these consultations along with its traditional oversight role and legislation as opportunities to clarify its performance expectations for agencies. This process may provide an opportunity for Congress to develop a more structured oversight agenda. To provide a framework to identify long-term goals and strategies to address issues that cut across federal agencies, Congress should consider amending GPRA to require the President to develop a governmentwide strategic plan. Agency Comments: We provided a copy of the draft report to OMB for comment. OMB's written comments are reprinted in appendix VIII. In general, OMB agreed with our findings and conclusions. OMB agreed to implement most of our recommendations, noting that these recommendations will enhance their efforts to make the government more results oriented. OMB agreed to (1) work with agencies to ensure they are provided adequate training in performance management, (2) revise its guidance to clarify the integrated and complementary relationship between GPRA and PART, and (3) continue to use PART to improve agency performance measurement practices and share those practices across government. In response to our recommendation that OMB fully implement GPRA's requirement to develop a governmentwide performance plan, OMB stated that the President's Budget represents the executive branch's governmentwide performance plan. However, according to GAO's review, the agency-by-agency focus of the budget over the past few years does not provide an integrated perspective of government performance, and thus does not meet GPRA's requirement to provide a "single cohesive picture of the annual performance goals for the fiscal year." To clarify this point, we added an example that illustrates the lack of integration between crosscutting issues in the budget. In response to our matter for congressional consideration that Congress should consider amending GPRA to require the President to develop a governmentwide strategic plan, OMB noted that the budget serves as the governmentwide strategic plan. However, the President's Budget focuses on establishing agency budgets for the upcoming fiscal year. Unlike a strategic plan, it provides neither a long-term nor an integrated perspective on the federal government's activities. A governmentwide strategic plan should provide a cohesive perspective on the long-term goals of the federal government and provide a basis for fully integrating, rather than primarily coordinating, a wide array of existing and relatively short-term federal activities. We provided relevant sections of the draft report to Education, DOE, HUD, SBA, SSA, and DOT. Education and SBA did not provide any comments, while DOT provided minor technical comments. DOE, HUD, and SSA disagreed with some of our observations on their strategic plans, performance plans, and performance reports; we changed or clarified relevant sections of the report, as appropriate. Written comments from DOE, HUD, and SSA are reprinted in appendixes IX, X, and XI, respectively, along with our responses. [End of section] Chapter 1: Introduction: From defending the homeland against terrorists, to preventing the spread of infectious diseases, to providing a reliable stream of social security income to retirees and supporting the transition from welfare to work, the federal government provides funding and services to the American public that can affect their lives in critical ways every day. However, the federal government is in a period of profound transition and faces an array of challenges and opportunities to enhance performance, ensure accountability, and position the nation for the future. A number of overarching trends, such as diffuse security threats and homeland security needs, increasing global interdependency, the shift to knowledge-based economies, and the looming fiscal challenges facing our nation, drive the need to reconsider the proper role for the federal government in the 21st century, how the government should do business (including how it should be structured), and in some instances, who should do the government's business. Without effective short-and long-term planning, which takes into account the changing environment and needs of the American public and the challenges they face and establishes goals to be achieved, federal agencies risk delivering programs and services that may or may not meet society's most critical needs. At a cost to taxpayers of over $2 trillion annually, the federal government should be able to demonstrate to the American public that it can anticipate emerging issues, develop sound strategies and plans to address them, and be accountable for the results that have been achieved. Concerned that the federal government was more focused on program activities and processes than the results to be achieved, Congress passed the Government Performance and Results Act of 1993 (GPRA).[Footnote 9] The act required federal agencies to develop strategic plans with long-term strategic goals, annual goals linked to achieving the long-term goals, and annual reports on the results achieved. Now that GPRA has been in effect for 10 years, you asked us to assess the effectiveness of GPRA in creating a focus on results in the federal government. Specifically, this report discusses (1) the effect of GPRA over the last 10 years in creating a governmentwide focus on results and the government's ability to deliver results to the American public, including an assessment of the changes in the overall quality of agencies' strategic plans, annual performance plans, and annual performance reports; (2) the challenges agencies face in measuring performance and using performance information in management decisions; and (3) how the federal government can continue to shift toward a more results-oriented focus. Impact of Emerging Trends and Fiscal Challenges: With the 21st century challenges we are facing, it is more vital than ever to maximize the performance of federal agencies in achieving their long-term goals. The federal government must address and adapt to major trends in our country and around the world. At the same time, our nation faces serious long-term fiscal challenges. Increased pressure also comes from world events: both from the recognition that we cannot consider ourselves "safe" between two oceans--which has increased demands for spending on homeland security--and from the United States (U.S.) role in combating terrorism in an increasingly interdependent world. To be able to assess federal agency performance and hold agency managers accountable for achieving their long-term goals, we need to know what the level of performance is. GPRA planning and reporting requirements can provide this essential information. Our country's transition into the 21st century is characterized by a number of key trends, including: * the national and global response to terrorism and other threats to our personal and national security; * the increasing interdependence of enterprises, economies, markets, civil societies, and national governments, commonly referred to as globalization; * the shift to market-oriented, knowledge-based economies; * an aging and more diverse U.S. population; * rapid advances in science and technology and the opportunities and challenges created by these changes; * challenges and opportunities to maintain and improve the quality of life for the nation, communities, families, and individuals; and: * the changing and increasingly diverse nature of governance structures and tools. As the nation and government policymakers grapple with the challenges presented by these evolving trends, they do so in the context of rapidly building fiscal pressures. GAO's long-range budget simulations show that this nation faces a large and growing structural deficit due primarily to known demographic trends and rising health care costs. The fiscal pressures created by the retirement of the baby boom generation and rising health costs threaten to overwhelm the nation's fiscal future. As figure 1 shows, by 2040, absent reform or other major tax or spending policy changes, projected federal revenues will likely be insufficient to pay more than interest on publicly held debt. Further, our recent shift from surpluses to deficits means the nation is moving into the future in a more constrained fiscal position. Figure 1: Composition of Spending as a Share of GDP Assuming Discretionary Spending Grows with GDP after 2003 and All Expiring Tax Provisions Are Extended: [See PDF for image] Notes: Although all expiring tax cuts are extended, revenue as a share of gross domestic product (GDP) increases through 2013 due to (1) real bracket creep, (2) more taxpayers becoming subject to the Alternative Minimum Tax, and (3) increased revenue from tax-deferred retirement accounts. After 2013, revenue as a share of GDP is held constant. This simulation assumes that currently scheduled Social Security benefits are paid in full throughout the simulation period. [End of figure] The United States has had a long-range budget deficit problem for a number of years, even during recent years in which we had significant annual budget surpluses. Unfortunately, the days of surpluses are gone, and our current and projected budget situation has worsened significantly. The bottom line is that our projected budget deficits are not manageable without significant changes in "status quo" programs, policies, processes, and operations. GPRA Background: GPRA is the centerpiece of a statutory framework that Congress put in place during the 1990s to help resolve the long-standing management problems that have undermined the federal government's efficiency and effectiveness and to provide greater accountability for results. In addition to GPRA, the framework comprises the Chief Financial Officers Act of 1990, as amended by the Government Management Reform Act of 1994, and information technology reform legislation, including the Paperwork Reduction Act of 1995 and the Clinger-Cohen Act of 1996. Together, these laws provide a powerful framework for developing and integrating information about agencies' missions and strategic priorities, the results-oriented performance goals that flow from those priorities, performance data to show the level of achievement of those goals, and the relationship of reliable and audited financial information and information technology investments to the achievement of those goals. GPRA was intended to address several broad purposes, including strengthening the confidence of the American people in their government; improving federal program effectiveness, accountability, and service delivery; and enhancing congressional decision making by providing more objective information on program performance. The basic requirements of GPRA for the preparation of strategic plans, annual performance plans, and annual program performance reports by executive branch agencies are the following: 1. The agency’s strategic plan must contain these six key elements: * a comprehensive agency mission statement; * agencywide long-term goals and objectives for all major functions and operations; * approaches (or strategies) and the various resources needed to achieve the goals and objectives; * a description of the relationship between the long-term goals and objectives and the annual performance goals; * an identification of key factors, external to the agency and beyond its control, that could significantly affect the achievement of the strategic goals; and * a description of how program evaluations were used to establish or revise strategic goals and a schedule for future program evaluations. 2. The agency must develop annual performance plans covering each program activity set forth in the agencies’ budgets:* Building on the decisions made as part of the strategic planning process, GPRA requires executive agencies to develop annual performance plans covering each program activity set forth in the agencies’ budgets. Annual performance plans, covering the upcoming fiscal year, are to be submitted to Congress after the President’s Budget is submitted, which generally occurs in February. Each plan is to contain an agency’s annual performance goals and associated measures, which the agency is to use in order to gauge its progress toward accomplishing its strategic goals. OMB is to use the agencies’ performance plans to develop an overall federal government performance plan that is to be submitted with the President’s Budget. The performance plan for the federal government is to present to Congress a single cohesive picture of the federal government’s annual performance goals for a given fiscal year. 3. The agency must prepare annual reports on program performance for the previous fiscal year, to be issued by March 31 each year: GPRA requires executive agencies to prepare annual reports on program performance for the previous fiscal year, to be issued by March 31 each year. In each report, an agency is to compare its performance against its goals, summarize the findings of program evaluations completed during the year, and describe the actions needed to address any unmet goals. Recent OMB guidance states that executive agencies must combine their program performance report with their accountability report and transmit the combined report for fiscal year 2003 by January 30, 2004, and the combined report for fiscal year 2004 by November 15, 2004. *Program activity refers to the lists of projects and activities in the appendix portion of the Budget of the United States Government. Program activity structures are intended to provide a meaningful representation of the operations financed by a specific budget account. The Office of Management and Budget (OMB) plays an important role in the management of the federal government's performance, and specifically GPRA implementation. Part of OMB's overall mission is to ensure that agency plans and reports are consistent with the President's Budget and administration policies. OMB is responsible for receiving and reviewing agencies' strategic plans, annual performance plans, and annual performance reports. To improve the quality and consistency of these documents, OMB issues annual guidance to agencies for their preparation, including guidelines on format, required elements, and submission deadlines.[Footnote 10] GPRA requires OMB to prepare the overall governmentwide performance plan, based on agencies' annual performance plan submissions. OMB also played an important role in the pilot phase of GPRA implementation by designating agencies for pilot projects in performance measurement, managerial accountability and flexibility, and performance budgeting, and assessing the results of the pilots. Finally, GPRA provides OMB with authority to grant agencies waivers to certain administrative procedures and controls. Recent OMB guidance[Footnote 11] requires agencies to submit "performance budgets" in lieu of annual performance plans for their budget submission to OMB and Congress. Performance budgets are to meet all the statutory requirements of GPRA for annual performance plans. In addition, agencies are to include all performance goals used in the assessment of program performance done under OMB's Program Assessment Rating Tool (PART) process.[Footnote 12] Moreover, the guidance states that until all programs have been assessed by PART, the performance budget will also for a time include performance goals for agency programs that have not yet been assessed using PART. The expectation is that agencies are to substitute new or revised performance goals resulting from OMB's review for goals it deemed unacceptable. In crafting GPRA, Congress recognized that managerial accountability for results is linked to managers having sufficient flexibility, discretion, and authority to accomplish desired results. GPRA authorizes agencies to apply for managerial flexibility waivers in their annual performance plans beginning with fiscal year 1999. The authority of agencies to request waivers of administrative procedural requirements and controls is intended to provide federal managers with more flexibility to structure agency systems to better support program goals. The nonstatutory requirements that OMB can waive under GPRA generally involve the allocation and use of resources, such as restrictions on shifting funds among items within a budget account. Agencies must report in their annual performance reports on the use and effectiveness of any GPRA managerial flexibility waivers that they receive. OMB was to designate at least five agencies from the first set of pilot projects to test managerial accountability and flexibility during fiscal years 1995 and 1996. We previously reported on the results of the pilot project to implement managerial flexibility waivers and found that the pilot did not work as intended.[Footnote 13] OMB did not designate any of the seven departments and one independent agency that submitted a total of 61 waiver proposals as GPRA managerial accountability and flexibility pilots. For about three-quarters of the waiver proposals, OMB or other central management agencies determined that the waivers were not allowable for statutory or other reasons or that the requirement for which the waivers were proposed no longer existed. For the remaining proposals, OMB or other central management agencies approved waivers or developed compromises by using authorities that were already available independent of GPRA. Under GPRA, another set of pilot projects, which were scheduled for fiscal years 1998 and 1999, were to test performance budgeting--i.e., the presentation of the varying levels of performance that would result from different budget levels. We previously reported that OMB initially deferred these pilots--originally to be designated in fiscal years 1998 and 1999--to give federal agencies time to develop the capability of calculating the effects of marginal changes in cost or funding on performance.[Footnote 14] When the pilots began in August 1999, OMB designed them as case studies prepared by OMB staff to demonstrate how performance information could be used to compare alternatives and to develop funding recommendations for incorporation into the President's fiscal year 2001 budget submission. On January 18, 2001, OMB reported the results of five performance budgeting pilots that explored agencies' capabilities of more formally assessing the effects of different funding levels on performance goals. OMB selected the pilots to reflect a cross section of federal functions and capabilities so that a representative range of measurement and reporting issues could be explored. In its report, OMB concluded that legislative changes were not needed. OMB reported that the pilots demonstrated that assuring further performance measurement improvements and steadily expanding the scope and quality of performance measures is paramount, and that the existing statute provides sufficient latitude for such improvement. Overall, OMB concluded that the pilots raised several key challenges about performance budgeting at the federal level including, for example, the following: * In many instances, measuring the effects of marginal, annual budget changes on performance is not precise or meaningful. * While continuing to change from an almost total reliance on output measures to outcome measures, it will be much more difficult to associate specific resource levels with those outcomes, particularly over short periods of time. * Establishing clear linkages between funding and outcomes will vary by the nature of the program and the number of external factors. * Delays in the availability of performance data, sometimes caused by agencies' reliance on nonfederal program partners for data collection, will continue to present synchronization problems during budget formulation. Scope and Methodology: To meet the three objectives stated earlier, we reviewed our extensive prior work on GPRA best practices and implementation and collected governmentwide data to assess the government's overall focus on results. We conducted a random, stratified, governmentwide survey of federal managers comparable to surveys we conducted in 1997 and 2000. We also held eight in-depth focus groups--seven comprised of federal managers from 23 federal agencies and one with GPRA experts. We also interviewed top appointed officials from the current and previous administrations. Finally, we judgmentally selected a sample of six agencies to review for changes in the quality of their strategic plans, performance plans, and performance reports since their initial efforts. The agencies we selected were the Departments of Education (Education), Energy (DOE), Housing and Urban Development (HUD), and Transportation (DOT) and the Small Business (SBA) and Social Security Administrations (SSA). In making this selection, we chose agencies that collectively represented the full range of characteristics in the following four areas: (1) agency size (small, medium, large); (2) primary program types (direct service, research, regulatory, transfer payments, and contracts or grants); (3) quality of fiscal year 2000 performance plan based on our previous review (low, medium, high);[Footnote 15] and (4) type of agency (cabinet department and independent agency). Appendix I contains a more detailed discussion of our scope and methodology. We performed our work in Washington, D.C., from January through November 2003 in accordance with generally accepted government auditing standards. Major contributors to this report are listed in appendix XII. [End of section] Chapter 2: GPRA Established the Foundation for a More Results-Oriented Federal Government: Among the purposes of GPRA cited by Congress was to improve federal program effectiveness and service delivery by promoting a new focus on results, service quality, and customer satisfaction by setting program goals, measuring performance against goals, and reporting publicly on progress. Furthermore, GPRA was to improve congressional decision making by providing better information on achieving objectives, and on the relative effectiveness and efficiency of federal programs and spending. Ten years after enactment, GPRA's requirements have laid a foundation of results-oriented agency planning, measurement, and reporting that have begun to address these purposes. Focus group participants, high-level political appointees, and OMB officials we interviewed cited positive effects of GPRA that they generally attributed to GPRA's statutory requirements for planning and reporting. Our survey results indicate that since GPRA went into effect governmentwide in 1997, federal managers reported having significantly more of the types of performance measures called for by GPRA-- particularly outcome-oriented performance measures. GPRA has also begun to facilitate the linking of resources to results, although much remains to be done in this area. GPRA Statutory Requirements Laid a Foundation for Agencywide Results- Oriented Management: Prior to enactment of GPRA, our 1992 review of the collection and use of performance data by federal agencies revealed that, although many agencies collected performance information at the program level, few agencies had results-oriented performance information to manage or make strategic policy decisions for the agency as a whole.[Footnote 16] Federal agencies surveyed indicated that many had a single, long-term plan that contained goals, standards, or objectives for the entire agency or program. Many of these agencies also reported they collected a wide variety of performance measures. However, in validating the survey responses with a sample of agencies, we found that measures were typically generated and used by program-level units within an agency and focused on measuring work activity levels and outputs or compliance with statutes. Little of this performance information was transparent to Congress, OMB, or the public and few of the agencies we visited used performance measures to manage toward long-term objectives. Few of the agencies surveyed had the infrastructure in place, such as a unified strategic plan with measurable goals, an office that collected performance measures, and regular consolidated reports, to tie plans and measures. GPRA addressed these shortcomings by creating a comprehensive and consistent statutory foundation of required agencywide strategic plans, annual performance plans, and annual performance reports. In contrast to prior federal government efforts to measure performance, GPRA explicitly emphasized that, in addition to performance indicators that agencies may need to manage programs on a day-to-day basis, such as quantity, quality, timeliness, and cost, agencies also needed outcome- oriented goals and measures that assess the actual results, effects, or impact of a program or activity compared to its intended purpose. Expert and agency focus group participants cited the creation of this statutory foundation as one of the key accomplishments of GPRA. Participants agreed that GPRA created a framework in statute for federal agencies to plan their activities in order to become more results oriented and provided a managerial tool for program accountability. Using this framework, agencies could develop and focus on strategies to carry out the programs they administer; set goals and identify performance indicators that will inform them whether or not they achieved the performance they expected; and determine what impact, if any, their programs have had on the American public. According to the experts in one of our focus groups, comparing federal agencies' current mission statements contained in their strategic plans to what they were in the past demonstrates that the agencies have done some "soul searching" to get a better sense of what their role is (or should be) and how they can achieve it. Given that GPRA is in statute, participants indicated that the use of this planning framework is likely to be sustained within agencies. One of the premises of GPRA is that both congressional and executive branch oversight of federal agency performance were seriously hampered by a lack of adequate results-oriented goals and performance information. As noted above, prior to the enactment of GPRA few agencies reported their performance information externally. OMB officials we interviewed as part of our current review suggested that OMB has been a key consumer of the performance information produced under GPRA and that it has provided a foundation for their efforts to oversee agency performance. For example, during the development of the fiscal year 2004 budget, OMB used PART to review and rate 234 federal programs. We recently reported that one of PART's major impacts was its ability to highlight OMB's recommended changes in program management and design.[Footnote 17] PART reviews look at four elements--program purpose and design, strategic planning, program management, and program results/accountability--and rate the program on how well each of these elements is executed. However, without the foundation of missions, goals, strategies, performance measures, and performance information generated under GPRA, such oversight would be difficult to carry out. Participants in most of our focus groups also agreed that GPRA has been a driving force behind many cultural changes that have occurred within federal agencies. Highlighting the focus on results, participants stated that GPRA had stimulated a problem-solving approach within federal agencies and encouraged agency managers to think creatively when developing performance indicators for their programs. GPRA has also changed the dialogue within federal agencies; front-line managers and staff at lower levels of the organization now discuss budget issues in connection with performance. Similarly, experts noted that information about performance management and resource investments are more frequently communicated between agency officials and Congress than in the past. Within agencies, GPRA documents can provide a context of missions, goals, and strategies that political appointees can use to articulate agencies' priorities. Views on GPRA's Effect on the Federal Government's Ability to Deliver Results to the American Public Were Mixed: A key purpose of GPRA was "to improve the confidence of the American people in the capability of the Federal Government, by systematically holding Federal agencies accountable for achieving program results." When asked about the direct effects of GPRA on the public in our 2003 survey, an estimated 23 percent of federal managers agreed to a moderate or greater extent that GPRA improved their agency's ability to deliver results to the American public; a larger percentage--38 percent--chose a "no basis to judge/not applicable" category. When a similar question was posed in our focus groups with experts and federal managers, participants' views were generally mixed. Some federal managers in our focus groups agreed that GPRA has had a positive effect on raising awareness on many performance issues, and that in and of itself is a way of delivering results. The information gathered and reported for GPRA allows agencies to make better-informed decisions, which improves their ability to achieve results. Other participants stated that while certain aspects of GPRA-related work have been positive, agencies' ability to deliver results and public awareness of their activities cannot always be exclusively attributed to GPRA. For example, some participants stated that many agencies rely on grant recipients to carry out their work, and delivering results to the American public depends, to a large extent, on the diligence of these organizations to implement their programs; such results would not change dramatically if GPRA were no longer a requirement. A number of the political appointees we interviewed cited examples of outcomes they believe would not have occurred without the structure of GPRA. For example, a former deputy secretary of the Department of Veterans Affairs (VA) stated that "the Results Act brought about a fundamental rethinking of how we managed our programs and processes. . . . We developed a strategic plan that was veteran-focused. . . . We made every effort to define program successes from the veteran's perspective." A former Chief Financial Officer (CFO) cited Customs Service goals to reduce the quantity of illegal drugs flowing into the United States and the Food and Drug Administration's focus on speeding up the approval of new drugs as examples of outcomes that can make a big difference in people's lives. Another major accomplishment of GPRA cited by our focus group participants is that GPRA improved the transparency of government results to the American public. As noted above, prior to GPRA, few agencies reported performance results outside of their agencies. Focus group participants indicated a key accomplishment of GPRA was its value as a communication tool by increasing the transparency to the public of what their agencies did in terms the public could understand. For example, information on agencies' strategic plans, performance goals, measures, and results are easily obtainable from agency Web sites. One focus group participant commented that GPRA helps bureaucrats explain to nonbureaucrats what the federal government does in terms they can better understand. Other comments indicated that because of GPRA agencies could now tell Congress and the American public what they are getting for their money. More Managers Reported Having Performance Measures: A fundamental element in an organization's efforts to manage for results is its ability to set meaningful goals for performance and to measure performance against those goals. From our 2003 survey we estimate that 89 percent of federal managers overall said there were performance measures for the programs they were involved with. This is a statistically significantly higher percentage than the 76 percent of managers who answered yes to this item on our 1997 survey. (See fig. 2.): Figure 2: Percentage of Federal Managers Who Reported That There Were Performance Measures for the Programs with Which They Were Involved: [See PDF for image] [A] There was a statistically significant difference between 1997 and 2003 surveys. [End of figure] Moreover, when we asked managers who said they had performance measures which of five types of measures they had to a great or very great extent, they reported increases in all five types of measures between 1997 and 2003,[Footnote 18] all of which were statistically significant. (See fig. 3.) Notably, managers indicated the existence of outcome measures, defined as "performance measures that demonstrate to someone outside the organization whether or not intended results are being achieved," grew from a low of 32 percent in 1997 to the current estimate of 55 percent, a level that is on par with output measures for the first time since we began our survey. Figure 3: Percentage of Federal Managers Who Reported Having Specific Types of Performance Measures to a Great or Very Great Extent: [See PDF for image] [A] There was a statistically significant difference between 1997 and 2003 surveys. [End of figure] Similarly, focus group participants commented on certain cultural changes that had taken place within their agencies since the passage of GPRA in which the "vocabulary" of performance planning and measurement- -e.g., a greater focus on performance management; orientation toward outcomes over inputs and outputs; and an increased focus on program evaluation--had become more pervasive. This perception is partly born out by our survey results. Since 1997 those reporting a moderate to extensive knowledge of GPRA and its requirements shifted significantly from 26 percent to 41 percent in 2003, while those reporting no knowledge of GPRA declined significantly from 27 percent to 20 percent. (See fig. 4.): Figure 4: Percentage of Federal Managers Who Reported Their Awareness of GPRA: [See PDF for image] [A] There was a statistically significant difference between 1997 and 2003 surveys. [End of figure] Consistent with our survey results indicating increases in results- oriented performance measures and increasing GPRA knowledge, we also observed a significant decline in the percentage of federal managers who agreed that certain factors hindered measuring performance or using the performance information. For example, as shown in figure 5, of those who expressed an opinion, the percentage of managers who noted that determining meaningful measures was a hindrance to a great or very great extent was down significantly from 47 percent in 1997 to 36 percent in 2003. Likewise, the percentage that agreed to a great or very great extent that different parties are using different definitions to measure performance was a hindrance also declined significantly from 49 percent in 1997 to 36 percent in 2003. Figure 5: Percentage of Federal Managers Who Reported Hindrances to Measuring Performance or Using the Performance Information to a Great or Very Great Extent: [See PDF for image] Note: Percentages are based on those respondents answering on the extent scale. [A] There was a statistically significant difference between 1997 and 2003. [End of figure] Finally, our survey data suggested that more federal managers, especially at the Senior Executive Service (SES) level, believed that OMB was paying attention to their agencies' efforts under GPRA. Moreover, there was no corresponding increase in their concern that OMB would micromanage the programs in their agencies. In our survey, we asked respondents to assess the extent to which OMB pays attention to their agencies' efforts under GPRA. As seen in figure 6, in 2003, the percentage of respondents who responded "Great" or "Very Great" to this question (31 percent) was significantly higher than in 2000 (22 percent). Of those, SES respondents showed an even more dramatic increase, from 33 to 51 percent. We also asked respondents the extent to which their concern that OMB would micromanage programs in their agencies was a hindrance to measuring performance or using performance information. The percentage among those expressing an opinion that it was a hindrance to a great or very great extent was low--around 24 percent in 2003--with no significant difference between 2000 and 2003. Figure 6: Percentage of Federal Managers and SES Managers Who Reported That OMB Paid Attention to Their Agency's Efforts under GPRA to a Great or Very Great Extent: [See PDF for image] [A] There was a statistically significant difference between 2000 and 2003 surveys. [End of figure] GPRA Has Begun to Establish a Link between Resources and Results: Among its major purposes, GPRA aims for a closer and clearer linkage between requested resources and expected results. The general concept of linking performance information with budget requests is commonly known as performance budgeting. Budgeting is and will remain an exercise in political choice, in which performance can be one, but not necessarily the only, factor underlying decisions. However, efforts to infuse performance information into resource allocation decisions can more explicitly inform budget discussions and focus them--both in Congress and in agencies--on expected results, rather than on inputs. GPRA established a basic foundation for performance budgeting by requiring that an agency's annual performance plan cover each program activity in the President's budget request for that agency. GPRA does not specify any level of detail or required components needed to achieve this coverage. Further, GPRA recognizes that agencies' program activity structures are often inconsistent across budget accounts and thus gives agencies the flexibility to consolidate, aggregate, or disaggregate program activities, so long as no major function or operation of the agency is omitted or minimized. In addition, OMB guidance has traditionally required agencies to display, by budget program activity, the funding level being applied to achieve performance goals. OMB's guidance on developing fiscal year 2005 performance budgets also encourages a greater link between performance and funding levels, however, it places greater emphasis on linking agencies' long-term and annual performance goals to individual programs. At a minimum, agencies are to align resources at the program level, but they are encouraged to align resources at the performance goal level. Resources requested for each program are to be the amounts needed to achieve program performance goal targets. Our 1998 assessment of fiscal year 1999 performance plans found that agencies generally covered the program activities in their budgets, but most plans did not identify how the funding for those program activities would be allocated to performance goals.[Footnote 19] However, our subsequent reviews of performance plans indicate that agencies have made progress in demonstrating how their performance goals and objectives relate to program activities in the budget. Over the first 4 years of agency efforts to implement GPRA, we observed that agencies continued to tighten the required link between their performance plans and budget requests.[Footnote 20] Of the agencies we reviewed over this period, all but three met the basic requirement of GPRA to define a link between their performance plans and the program activities in their budget requests, and most of the agencies in our review had moved beyond this basic requirement to indicate some level of funding associated with expected performance described in the plan. Most importantly, more of the agencies we reviewed each year--almost 75 percent in fiscal year 2002 compared to 40 percent in fiscal year 1999- -were able to show a direct link between expected performance and requested program activity funding levels--the first step in defining the performance consequences of budgetary decisions. However, we have also observed that the nature of these linkages varied considerably. Most of the agencies in our review of fiscal year 2002 performance plans associated funding requests with higher, more general levels of expected performance, rather than the more detailed "performance goals or sets of performance goals" suggested in OMB guidance. Although not cited by our group of experts, participants at six of our seven focus groups with federal managers cited progress in this area as a key accomplishment of GPRA. However, the participants also commented that much remains to be done in this area. The comments ranged from the general--GPRA provides a framework for planning and budgeting, to the more specific--GPRA created a definition of programs and how they will help the agency achieve its goals/objectives and the amount of money that will be required to achieve said goals/objectives. One of the comments implied that GPRA has helped to prioritize agency efforts by helping agencies align their efforts with programs or activities that make a difference. A political appointee we interviewed echoed this comment, stating that GPRA was pushing the department to think about what it gets out of the budget, not just what it puts into it--12 to 15 years ago the "so what" was missing from the budget process. Another political appointee we interviewed stated that the department was in the process of tying its goals to its budget formulation and execution processes and linking program costs to departmental goals. A former political appointee discussed how his department used program performance information to inform a major information systems investment decision. Furthermore, GAO case studies on the integration of performance information in budget decision making found that performance information has been used to inform the allocation of resources and for other management purposes at selected agencies. For example, the Veterans Health Administration provides its health care networks with performance information on patterns of patient care and patient health outcomes, which can be used to analyze resource allocation and costs and reallocate resources as appropriate.[Footnote 21] Officials at the Administration for Children and Families said that training and technical assistance and salaries and expense funds are often allocated based on program and performance needs.[Footnote 22] The Nuclear Regulatory Commission monitors performance against targets and makes resource adjustments, if needed, to achieve those targets.[Footnote 23] Although there has been progress in formally establishing the linkages between budgets and plans, our survey results are somewhat conflicting and have not reflected any notable changes either in managers' perceptions governmentwide as to their personal use of plans or performance information when allocating resources, or in their perceptions about the use of performance information when funding decisions are made about their programs. Our 2003 survey data show that a large majority of federal managers reported that they consider their agency's strategic goals when they are allocating resources. As shown in figure 7, on our 2003 survey, an estimated 70 percent of all federal managers agreed to a great or very great extent that they considered their agency's strategic goals when allocating resources. However, using our 1997 survey responses as a baseline, it was not a statistically significant increase over 64 percent of the managers who responded comparably then. As shown in figure 8, a similar, but somewhat smaller, majority (60 percent) of managers who expressed an opinion on our 2003 survey agreed to a great or very great extent that they used information from performance measurement when they were involved in allocating resources. In 1997, the comparable response was about the same at 62 percent. When we asked managers on another item, however, about the extent to which they perceived funding decisions for their programs being based on results or outcome-oriented performance information, only 25 percent of federal managers in 2003 endorsed this view to a great or very great extent. In 1997, 20 percent of managers expressed a comparable view, again not a significant increase. (See fig. 9.): Figure 7: Percentage of Federal Managers Who Reported They Considered Strategic Goals to a Great or Very Great Extent When Allocating Resources: [See PDF for image] [End of figure] Figure 8: Percentage of Federal Managers Who Reported They Considered Performance Information to a Great or Very Great Extent When Allocating Resources: [See PDF for image] Note: Percentages are based on those respondents answering on the extent scale. [End of figure] Figure 9: Percentage of Federal Managers Who Reported That Funding Decisions Were Based on Results or Outcome-Oriented Performance Information to a Great or Very Great Extent: [See PDF for image] [End of figure] [End of section] Chapter 3: Agencies Have Addressed Many Critical Performance Planning and Reporting Challenges, but Weaknesses Persist: [End of section] Beginning with federal agencies' initial efforts to develop effective strategic plans in 1997 and annual performance plans and reports for fiscal year 1999, Congress, GAO, and others have commented on the quality of those efforts and provided constructive feedback on how agency plans and reports could be improved. On the basis of our current review of the strategic plans, annual performance plans, and annual performance reports of six selected agencies--Education, DOE, HUD, DOT, SBA, and SSA--we found that these documents reflect much of the feedback that was provided. For example, goals were more quantifiable and results oriented, and agencies were providing more information about goals and strategies to address performance and accountability challenges and the limitations to their performance data. However, certain weaknesses, such as lack of detail on how annual performance goals relate to strategic goals and how agencies are coordinating with other entities to achieve common objectives, persist. A detailed discussion of our scope and methodology and the results of our reviews of the six agencies' most recent strategic plans, annual performance plans, and annual performance reports compared to initial efforts are contained in appendixes III, IV, and V, respectively. Quality of Selected Strategic Plans Reflects Improvements over Initial Drafts: Under GPRA, strategic plans are the starting point and basic underpinning for results-oriented management. GPRA requires that an agency's strategic plan contain six key elements: (1) a comprehensive agency mission statement; (2) agencywide long-term goals and objectives for all major functions and operations; (3) approaches (or strategies) and the various resources needed to achieve the goals and objectives; (4) a description of the relationship between the long-term goals and objectives and the annual performance goals; (5) an identification of key factors, external to the agency and beyond its control, that could significantly affect the achievement of the strategic goals; and (6) a description of how program evaluations were used to establish or revise strategic goals and a schedule for future program evaluations. Our 1997 review of agencies' draft strategic plans found that a significant amount of work remained to be done by executive branch agencies if their strategic plans were to fulfill the requirements of GPRA, serve as a basis for guiding agencies, and help congressional and other policymakers make decisions about activities and programs.[Footnote 24] Our assessment of 27 agencies' initial draft strategic plans revealed several critical strategic planning issues that needed to be addressed. These planning issues were as follows: * Most of the draft plans did not adequately link required elements in the plans, such as strategic goals to annual performance goals. * Long-term strategic goals often tended to have weaknesses. * Many agencies did not fully develop strategies explaining how their long-term strategic goals would be achieved. * Most agencies did not reflect in their draft plans the identification and planned coordination of activities and programs that cut across multiple agencies. * The draft strategic plans did not adequately address program evaluations. We noted that Congress anticipated that it may take several planning cycles to perfect the process and that strategic plans would be continually refined as various planning cycles occur. We also recognized that developing a strategic plan is a dynamic process and that agencies, with input from OMB and Congress, were continuing to improve their plans. Agencies have now had 6 years to refine their strategic planning processes. Although the six strategic plans we looked at for this review reflected many new and continuing strengths as well as improvements over the 1997 initial drafts, we continued to find certain persistent weaknesses. As depicted in table 1, of the six elements required by GPRA, the plans generally discussed all but one--program evaluation, an area in which we have found capacity is often lacking in federal agencies.[Footnote 25] Although the strategic plans generally listed the program evaluations agencies planned to complete over the planning period, they generally did not address how the agencies planned to use their evaluations to establish new or revise existing strategic goals, as envisioned by GPRA. Finally, although not required by GPRA, the strategic plans would have benefited from more complete discussions of how agencies planned to coordinate with other entities to address common challenges or achieve common or complementary goals. Appendix III provides a more detailed discussion of (1) the required and other useful elements we reviewed to assess strategic plan strengths and weaknesses and (2) changes in the quality of the six agencies' strategic plans we reviewed. Table 1: Agencies' Progress in Addressing Required Elements of Strategic Planning under GPRA: Agency strategic plans: Department of Education; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Education; Plan year: 2002; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Energy; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Energy; Plan year: 2003[A]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Housing and Urban Development; Plan year: 1997; Element included in agency strategic plan? Mission statement: No; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: No; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Housing and Urban Development; Plan year: 2003; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long- term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Small Business Administration; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Small Business Administration; Plan year: 2001[B]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long- term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Social Security Administration; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Social Security Administration; Plan year: 2003; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Transportation; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: No; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Transportation; Plan year: Plan year: 2003[A]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Sources: GAO analysis of agencies' strategic plans in effect at the time of our review. See also, U.S. General Accounting Office, The Results Act: Observations on the Department of Education's June 1997 Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R] (Washington, D.C.: July 18, 1997); Results Act: Observations on DOE's Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R] (Washington, D.C.: July 11, 1997); The Results Act: Observations on the Department of Transportation's Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R] (Washington, D.C.: July 30, 1997); The Results Act: Observations on the Social Security Administration's June 1997 Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R] (Washington, D.C.: July 22, 1997); The Results Act: Observations on the Small Business Administration's Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R] (Washington, D.C.: July 11, 1997); The Results Act: Observations on the Department of Housing and Urban Development's Draft Strategic Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R] (Washington, D.C.: Aug. 8, 1997). [A] The 2003 plans for DOE and DOT were in draft form during the time of our review. [B] At the time of our review, the most recent SBA strategic plan was for fiscal years 2001-2008. SBA released a new strategic plan for fiscal years 2003-2008 in October 2003. [End of table] Strategic Planning Strengths and Improvements from Initial Draft Plans: Consistent with our review of agencies' 1997 strategic plans, the recent strategic plans we reviewed generally contained mission statements that were results oriented, distinct from other agencies, and covered the agencies' major activities. DOT's mission statement had improved by reflecting additional language from its enabling legislation that we recommended adding during our 1997 review. Still improvement could be made in this area as is shown by DOE's mission statement. DOE's mission was results oriented but did not address the department's activities related to energy supply and conservation. Our review of the current strategic plans also revealed improvements in the development of agencies' long-term, strategic goals--essential for results-oriented management. Although GPRA does not require that all of an agency's long-term, strategic goals be results oriented, the intent of GPRA is to have agencies focus their strategic goals on results to the extent feasible. In addition, as required by GPRA, the goals should be expressed in a manner that could be used to gauge success in the future and should cover an agency's major functions or activities. All of the strategic plans we reviewed contained long-term, strategic goals that demonstrated improvements in the quality of their 1997 goals. Agencies' long-term strategic goals generally covered their missions, were results oriented, and were expressed in a manner that could be used to gauge future success. For example, SBA improved the quality of its long-term goals by focusing more on key outcomes to be achieved and less on process improvements, as was the case in its 1997 plan. In some cases, we observed strategic goals that addressed the agency's organizational capacity to achieve results, such as SSA's long-term goal to strategically manage and align staff to support its mission. We also found improvements in how agencies' current plans addressed performance and accountability challenges we had identified, a key weakness we identified in our earlier review. Each of the agency plans we reviewed discussed the long-term goals and strategies to address the challenges that we had identified. For example, Education's strategic plan contained a long-term strategic goal to modernize the Federal Student Assistance programs and address identified problems in this area, which we have designated as high risk since 1990.[Footnote 26] SSA noted that it considered GAO-identified performance and accountability challenges when it determined its strategic goals and objectives, however not all of the challenges are clearly addressed in the plan. A third area of improvement we observed was in the description of the strategies agencies planned to use to achieve their long-term strategic goals. In our review of agencies' 1997 draft strategic plans, we found that many agencies did not fully develop strategies explaining how their long-term strategic goals would be achieved. In contrast, all six of the current strategic plans we reviewed contained strategies that appeared logically linked to achieving the agencies' long-term goals. Other strengths and improvements we observed in meeting GPRA's basic requirements involved the reporting of external factors that could affect the achievement of the long-term goals and the identification of crosscutting activities, although as indicated below these discussions could be improved. The six agencies reviewed for this report each reported on external factors in current strategic plans. For example, for each of the strategic objectives in DOT's strategic plan, DOT lists factors external to its control and how those factors could affect the achievement of its objectives. Although not a requirement, some of the better plans we reviewed discussed strategies to ameliorate the effect of external factors. For example, for an external factor on teacher certification under a goal on reading, Education's plan states that the agency "will work with the states and national accreditation bodies to encourage the incorporation of research-based reading instruction into teacher certification requirements.": We have frequently reported that a focus on results, as envisioned by GPRA, implies that federal programs contributing to the same or similar results should be closely coordinated to ensure that goals are consistent and, as appropriate, program efforts are mutually reinforcing. This means that federal agencies are to look beyond their organizational boundaries and coordinate with other agencies to ensure that their efforts are aligned. During our 1997 review, we found that most agencies did not reflect in their draft plans the identification and planned coordination of activities and programs that cut across multiple agencies. In contrast, each of the six current agency strategic plans that we reviewed identified at least some activities and programs that cut across multiple agencies. For example, SBA's 1997 plan contained no evidence of how the agency coordinated with other agencies, but the current plan contained a separate section describing crosscutting issues in the areas of innovation and research assistance, international trade assistance, business development assistance, veterans affairs, and disaster assistance. Critical Strategic Planning Issues Needing Further Improvement: First, consistent with our 1997 review, the strategic plans we reviewed did not adequately link required elements in the plans. Although all of the agencies we reviewed provided some information on the relationship between their long-term and annual goals, the extent of information provided on how annual goals would be used to measure progress in achieving the long-term goals varied greatly. In the case of DOE, the plan provides a very brief description of the overall relationship between its long-term and annual goals with examples, but does not demonstrate how it will assess progress for each of its long-term goals and objectives. Another plan, DOT's, refers the reader to the annual performance plan for information about annual goals. We have reported that this linkage is critical for determining whether an agency has a clear sense of how it will assess progress toward achieving its intended results. Second, although the agencies' descriptions of their strategies had improved since our initial reviews, with few exceptions, their strategies generally did not include information on how the agencies plan to align their activities, core processes, human capital, and other resources to support their mission-critical outcomes and whether they have the right mix of activities, skills, and resources to achieve their goals. Such information is critical to understanding the viability of the strategies. Furthermore, none of the agencies discussed alternative strategies they had considered in developing their plans. Without such discussions, it is unclear whether agency planning processes were truly strategic or simply a recasting of existing activities, processes, etc. HUD was the only agency that provided any details of how it intended to coordinate with other agencies to achieve common or complementary goals for its crosscutting programs or activities. For example, to support its goal of "Equal Opportunity in Housing," HUD's plan states that HUD and the Department of Justice continue to coordinate their fair housing enforcement activities, especially with respect to responding quickly and effectively to Fair Housing Act complaints that involve criminal activity (e.g., hate crimes), a pattern and practice of housing discrimination, or the legality of state and local zoning or other land use laws or ordinances. We have reported that mission fragmentation and program overlap are widespread throughout the federal government.[Footnote 27] As such, interagency coordination is important for ensuring that crosscutting programs are mutually reinforcing and efficiently implemented. Finally, the draft strategic plans did not adequately address program evaluations. In combination with an agency's performance measurement system, program evaluations can provide feedback to the agency on how well its activities and programs contributed to achieving strategic goals. For example, evaluations can be a potentially critical source of information for Congress and others in assessing (1) the appropriateness and reasonableness of goals; (2) the effectiveness of strategies by supplementing performance measurement data with impact evaluation studies; and (3) the implementation of programs, such as identifying the need for corrective action. Evaluations are important because they potentially can be critical sources of information for ensuring that goals are reasonable, strategies for achieving goals are effective, and that corrective actions are taken in program implementation. Five out of the six current plans that we reviewed included a discussion of program evaluations, however for most of these plans the discussions lacked critical information required by GPRA, such as a discussion of how evaluations were used to establish strategic goals or a schedule of future evaluations. For example, DOE's plan stated that internal, GAO, and Inspector General (IG) evaluations were used as resources to develop its draft strategic plan, but specific program evaluations were not identified. Fiscal Year 2004 Annual Performance Plans Addressed Some Weaknesses of Earlier Plans, but Still Have Room for Significant Improvement: According to our review of agencies' first annual performance plans, which presented agencies' annual performance goals for fiscal year 1999,[Footnote 28] we found that substantial further development was needed for these plans to be useful in a significant way to congressional and other decision makers. Most of the fiscal year 1999 plans that we reviewed contained major weaknesses that undermined their usefulness in that they (1) did not consistently provide clear pictures of agencies' intended performance, (2) generally did not relate strategies and resources to performance, and (3) provided limited confidence that agencies' performance data will be sufficiently credible. Although all of the fiscal year 1999 plans contained valuable information for decision makers, their weaknesses caused their usefulness to vary considerably within and among plans. As shown in table 2, our current review of agencies' fiscal year 2004 performance plans found that five agencies--Education, HUD, SBA, SSA, and DOT--improved their efforts to provide a clear picture of intended performance, with SSA and DOT being the clearest. Furthermore, the same five agencies improved the specificity of the strategies and resources they intended to use to achieve their performance goals, with DOT being the most specific. Finally, the same five agencies--Education, HUD, SBA, SSA, and DOT--made improvements in the area of greatest weakness- -reporting on how they will ensure performance data will be credible. However, only DOT's plan provided a full level of confidence that the performance data the agency intended to collect would be credible. Appendix IV provides a more detailed discussion of (1) the required and other useful elements we reviewed to assess the clarity of the picture of intended performance, the specificity of the strategies and resources, and the level of confidence in the performance data and (2) changes in the quality of the six agencies' annual performance plans we reviewed. Table 2: Characterizations of Agencies' Fiscal Year 1999 and 2004 Annual Performance Plans: Agency: Department of Education; Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Picture of intended performance: (unclear, limited, general, clear): 2004: General; Strategies and resources: (no, limited, general, specific): 1999: Limited; Strategies and resources: (no, limited, general, specific): 2004: General; Data credible: (no, limited, general, full): 1999: Limited; Data credible: (no, limited, general, full): 2004: General. Agency: Department of Energy; Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Picture of intended performance: (unclear, limited, general, clear): 2004: Limited; Strategies and resources: (no, limited, general, specific): 1999: General; Strategies and resources: (no, limited, general, specific): 2004: General; Data credible: (no, limited, general, full): 1999: Limited; Data credible: (no, limited, general, full): 2004: Limited. Agency: Department of Housing and Urban Development; Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Picture of intended performance: (unclear, limited, general, clear): 2004: General; Strategies and resources: (no, limited, general, specific): 1999: Limited; Strategies and resources: (no, limited, general, specific): 2004: General; Data credible: (no, limited, general, full): 1999: Limited; Data credible: (no, limited, general, full): 2004: General. Agency: Small Business Administration; Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Picture of intended performance: (unclear, limited, general, clear): 2004: General; Strategies and resources: (no, limited, general, specific): 1999: Limited; Strategies and resources: (no, limited, general, specific): 2004: General; Data credible: (no, limited, general, full): 1999: Limited; Data credible: (no, limited, general, full): 2004: General. Agency: Social Security Administration; Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Picture of intended performance: (unclear, limited, general, clear): 2004: Clear; Strategies and resources: (no, limited, general, specific): 1999: Limited; Strategies and resources: (no, limited, general, specific): 2004: General; Data credible: (no, limited, general, full): 1999: No; Data credible: (no, limited, general, full): 2004: General. Agency: Department of Transportation; Picture of intended performance: (unclear, limited, general, clear): 1999: General; Picture of intended performance: (unclear, limited, general, clear): 2004: Clear; Strategies and resources: (no, limited, general, specific): 1999: General; Strategies and resources: (no, limited, general, specific): 2004: Specific; Data credible: (no, limited, general, full): 1999: Limited; Data credible: (no, limited, general, full): 2004: Full. Sources: GAO analysis of agencies' fiscal year 2004 annual performance plans and U.S. General Accounting Office, Results Act: Observations on the Department of Education's Fiscal Year 1999 Annual Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-172R] (Washington, D.C.: June 8, 1998); Results Act: Observations on DOE's Annual Performance Plan for Fiscal Year 1999, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-194R] (Washington, D.C.: May 28, 1998); Results Act: Observations on the Department of Housing and Urban Development's Fiscal Year 1999 Annual Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-159R] (Washington, D.C.: June 5, 1998); Results Act: Observations on the Small Business Administration's Fiscal Year 1999 Annual Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98- 200R] (Washington, D.C.: May 28, 1998); The Results Act: Observations on the Social Security Administration's Fiscal Year 1999 Annual Performance Plan, [Hyperlink, http://www.gao.gov/cgi- bin/getrpt?GAO/HEHS-98-178R] (Washington, D.C.: June 9, 1998); Results Act: Observations on the Department of Transportation's Annual Performance Plan for Fiscal Year 1999, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-180R] (Washington, D.C.: May 12, 1998). [End of table] Plans Generally Provided a Clearer Picture of Intended Performance, Except for Crosscutting Areas: At the most basic level, an annual performance plan is to provide a clear picture of intended performance across the agency. Such information is important to Congress, agency managers, and others for understanding what the agency is trying to achieve, identifying subsequent opportunities for improvement, and assigning accountability. Our current review of agencies' fiscal year 2004 performance plans found that five of the six agencies provided a clearer picture of intended performance than their fiscal year 1999 plans did, although only two of the 2004 plans--DOT's and SSA's-- received the highest rating possible. As shown in table 2, except for DOT, the six agencies we reviewed for this report initially provided a limited picture of intended performance. Most of the fiscal year 1999 performance plans we previously reviewed had at least some objective, quantifiable, and measurable goals, but few plans consistently included a comprehensive set of goals that focused on the results that programs were intended to achieve. Moreover, agencies did not consistently follow OMB's guidance that goals for performance and accountability challenges be included in the plans. Agencies' plans generally showed how their missions and strategic goals were related to their annual performance goals and covered all of the program activities in the agencies' budget requests.[Footnote 29] In addition, many agencies took the needed first step of identifying their crosscutting efforts, with some including helpful lists of other agencies with which they shared a responsibility for addressing similar national issues. However, the plans generally did not go further to describe how agencies expected to coordinate their efforts with other agencies. The fiscal year 2004 plans improved the picture of performance by making annual goals and performance measures more results oriented, objective, and quantifiable. For example, Education's plan included a measure for the number of states meeting their eighth grade mathematics achievement targets under the long-term goal to improve mathematics and science achievement for all students. We previously criticized Education's 1999 plan for lacking such outcome-oriented measures. Another overall improvement we observed was that all of the plans described intended efforts to address performance and accountability challenges we and others had previously identified. For instance, to address the governmentwide high-risk area of strategic human capital management, HUD states that to develop its staff capacity, it will complete a comprehensive workforce analysis in 2004 to serve as the basis to fill mission critical skill gaps through succession planning, hiring, and training initiatives in a 5-year human capital management strategy. The clarity of DOE's plan remained limited because its annual goals were not clearly linked to its mission, the long-term goals in its strategic plan, or the program activities in its budget request. Although five of the six agencies improved the clarity of the picture of intended performance, improvement is still needed in reporting on crosscutting efforts. In both the 1999 and 2004 plans, many agencies identified their crosscutting efforts, with some including helpful lists of other agencies with which they shared a responsibility for addressing similar national issues. Our review of fiscal year 2004 plans shows that the six agencies we reviewed still did not discuss how they expected to coordinate with other agencies to address common challenges or to achieve common or complementary performance goals. As we have reported previously, improved reporting on crosscutting efforts can help Congress use the annual performance plan to evaluate whether the annual goals will put the agency on a path toward achieving its mission and long-term strategic goals. In addition, the plans can aid in determining efforts to reduce significant program overlap and fragmentation that can waste scarce resources, confuse and frustrate program customers, and limit overall program effectiveness. None of the six agencies' plans indicated an intention to request waivers of specific administrative procedural requirements and controls that may be impeding an agencies' ability to achieve results. This provision of GPRA allows agencies greater managerial flexibility in exchange for accountability for results. We previously reported on the results of the pilot project to implement this provision of GPRA and found that the pilot did not work as intended.[Footnote 30] OMB did not designate any of the seven departments and one independent agency that submitted a total of 61 waiver proposals as GPRA managerial accountability and flexibility pilots. For about three-quarters of the waiver proposals, OMB or other central management agencies determined that the waivers were not allowable for statutory or other reasons or that the requirement for which the waivers were proposed no longer existed. For the remaining proposals, OMB or other central management agencies approved waivers or developed compromises by using authorities that were already available independent of GPRA. Plans More Specifically Related Strategies and Resources to Performance Goals: To judge the reasonableness of an agency's proposed strategies and resources, congressional and other decision makers need complete information on how the proposed strategies and resources will contribute to the achievement of agency goals. Agencies generally improved their plans by better relating strategies and resources to performance. Education's, HUD's, SBA's, and SSA's 1999 plans had a limited discussion, while DOE's and DOT's 1999 plans had a general discussion. In 2004, five of the six plans--Education's, DOE's, HUD's, SBA's, and SSA's--provided general discussions of how their strategies and resources would contribute to achieving their performance goals. DOT's 2004 plan improved to include a specific discussion. Our review of the 1999 plans found that most agencies' performance plans did not provide clear strategies that described how performance goals would be achieved. In contrast, the 2004 performance plans we reviewed generally provided lists of the agencies' current array of programs and initiatives. Several plans provided a perspective on how these programs and initiatives were necessary or helpful for achieving results. For example, DOE and HUD included in their plans a "means and strategies" section for each of their goals that described how the goal would be achieved. One strategy DOE identified to meet its goal of contributing unique, vital facilities to the biological environmental sciences was to conduct peer reviews of the facilities to assess the scientific output, user satisfaction, the overall cost-effectiveness of each facility's operations, and their ability to deliver the most advanced scientific capability. In addition, each of the agencies' plans identified the external factors that could influence the degree to which goals are achieved. Some of the better plans, such as DOT's and SBA's, provided strategies to mitigate the negative factors or take advantage of positive factors, as appropriate. For example, for its transportation accessibility goals, DOT's plan states that as the population ages, more people will require accessible public transit, for which states and local agencies decide how best to allocate federally provided resources. One of the strategies DOT intends to employ to address this external factor is the "Special Needs of Elderly Individuals and Individuals with Disabilities" grant program. The plan states the grant program will help meet transportation needs of the elderly and persons with disabilities when regular transportation services are unavailable, insufficient, or inappropriate to meet their needs. Agencies' 2004 plans did not consistently describe all the resources needed and how they would be used to achieve agency goals. Our review of agencies' fiscal year 1999 plans found that most did not adequately describe--or reference other appropriate documents that describe--the capital, human, information, and financial resources needed to achieve their agencies' performance goals. The 2004 plans we reviewed generally described the funding levels needed to achieve their performance goals overall and in some cases broke out funding needs by high-level performance goal. For example, SSA's plan provides a general perspective on the financial resources needed to achieve its performance goals because it provides budget information by account and program activity. However, the plan is neither structured by budget program activity or account, nor does it provide a crosswalk between the strategic goals and budget program accounts. In contrast, HUD's plan presented its requested funding and staffing levels at the strategic goal level, but did not present budget information at the level of its annual goals. In addition, although the plans make brief mention of nonfinancial resources, such as human capital, information technology, or other capital investments, little information is provided on how such resources would be used to achieve performance goals. Plans Continue to Provide Less Than Full Confidence That Performance Data Will Be Credible: Credible performance information is essential for accurately assessing agencies' progress towards the achievement of their goals and, in cases where goals are not met, identifying opportunities for improvement or whether goals need to be adjusted. Under GPRA, agencies' annual performance plans are to describe the means that will be used to verify and validate performance data. To help improve the quality of agencies' performance data, Congress amended GPRA through the Reports Consolidation Act of 2000 to require that agencies assess the completeness and reliability of the performance data in their performance reports. Agencies were also required to discuss in their report any material inadequacies in the completeness and reliability of their performance data and discuss actions to address these inadequacies. Meeting these new requirements suggests the need for careful planning to ensure that agencies can comment accurately on the quality of the performance data they report to the public. As shown in table 2, although five of the six agencies we reviewed improved in reporting how they plan to ensure that performance data will be credible, only one agency--DOT--improved enough over its 1999 plan to provide a full level of confidence in the credibility of its performance data. Four agencies--Education, HUD, SBA, and SSA--improved enough to provide a general level of confidence. However, DOE provided the same limited level of confidence in the credibility of the performance data as in its 1999 plan. Regarding all 24 of the fiscal year 1999 performance plans we reviewed, we found most provided only superficial descriptions of procedures that agencies intended to use to verify and validate performance data. Moreover, in general, agencies' performance plans did not include discussions of documented limitations in financial and other information systems that may undermine efforts to produce high-quality data. As we have previously noted, without such information, and strategies to address those limitations, Congress and other decision makers cannot assess the validity and reliability of performance information. We found that each of the 2004 plans we reviewed contained some discussion of the procedures the agencies would use to verify and validate performance information, although in some cases the discussion was inconsistent or limited. For example, the discussions of SBA's verification and validation processes for its indicators in the 2004 plan were generally one-or two-sentence statements. SBA also noted that it does not independently verify some of the external data it gathers or that it does not have access to the data for this purpose. In contrast, the DOT plan referred to a separate compendium available on- line that provides source and accuracy statements, which give more detail on the methods used to collect performance data, sources of variation and bias in the data, and methods used to verify and validate the data. In addition, all of the agencies except DOE discussed known limitations to performance data in their plans. These agencies' plans generally provided information about the quality of each performance measure, including any limitations. According to DOE officials, DOE's plan generally does not discuss data limitations because the department selected goals for which data are expected to be available and therefore did not anticipate finding any limitations. However, in our 2003 Performance and Accountability Series report on DOE, we identified several performance and accountability challenges where data were a concern, such as the need for additional information on the results of contractors' performance to keep projects on schedule and within budget.[Footnote 31] DOE's contract management continues to be a significant challenge for the department and remains at high risk. Finally, the remaining five agencies also discussed plans to address limitations to the performance data. For example, DOT's plan provided a general discussion of the limitations to the internal and external sources of data used to measure performance. Detailed discussions were contained in an appendix to the plan and separate source and accuracy statements. This information had been lacking in its 1999 plan. Education, HUD, SBA, and SSA also provided information on limitations to their performance data and plans for improvement. Strengths and Weaknesses of Selected Agencies' Fiscal Year 2002 Annual Performance Reports: Key to improving accountability for results as Congress intended under GPRA, annual performance reports are to document the results agencies have achieved compared to the goals they established. To be useful for oversight and accountability purposes, the reports should clearly communicate performance results, provide explanations for any unmet goals as well as actions needed to address them, and discuss known data limitations as well as how the limitations are to be addressed in the future. Compared to our reviews of the six agencies' fiscal year 1999 performance reports, we identified a number of strengths and improvements as well as areas that continued to need improvement. Because the scope of our review of the fiscal year 2002 reports was broader than that for the fiscal year 1999 reports we previously reviewed, we were unable to make specific comparisons for the three characteristics we used to assess the fiscal year 2002 reports. However, we discuss comparative information on aspects of the reports where available. Table 3 shows the results of our assessment of the six agencies' annual performance reports for fiscal year 2002. Appendix V provides a more detailed discussion of (1) the required and other useful elements we reviewed to assess the clarity of the picture of performance, the clarity of the linkage between costs and performance, and the level of confidence in the performance data and (2) changes in the quality of the six agencies' annual performance plans we reviewed. Table 3: Characterizations of Agencies' 2002 Annual Performance Reports: Agency: Department of Education; Picture of performance: (unclear, limited, general, clear): Limited; Resources linked to results: (no, limited, general, clear): Clear; Data credible: (no, limited, general, full): General. Agency: Department of Energy; Picture of performance: (unclear, limited, general, clear): General; Resources linked to results: (no, limited, general, clear): Limited; Data credible: (no, limited, general, full): Limited. Agency: Department of Housing and Urban Development; Picture of performance: (unclear, limited, general, clear): General; Resources linked to results: (no, limited, general, clear): No; Data credible: (no, limited, general, full): General. Agency: Small Business Administration; Picture of performance: (unclear, limited, general, clear): Limited; Resources linked to results: (no, limited, general, clear): General; Data credible: (no, limited, general, full): General. Agency: Social Security Administration; Picture of performance: (unclear, limited, general, clear): General; Resources linked to results: (no, limited, general, clear): Limited; Data credible: (no, limited, general, full): General. Agency: Department of Transportation; Picture of performance: (unclear, limited, general, clear): General; Resources linked to results: (no, limited, general, clear): No; Data credible: (no, limited, general, full): Full. Sources: GAO analysis of agencies' fiscal year 2002 annual performance reports and U.S. General Accounting Office, Observations on the Department of Education's Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-00-128R]; (Washington, D.C. June 30, 2000); Observations on the Department of Energy's Fiscal Year 1999 Accountability Report and Fiscal Years 2000 and 2001 Performance Plans, GAO/RCED-00-209R; (Washington, D.C. June 30, 2000); Observations on the Department of Housing and Urban Development's Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00- 211R]; (Washington, D.C. June 30, 2000); Observations on the Small Business Administration's Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00-207R]; (Washington, D.C. June 30, 2000); Observations on the Social Security Administration's Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-00-126R]; (Washington, D.C. June 30, 2000); and Observations on the Department of Transportation's Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance Plan, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00-201R]; (Washington, D.C. June 30, 2000). [End of table] Progress in Providing a Clear Picture of Performance: The six agency reports that we reviewed contained a number of strengths, some of which we can describe as improvements over the reports on fiscal year 1999 performance. A key strength of four of the 2002 reports (Education, HUD, DOT, SSA) was a discussion of the relationship between the strategic plan, performance plan, and performance report. For example, SSA's report identified relevant results that were linked to its strategic objective to deliver "citizen-centered, world-class service," such as maintaining the accuracy, timeliness, and efficiency of service to people applying for its benefit programs. The clarity of the DOE and SBA reports was limited by not clearly relating agency performance results to strategic and annual performance goals. For example, the structure of SBA's report reflected the objectives in its draft 2003 to 2008 strategic plan rather than those in its 2002 performance plan, making it difficult to assess progress against the original 2002 objectives. Furthermore, although there is no "right" number of performance measures to be used to assess progress, a number of the plans allowed for an easier review of results by limiting the overall number of measures presented or by highlighting key performance measures of greatest significance to their programs. For example, SBA discussed a total of 19 performance goals and DOT discussed a total of 40. Although SSA discussed a total of 69 performance goals, the report highlighted its progress in achieving 14 key goals. In contrast, Education, HUD, and DOE presented a total of 120, 184, and 260 measures, respectively. Furthermore, while Education and SSA each provided a table showing progress across all its measures, the other agencies did not provide such summary information. As we found in our earlier reviews, the six agencies' fiscal year 2002 reports generally allowed for an assessment of progress made in achieving agency goals. Some of the reports made this assessment easier than others by providing easy-to-read summary information. For example, SSA provided a table at the beginning of the report that summarized the results for each of its 69 indicators with the following dispositions: met, not met, almost met, and data not yet available. Other reports, such as HUD's, required an extensive review to make this assessment. In addition, to place current performance in context, each of the agencies' reports contained trend information, as required by GPRA, which allowed for comparisons between current year and prior year performance. In addition, the majority of agencies maintained, or demonstrated improvements over, the quality of their 1999 reports in discussing the progress achieved in addressing performance and accountability challenges identified by agency IGs and GAO. For example, SBA's report contained two broad overviews and an appendix describing the status of GAO audits and recommendations, as well as a description of the most serious management challenges SBA faces as identified by the agency's IG. Unfortunately, many of the weaknesses we identified in the agencies' fiscal year 2002 reports were similar to those we found in their fiscal year 1999 reports related to the significant number of performance goals (1) which were not achieved and lacked explanations or plans for achieving the goal in the future and (2) for which performance data were unavailable. Three of the six agencies we reviewed--HUD, SSA, and Transportation--did not consistently report the reasons for not meeting their goals. For example, Transportation provided explanations for only 5 of the 14 goals it did not meet. In addition, similar to our 1999 report findings, three of the six agencies we reviewed--HUD, SBA, and DOT--did not discuss their plans or strategies to achieve unmet goals in the future. For example, HUD reported "substantially meeting" only 47 percent of the performance targets in fiscal year 2002. However, although HUD provides various reasons for not meeting all its targets, it offers no information on plans or time frames to achieve the goals in the future. Finally, we continued to observe a significant number of goals for which performance data were unavailable. For example, performance data for 10 of SBA's 19 performance goals were unavailable. In addition, the majority of the reports we reviewed did not include other GPRA requirements. The reports generally did not evaluate the performance plan for the current year relative to the performance achieved toward the performance goals in the fiscal year covered by the report. The reports also did not discuss the use or effectiveness of any waivers in achieving performance goals. In addition, for two of the agencies--DOE and SBA--program evaluation findings completed during the fiscal year were not summarized. As we have previously noted, such evaluations could help agencies understand the relationship between their activities and the results they hope to achieve. Progress in Linking Resources to Results: Although linking costs to performance goals is not a requirement of GPRA, both GPRA and the CFO Act emphasized the importance of linking program performance information with financial information as a key feature of sound management and an important element in presenting to the public a useful and informative perspective on federal spending. The committee report for GPRA suggested that developing the capacity to relate the level of program activity with program costs, such as cost per unit of result, cost per unit of service, or cost per unit of output, should be a high priority. In our survey of federal managers, this year we asked for the first time the extent to which federal managers had measures of cost-effectiveness for the programs they were involved with. Only 31 percent of federal managers we surveyed reported having such measures to a great or very great extent, lower than any of the other types of measures associated with GPRA we asked about by at least 12 percent (see fig. 3 in ch. 2). Under the PMA, the current administration has set an ambitious agenda for performance budgeting, calling for agencies to better align budgets with performance goals and focus on capturing full budgetary costs and matching those costs with output and outcome goals. All this suggests that agencies will need to develop integrated financial and performance management systems that will enable the reporting of the actual costs associated with performance goals and objectives along with presentations designed to meet other budgetary or financial purposes, such as the accounts and program activities found in the President's Budget and responsibility segments found in financial statements.[Footnote 32] Of the six agencies we reviewed, only Education's report clearly linked its budgetary information to the achievement of its performance goals or objectives. Education's report laid out, using both graphics and text, the estimated appropriations associated with achieving each of its 24 objectives. In addition the report provided the staffing in full-time equivalent employment (FTEs) and an estimate of the funds from salaries and expenses contributing to the support of each of these objectives. SBA's report contained crosswalks that showed the relationship between SBA's strategic goals, outcome goals, performance goals, and programs. Because SBA shows the resources for each program, a reader can infer a relationship between SBA's resources and performance goals. However, the linkage between resources and results would be clearer if results and resources were presented by performance goal as well. SSA provided a limited view of the costs of achieving its performance goals by providing the costs associated with four out of five of its strategic goals.[Footnote 33] However, as reported by the IG, SSA needs to further develop its cost accounting system, which would help link costs to performance.[Footnote 34] DOE also provided a limited view of the costs of achieving its performance goals by organizing its performance information by budget program activity and associated net costs. According to DOE officials, the department plans to link its individual performance measures to the costs of program activities in future reports. Neither HUD nor DOT provided information on the cost of achieving individual performance goals or objectives. Progress in Providing Confidence in the Credibility of Performance Data: To assess the degree to which an agency's report provided full confidence that the agency's performance information was credible, we examined the extent to which the reports discussed the quality of the data presented. As shown in table 3, only DOT's report provided a full level of confidence in the quality of the data. The other agencies provided general or limited confidence in their data. All six agencies in our current review complied with the Reports Consolidation Act of 2000 by including assessments of the completeness and reliability of their performance data in their transmittal letters. In contrast, we found that only 5 of the 24 CFO Act agencies included this information in their fiscal year 2000 performance reports.[Footnote 35] Of the six agencies in our current review, only DOE provided this assessment in its fiscal year 2000 report. For example, the Secretary of DOT stated in the transmittal letter that the 2002 report "contains performance and financial data that are substantially complete and reliable." However, only two of the six agencies also disclosed material inadequacies in the completeness and reliability of their performance data and discussed actions to address the inadequacies in their transmittal letters. For example, SBA stated in its transmittal letter that it is "working to improve the completeness and reliability of the performance data for the advice provided to small business through SBA's resource partners." SBA explained that data for this aspect of its performance are collected through surveys, which are inconsistent and not comparable, and for which client responses are difficult to obtain. SBA stated that it is working to improve the survey instruments it uses to obtain performance data. In addition to the requirements of the Reports Consolidation Act, we have previously reported on other practices that enhance the credibility of performance data that are not specifically required by GPRA. For instance, discussions of standards and methods used by agencies to assess the quality of their performance data in their performance reports provide decision makers greater insight into the quality and value of the performance data. None of the reports explicitly referred to a specific standard they used, however, DOE described its method for assuring data quality. The report states that the heads of DOE's organizational elements certified the accuracy of their performance data. DOE subsequently reviewed the data for quality and completeness. Other useful practices that help foster transparency to the public and assist decision makers in understanding the quality of an agency's data include: (1) discussion of data quality, including known data limitations and actions to address the limitations, and (2) discussion of data verification and validation procedures, including proposals to review data collection and verification and validation procedures. All six agencies' reports described data limitations, although discussions were mostly brief and very high level. One exception was DOT, which directed readers to the DOT Web site to obtain an assessment of the completeness and reliability of its performance data and detailed information on the source, scope, and limitations of the performance data. HUD and SBA also discussed plans for addressing the limitations. For example, HUD stated that to address problems with its indicator on the number of homeowners who have been assisted with the Home Investment Partnership Program (HOME), HUD has established a team of managers, technical staff, and contractors to make a series of improvements to the Integrated Disbursement and Information System beginning in fiscal year 2003 that should reduce the need to clean up the data. Each of the six agencies' reports also discussed the procedures they used to verify and validate their performance data. However, these discussions ranged from the very general description of the DOE method (noted previously), to the very detailed discussions provided by DOT. DOT provides an on-line compendium that discusses the source and accuracy of its data. Furthermore, DOT's 2002 report also describes strategies being undertaken to address the quality of its data. The report states that a DOT intermodal working group addressed data quality issues by developing departmental statistical standards and by updating source and accuracy statements for all of DOT's data programs. The working group also worked to improve quality assurance procedures, evaluate sampling and nonsampling errors, and develop common definitions for data across modes. [End of section] Chapter 4: Challenges to GPRA Implementation Persist: While a great deal of progress has been made in making federal agencies more results oriented, numerous challenges still exist to effective implementation of GPRA. The success of GPRA depends on the commitment of top leadership within agencies, OMB, and Congress. However, according to federal managers surveyed, top leadership commitment to achieving results has not grown significantly since our 1997 survey. Furthermore, although OMB has recently shown an increased commitment to management issues, it significantly reduced its guidance to agencies on GPRA implementation compared to prior years, and it is not clear how the program goals developed through its PART initiative will complement and integrate with the long-term, strategic focus of GPRA. Obtaining leadership commitment to implement a strategic plan depends in part on the usefulness and relevance of agency goals and strategies to agency leaders, Congress, and OMB. However, GPRA's requirement to update agency strategic plans every 3 years is out of sync with presidential and congressional terms and can result in updated plans that do not have the support of top administration leadership and key congressional stakeholders. As noted in chapter 2, more federal managers surveyed reported having results-oriented performance measures for their programs and we would expect to have seen similar increases in the use of this information for program management. However, we did not observe any growth in their reported use of this information for key management activities, such as adopting new program approaches or changing work processes. Additionally, managers noted human capital-related challenges that impede results-oriented management, including a lack of authority and training to carry out GPRA requirements, as well as a lack of recognition for the results achieved. Consistent with our previous work, federal managers in our focus groups reported that significant challenges persist in setting outcome- oriented goals, measuring performance, and collecting useful data. However, our survey data suggested that federal managers do not perceive issues, such as "difficulty distinguishing between the results produced by the program and results caused by other factors" and "difficulty obtaining data in time to be useful," to be substantial hindrances to measuring performance or using performance information. Additionally, mission fragmentation and overlap contribute to difficulties in addressing crosscutting issues, particularly when those issues require a national focus, such as homeland security, drug control, and the environment. GAO has previously reported on a variety of barriers to interagency cooperation, such as conflicting agency missions, jurisdiction issues, and incompatible procedures, data, and processes. We have also reported that OMB could use the provision of GPRA that calls for OMB to develop a governmentwide performance plan to integrate expected agency-level performance. Unfortunately, this provision has not been fully implemented and the federal government lacks a tool, such as a strategic plan, that could provide a framework for a governmentwide reexamination of existing programs, as well as proposals for new programs. Finally, federal managers in our focus groups and political appointees we interviewed believed that Congress does not use performance information to the fullest extent to conduct oversight and to inform appropriations decisions. While there is concern regarding Congress' use of performance information, it is important to make sure that this information is initially useful. As a key user of performance information, Congress needs to be considered a partner in shaping agency goals at the outset. GPRA provides Congress opportunities to influence agency performance goals through the consultation requirement for strategic plans and through Congress' traditional oversight role. Top Leadership Does Not Consistently Show Commitment to Achieving Results: We have previously testified that perhaps the single most important element of successful management improvement initiatives is the demonstrated commitment of top leaders to change.[Footnote 36] This commitment is most prominently shown through the personal involvement of top leaders in developing and directing reform efforts. Organizations that successfully address their long-standing management weaknesses do not "staff out" responsibility for leading change. Top leadership involvement and clear lines of accountability for making management improvements are critical to overcoming organizations' natural resistance to change, marshalling the resources needed in many cases to improve management, and building and maintaining the organizationwide commitment to new ways of doing business. Results from our surveys show that while the majority of managers continue to indicate top leadership demonstrates a strong commitment to achieving results, we have not seen a noteworthy improvement in the percentage of managers expressing this view. From our 1997 survey, we estimated about 57 percent of managers overall reported such commitment to a great or very great extent. On our 2003 survey, 62 percent of managers expressed a comparable view--a higher but not statistically significant increase. (See fig. 10.): Figure 10: Percentage of Federal Managers Who Reported to a Great or Very Great Extent Their Top Leadership Has a Strong Commitment to Achieving Results: [See PDF for image] [End of figure] As shown in figure 11, however, we continued to see a significant difference between the perceptions of SES and non-SES managers on this issue. That is, the percentage of SES managers reporting that top leadership demonstrated strong commitment to a great or very great extent in 2003 was 22 percent higher than for non-SES managers. Figure 11: Percentage of SES and Non-SES Managers Who Reported to a Great or Very Great Extent Their Agency Top Leadership Demonstrated Strong Commitment to Achieving Results: [See PDF for image] [A] There was a statistically significant difference between SES and non-SES. [End of figure] We observed in our 1997 and 2000 reports on governmentwide implementation of GPRA that we would expect to see managers' positive perceptions on items, such as the extent to which top leadership is committed to achieving results, become more prevalent and the gap between SES and non-SES managers begin to narrow as GPRA and related reforms are implemented; however, these changes do not appear to be happening as expected.[Footnote 37] Demonstrating the willingness and ability to make decisions and manage programs based on results and the ability to inspire others to embrace such a model are important indicators of leadership commitment to results-oriented management. However, in both our 1997 and 2000 surveys, only about 16 percent of managers reported that changes by management above their levels to the programs for which they were responsible were based on results or outcome-oriented performance information to a great or very great extent. In our 2003 survey, this indicator increased to 23 percent, a statistically significant increase from prior surveys. Twenty-eight percent of federal managers surveyed who expressed an opinion reported that the lack of ongoing top executive commitment or support for using performance information to make program/funding decisions hindered measuring performance or using performance information to a great or very great extent. Our interviews with 10 top political appointees from the Clinton and current Bush administrations indicated a high level of support and enthusiasm for effectively implementing the principles embodied in GPRA. For example, one appointee noted that GPRA focused senior management on a set of goals and objectives to allow the organization to understand what is important and how to deal with accomplishment at a macro-level, as well as provided a structure for problem solving. Another political appointee noted that GPRA has made it important to look at what you get out of the budget, not just what you put into it, while another concluded that GPRA brought about a fundamental rethinking of how they managed their programs and processes. Such indications of support for GPRA are promising. However, to support the transition to more results-oriented agency cultures, top agency management will need to make a more concerted effort to translate their enthusiasm for GPRA into actions that communicate to employees that top management cares about performance results and uses the information in its decision making. The need for strong, committed leadership extends to OMB as well. OMB has shown a commitment to improving the management of federal programs, both through its leadership in reviewing agency program performance using the PART tool as well as through the PMA, which calls for improved financial performance, strategic human capital management, competitive sourcing, expanded electronic government, and performance budget integration. Using the foundation of information generated by agencies in their strategic plans, annual performance plans, and program performance reports, OMB has used the PART tool to exercise oversight of selected federal programs by assessing program purpose and design, the quality of strategic planning, the quality of program management, and the extent to which programs can demonstrate results. PART provides OMB a lens through which to view performance information for use in the budget formulation process. PART, and OMB's use of performance data in the budget formulation process, potentially can complement GPRA's focus on increasing the supply of credible performance information by promoting the demand for this information in the budget formulation process. As we reported in chapter 2, more federal managers noted that OMB was paying attention to their agencies' efforts under GPRA. (See fig. 6.) Additionally, OMB convened a performance measurement workshop in April 2003 to identify practical strategies for addressing common performance measurement challenges. As a result of this workshop, it produced a paper in June 2003 that included basic performance measurement definitions and concepts and common performance measurement problems that were discussed at the workshop. This was part of OMB's continuing efforts to improve PART as an evaluation tool. However, there are areas where OMB could further enhance its leadership. OMB has stated that the PART exercise presents an opportunity to inform and improve on agency GPRA plans and reports and establish a meaningful, systematic link between GPRA and the budget process. OMB has instructed agencies that, in lieu of a performance plan, they are to submit a performance budget that includes information from the PART assessments, including all performance goals used in the assessment of program performance done under the PART process. The result is that program-specific performance measures developed through the PART review are to substitute for other measures developed by the agency through its strategic planning process. GPRA is a broad legislative framework that was designed to be consultative with Congress and other stakeholders and address the needs of many users of performance information--Congress to provide oversight and inform funding decisions, agency managers to manage programs and make internal resource decisions, and the public to provide greater accountability. Changing agency plans and reports for use in the budget formulation process may not satisfy the needs of these other users. Users other than OMB are not likely to find the information useful unless it is credible and valid for their purposes. PART's program-specific focus may fit with OMB's agency-by-agency budget reviews, but it is not well suited to achieving one of the key purposes of strategic plans--to convey agencywide, long-term goals and objectives for all major functions and operations. PART's focus on program-specific measures does not substitute for the strategic, long-term focus of GPRA on thematic goals and department-and governmentwide crosscutting comparisons. To reach the full potential of performance management, agency planning and reporting documents need to reflect the full array of uses of performance information, which may extend beyond those needed for formulating the President's Budget. However, it is not yet clear whether the results of those reviews, such as changes to agencies' program performance measures, will complement and be integrated with the long-term, strategic goals and objectives agencies have established in consultation with Congress and other stakeholders under GPRA. OMB has not yet clearly articulated how PART is to complement GPRA. Focus group participants suggested that the administration and OMB needed to reinforce GPRA's usefulness as a management tool for agencies. They also emphasized the need for OMB to help agencies understand how to integrate GPRA with other management initiatives, such as PART. As we noted in chapter 3, agencies' plans and reports still suffer from persistent weaknesses and could improve in a number of areas, such as attention to issues that cut across agency lines, and better information about the quality of the data that underlie agency performance goals. However, OMB's July 2003 guidance for the preparation and submission of strategic plans, annual performance plans, and annual performance reports is significantly shorter and less detailed than its 2002 guidance. For example, OMB no longer provides detailed guidance to agencies for the development of performance plan components. OMB's 2002 guidance on the preparation and submission of annual performance plans is approximately 39 pages long; in its 2003 guidance, that discussion spans only 2 pages. The 2003 guidance in this area does not include entire sections found in the 2002 guidance, such as principles for choosing performance goals and indicators for inclusion in the annual plan, types of performance goals, crosscutting programs, and requirements for verifying and validating data. OMB needs to maintain and strengthen its leadership role in working with agencies to help them produce the highest quality GPRA documents through its formal guidance and reviews of strategic plan and report submissions. Focus group participants discussed the need for consistent guidance on how to implement GPRA. Furthermore, there is no evidence that agencies have institutional knowledge of GPRA requirements that would obviate the need for OMB's guidance. New managers will need a consistent resource that provides practical guidance on what agencies need to include in their planning and reporting documents to comply with GPRA and reflect best practices. Consistent, explicit OMB guidance on preparing GPRA documents can help ensure that gains in the quality of GPRA documents are maintained and provide a resource for agencies to make further improvements in those documents. For example, guidance on how to discuss coordination of crosscutting programs or improvements to the credibility of performance data in agency performance plans goes hand-in-hand with OMB's enhanced oversight of agency performance through the PART exercise. The success of GPRA depends on the commitment of top leadership within agencies, OMB, and Congress. Obtaining such leadership commitment depends in part on the usefulness and relevance of agency goals and strategies to these parties. GPRA requires an agency to develop a strategic plan at least every 3 years to cover the following 5-year period. Thus, there have been two required updates of strategic plans since the initial strategic plans were submitted for fiscal year 1997- -fiscal year 2000 and fiscal year 2003. The fiscal year 2000 update occurred the year before a new presidential term began. According to our focus group participants--both the experts and federal managers--it makes little sense to require an update of a strategic plan shortly before a new administration is scheduled to take office. For example, changes in political leadership generally result in a new agenda with new objectives. Such changes force agencies to revise their plans, management initiatives, and strategies, which translates into additional GPRA-related work. A strategic plan that does not reflect the participation and buy-in of top administration leadership and key congressional stakeholders is unlikely to be successfully implemented. Therefore, GPRA's requirement to update agency strategic plans according to a schedule that is out of sync with presidential and congressional terms means that effort may be wasted on plans that lack the support of top leadership. Managers Report Mixed Results in Use of Performance Information: GPRA's usefulness to agency leaders and managers as a tool for management and accountability was cited as a key accomplishment numerous times by focus group participants. However, a number of alternative views indicated use of performance information for key management decisions has been mixed. For example, one participant said they did not believe GPRA has been used as a tool yet, while another participant commented that only better managers take advantage of GPRA as a management tool. According to focus group participants, although many federal managers understand and use results-oriented management concepts in their day-to-day activities, such as strategic planning and performance measurement, they do not always connect these concepts to the requirements of GPRA. This view was strongly supported by our survey results. Prior to mentioning GPRA in our survey, we asked federal managers the extent to which they consider their agency's strategic goals when engaging in key management tasks such as setting program activities, allocating resources, or considering changes in their programs. A relatively high percentage of managers--ranging from 66 to 79 percent--responded to a great or very great extent. However, when we asked similar questions about the extent to which they considered their agency's annual performance goals as set forth in the agency's GPRA annual performance plan for the same activities, the comparable responses were considerably lower, ranging from 22 to 27 percent. Because the benefit of collecting performance information is only fully realized when this information is actually used by managers, we asked them about the extent to which they used the information obtained from measuring performance for various program management activities. As shown in figure 12, for seven of the nine activities we asked about, the majority of managers who expressed an opinion reported using performance information to a great or very great extent in 2003. Across all nine activities, the percentage of managers saying they used performance information to a great or very great extent ranged from 41 percent for developing and managing contracts to 60 percent for allocating resources, setting individual job expectations, and rewarding staff. While we had observed a decline in the reported use of performance information to this extent for many of these activities between 1997 and 2000, our 2003 results increased to levels not significantly different from 1997 for all but one category--adopting new program approaches or changing work processes. This category of use continued to be significantly lower at 56 percent in 2003 than it was in 1997 at 66 percent. Although another category, coordinating program efforts with other internal or external organizations, shows a similar pattern of limited recovery, the difference between the 1997 and 2003 results is not statistically significant. Figure 12: Percentage of Federal Managers Who Reported Using Information Obtained from Performance Measurement to a Great or Very Great Extent for Various Management Activities: [See PDF for image] Note: Percentages are based on those respondents answering on the extent scale. [A] There was a statistically significant difference between the 1997 and 2003 surveys. [B] This question was not asked in 1997. [End of figure] We have reported that involving program managers in the development of performance goals and measures is critical to increasing the relevance and usefulness of this information to their day-to-day activities.[Footnote 38] Yet, our survey data indicate that participation in activities related to the development and use of performance information has also been mixed. In 2003, only 14 percent of managers believed to a great or very great extent that their agencies considered their contributions to or comments on their agency's GPRA plans or reports. However, significantly more SES managers (43 percent) than non-SES managers (12 percent) expressed this view. Also, when compared to our 2000 survey when we first asked this question, the percentage of SES managers expressing this view in 2003 was significantly higher than in 2000 (32 percent). The percentage of non-SES managers was essentially unchanged from 2000 (10 percent). Furthermore, as shown in figure 13, overall around half or fewer of managers responded "yes" on our 2003 survey to questions about being involved in developing ways to measure whether program performance goals are being achieved (46 percent), gathering and analyzing data to measure whether programs were meeting their specific performance goals (51 percent), or using measures for program performance goals to determine if the agency's strategic goals were being achieved (43 percent). None of these overall results were significantly different from our 1997 results. We did find, however, that significantly more SES managers responded "yes" on the 2003 survey (72 percent) than the 1997 survey (55 percent) with regard to being involved in using performance measurement information to determine if the agency's strategic goals were being achieved when compared to our 1997 results. Figure 13: Percentage of Federal Managers Responding "Yes" about Being Involved in the Following Activities: [See PDF for image] [End of figure] Managers Continue to Confront a Range of Human Capital Management Challenges: Managing people strategically and maintaining a highly skilled and energized workforce that is empowered to focus on results are critically important. Such human capital management practices are essential to the success of the federal government in the 21st century and to maximizing the value of its greatest asset--its people. Our survey results showed continuing challenges related to the adequacy of managerial decision making authority, training, and incentives. Federal Managers Report That They Are Held Accountable for Program Results but Do Not Have the Decision-Making Authority They Need to Accomplish Agency Goals: High-performing organizations seek to shift the focus of management and accountability from activities and processes to contributions and achieving results. In each of our three surveys, we asked managers about the amount of decision-making authority they had and the degree to which they were held accountable for results. As shown in figure 14, for 2003, an estimated 40 percent of federal managers overall reported that they had the decision-making authority they needed to help the agency accomplish its strategic goals to a great or very great extent. This was a statistically significant increase over our 1997 estimate of 31 percent. While there were more SES and non-SES managers expressing this view on our 2003 survey than the 1997 survey, it was the non-SES managers that showed the significant increase. Despite this promising trend, however, there continued to be substantial differences in 2003, as well as on the two previous surveys, between the responses of SES and lower-level managers on this question. Compared to the 57 percent of SES managers who reported having such authority to a great or very great extent in 2003, only 38 percent of non-SES managers reported having such authority to a great or very great extent. Figure 14: Percentage of Federal Managers Reporting to a Great or Very Great Extent That Managers/Supervisors at Their Levels Had the Decision-Making Authority They Needed to Help the Agency Accomplish Its Strategic Goals: [See PDF for image] [A] There was a statistically significant difference between the 1997 and 2003 surveys. [B] There was a statistically significant difference between SES compared to non-SES for each survey. [End of figure] However, when asked the extent to which managers or supervisors at their levels were held accountable for the accomplishment of agency strategic goals, 57 percent responded to a great or very great extent in 2003. Unlike in other areas, where SES managers had significantly different views from non-SES managers, there was little difference in the area of accountability. (See fig. 15.): Figure 15: Percentage of Federal Managers, SES, and Non-SES in 2003 Reporting to a Great or Very Great Extent That They Were Held Accountable for the Accomplishment of Agency Strategic Goals: [See PDF for image] [End of figure] This 57 percent is significantly higher than the 40 percent of managers overall who indicated that they had comparable decision-making authority. However, in contrast to the question on authority, as shown in figure 14, where more SES managers than non-SES managers expressed the view that they had the authority, there was little difference, as shown in figure 15, between the two groups in their views about being held accountable for achieving agency strategic goals to a great or very great extent. As figures 14 and 15 further illustrate, roughly the same percentage of SES managers perceived to a great or very great extent that managers at their level had decision-making authority and accountability for achieving agency strategic goals. This result suggests that their authority was perceived to be on par with their accountability. In contrast, only 38 percent of non-SES managers perceived that managers at their levels had the decision-making authority they needed to a great or very great extent, while 57 percent perceived that they were held accountable to a comparable extent. Managers are hard-pressed to achieve results when they do not have sufficient authority to act. In our report containing the results of our 1997 survey, we noted that agencies needed to concentrate their efforts on areas where managers were not perceiving or experiencing progress, such as that concerning devolving decision-making authority to managers throughout their organizations. While authority for achieving results appears to be in a modestly upward trend, the balance between authority and accountability that fosters decision making to achieve results could be further improved, particularly among non-SES managers. Fewer Than Half of Managers Reported Training on Key Tasks: We previously reported on the need for agencies to expend resources on effective training and professional development to equip federal employees to work effectively.[Footnote 39] Among the resources focus group participants cited as lacking included federal managers and staff with competencies and skills needed to plan strategically, develop robust measures of performance, and analyze what the performance data mean. Our 2003 Guide calls for training and development efforts to be strategically focused on improving performance toward the agency's goals and put forward with the agency's organizational culture firmly in mind.[Footnote 40] Throughout this process it is important that top leaders in the agencies communicate that investments in training and development are expected to produce clearly identified results. By incorporating valid measures of effectiveness into the training and development programs they offer, agencies can better ensure that they will adequately address training objectives and thereby increase the likelihood that desired changes will occur in the target population's skills, knowledge, abilities, attitudes, or behaviors. Furthermore, if managers understand and support the objectives of training and development efforts, they can provide opportunities to successfully use the new skills and competencies on the job and model the behavior they expect to see in their employees. In response to our 2003 survey, fewer than half of managers answered "yes" when we asked them whether, during the past 3 years, their agencies had provided, arranged, or paid for training that would help them accomplish any of seven critical results-oriented management- related tasks. However, progress is indicated in our survey results. As shown in figure 16, more managers answered "yes" in 2003 on all seven training areas than in previous surveys. These increases were statistically significant for five of the tasks--setting program performance goals, using program performance information to make management decisions, linking program performance to the achievement of agency strategic goals, and implementing the requirements of GPRA. Figure 16: Percentage of Federal Managers in Each Survey Year Who Reported That during the Past 3 Years Their Agencies Provided, Arranged, or Paid for Training That Would Help Them Accomplish Specific Tasks: [See PDF for image] [A] This question was not asked in the 1997 survey. [B] There was a statistically significant difference between the 2000 and 2003 surveys. [C] There was a statistically significant difference between the 1997 and 2003 surveys. [End of figure] As with our 2000 survey results, the 2003 survey results continued to demonstrate that there is a positive relationship between agencies providing training and development on setting program performance goals and the use of performance information when setting or revising performance goals. For those managers who responded "yes" to training on setting performance goals, 60 percent also reported that they used information obtained from performance measurement when setting new or revising existing performance goals to a great or very great extent. In contrast, for those managers who responded "no" to training on setting performance goals, only 38 percent reported that they used information obtained from performance measurement for setting new or revising existing performance goals to a great or very great extent. The difference between these percentages is statistically significant. Effective training and development programs are an integral part of a learning environment that can enhance the federal government's ability to attract and retain employees with the skills and competencies needed to achieve results. Training and developing new and current staff to fill new roles and work in different ways will be a crucial part of the federal government's endeavors to meet its transformation challenges. Ways that employees learn and achieve results will also continue to transform how agencies do business and engage employees in further innovation and improvements. Managers Perceive a Lack of Positive Recognition for Helping Agencies Achieve Results: Another fundamental aspect of the human capital management challenge agencies face is providing the incentives to their employees to encourage results-oriented management. Monetary and nonmonetary incentives can be used as a method for federal agencies to reward employees and to motivate them to focus on results. Overall, an increasing but still small percentage of managers reported in 1997, 2000, and 2003 that employees in their agencies received positive recognition to a great or very great extent for helping agencies accomplish their strategic goals. In 1997, 26 percent of federal managers reported such an extent of positive recognition as compared to 37 percent in 2003, a statistically significant increase. Interestingly, this improvement is seen in the responses of non-SES managers. As shown in figure 17, the percentage of SES managers expressing this view stayed at about the same level over the three surveys, while the percentage of non-SES managers holding this view was significantly higher in 2003 than in 1997. Even with this improvement on the part of the responses from non-SES managers, significantly more SES managers (47 percent) than non-SES managers (36 percent) expressed this perception to a comparable extent in 2003. Figure 17: Percentage of Federal Managers Who Reported to a Great or Very Great Extent That Employees in Their Agencies Received Positive Recognition for Helping Their Agencies Accomplish Their Strategic Goals: [See PDF for image] [A] There was a statistically significant difference between the 1997 and 2003 surveys. [End of figure] Unfortunately, most existing federal performance appraisal systems are not designed to support a meaningful performance-based pay system in that they fail to link institutional, program, unit, and individual performance measurement and reward systems. In our view, one key need is to modernize performance management systems in executive agencies so that they link to the agency's strategic plan, related goals, and desired outcomes and are therefore capable of adequately supporting more performance-based pay and other personnel decisions. We have reported federal agencies can develop effective performance management systems by implementing a selected, generally consistent set of key practices. These key practices helped public sector organizations both in the United States and abroad create a clear linkage--"line of sight"--between individual performance and organizational success and, thus, transform their cultures to be more results oriented, customer-focused, and collaborative in nature. Examples of such practices include: * aligning individual performance expectations with organizational goals, * connecting performance expectations to crosscutting goals, * linking pay to individual and organizational performance, and: * making meaningful distinctions in performance.[Footnote 41] Beyond implementing these key practices, high-performing organizations understand that their employees are assets whose value to the organization must be recognized, understood, and enhanced. They view an effective performance management system as an investment to maximize the effectiveness of people by developing individual potential to contribute to organizational goals. To maximize this investment, an organization's performance management system is designed, implemented, and continuously assessed by the standard of how well it helps the employees help the organization achieve results and pursue its mission. Persistent Challenges in Setting Outcome-Oriented Goals, Measuring Performance, and Collecting Useful Data: In prior reports, we have described difficulties faced by federal managers in developing useful, outcome-oriented measures of performance and collecting data indicating progress achieved.[Footnote 42] One of the most persistent challenges has been the development of outcome- oriented performance measures. Additionally, it is difficult to distinguish the impact of a particular federal program from the impact of other programs and factors, thus making it difficult to attribute specific program performance to results. The lack of timely and useful performance information can also hinder GPRA implementation. Meaningful, Outcome-Oriented Performance Measures Are Sometimes Hard to Develop: In the past, we have noted that federal managers found meaningful performance measures difficult to develop. Focus group participants and survey respondents noted that outcome-oriented performance measures were especially difficult to establish when the program or line of effort was not easily quantifiable. The challenge of the "complexity of establishing outcome-oriented goals and measuring performance" was cited by six of the eight focus groups as one of the key challenges that managers face in implementing GPRA. Focus group participants agreed that they often felt as if they were trying to measure the immeasurable, not having a clear understanding of which performance indicators could accurately inform the agency how it is carrying out a specific activity. Managers from agencies engaged in basic science research and development and grant-making functions noted that this effort was particularly difficult for them because federal programs, especially those that are research-based, often take years to achieve the full scope of their goals. On our most recent survey, we estimated that 36 percent of federal managers who had an opinion indicated that the determination of meaningful measures hindered the use of performance information or performance measurement to a great or very great extent. While this number was significantly lower than the percentage of managers expressing the comparable view on the 1997 or 2000 survey and may reflect some lessening of this as a hindrance to some managers, it nonetheless continues to be among those items having the largest percentage of managers citing it as a substantial hindrance. Impact of Federal Programs Difficult to Discern: In our June 1997 report on GPRA, we noted that "the often limited or indirect influence that the federal government has in determining whether a desired result is achieved complicates the effort to identify and measure the discrete contribution of the federal initiative to a specific program result."[Footnote 43] This occurs primarily because many federal programs' objectives are the result of complex systems or phenomena outside the program's control. In such cases, it is particularly challenging for agencies to confidently attribute changes in outcomes to their program--the central task of program impact evaluation. This is particularly challenging for regulatory programs, scientific research programs, and programs that deliver services to taxpayers through third parties, such as state and local governments. We have reported that determining the specific outcomes resulting from federal research and development has been a challenge that will not be easily resolved.[Footnote 44] Due to the difficulties in identifying outcomes, research and development agencies typically have chosen to measure a variety of proxies for outcomes, such as the number of patents resulting from federally funded research, expert review and judgments of the quality and importance of research findings, the number of project-related publications or citations, and contributions to expanding the number of research scientists. We have also reported that implementing GPRA in a regulatory environment is particularly challenging.[Footnote 45] Although federal agencies are generally required to assess the potential benefits and costs of proposed major regulatory actions, they generally do not monitor the benefits and costs of how these and other federal programs have actually performed. For example, in the case of the Environmental Protection Agency (EPA), to determine if existing environmental regulations need to be retained or improved, we previously recommended that EPA study the actual costs and benefits of such regulations.[Footnote 46] In the past, regulatory agencies have cited numerous barriers to their efforts to establish results-oriented goals and measures. These barriers included problems in obtaining data to demonstrate results, accounting for factors outside of the agency's control that affect results, and dealing with the long time periods often needed to see results. Our prior work discussed best practices for addressing challenges to measuring the results of regulatory programs. In particular, to address the challenge of discerning the impact of a federal program, when other factors also affect results, we suggested agencies "establish a rationale of how the program delivers results." Establishing such a rationale involves three related practices: (1) taking a holistic or "systems" approach to the problem being addressed, (2) building a program logic model that described how activities translated to outcomes, and (3) expanding program assessments and evaluations to validate the model linkages and rationale. We have also reported on the difficulties encountered in meeting GPRA reporting requirements for intergovernmental grant programs.[Footnote 47] Programs that do not deliver a readily measurable product or service are likely to have difficulty meeting GPRA performance measurement and reporting requirements. Intergovernmental grant programs, particularly those with the flexibility inherent in classic block grant design, may be more likely to have difficulty producing performance measures at the national level and raise delicate issues of accountability. Although most flexible grant programs we reviewed reported simple activity or client counts, relatively few of them collected uniform data on the outcomes of state or local service activities. Collecting such data requires conditions (such as uniformity of activities, objectives, and measures) that do not exist under many flexible program designs, and even where overall performance of a state or local program can be measured, the amount attributable to federal funding often cannot be separated out. Focus group participants also suggested that they faced challenges in obtaining timely performance data from relevant partner organizations and in identifying what the federal government's contribution has been to a specific outcome. Furthermore, survey respondents provided some corroboration for these views. Across all three of our surveys, we estimate that roughly a quarter of all federal managers reported this difficulty--distinguishing between the results produced by the program they were involved with and results caused by other factors--as a substantial hindrance. In response to a survey question about what the federal government could do to improve its overall focus on managing for results, one respondent noted: "Defining meaningful measures for the work we do is extremely difficult; and even if they could be defined, performance and accomplishment is (sic) dependent on so many factors outside our control that it is difficult, if not impossible, to make valid conclusions.": Timely, Useful Performance Information Not Always Available: In February 2000, we reported that intergovernmental programs pose potential difficulties in collecting timely and consistent national data.[Footnote 48] We also noted that agencies had limited program evaluation capabilities and weaknesses in agencies' financial management capabilities make it difficult for decision makers to effectively assess and improve many agencies' financial performance. On the basis of our current findings, these issues still exist. Federal managers who participated in our focus groups cited difficulties in gathering data from state or local entities, as well as statutory limitations regarding the nature and breadth of data that they were permitted to collect. However, in our 2003 survey, only 27 percent of federal managers indicated that obtaining data in time to be useful was a substantial hindrance; 31 percent expressed a comparable view with regard to obtaining valid or reliable data. Focus group participants also noted that OMB's accelerated time frames for reporting performance information will contribute to the challenge of producing complete, timely information in their agencies' performance reports. Over the past 2 fiscal years, OMB has moved the deadline for submission of agencies' performance reports (now performance and accountability reports) back from the statutory requirement of March 31; for fiscal year 2003 data, the deadline is January 30, 2004. In fiscal year 2004, these reports will be due on November 15, 2004. According to the managers, individual agencies may work on different time frames based partially on the population they serve or the stakeholders they must work with, such as state or local agencies. This "one size fits all" approach does not take such differences into account. Additionally, OMB requires agencies to report on their performance data quarterly; managers noted that this was particularly difficult for outcomes that may be achieved over extended periods of time, such as outcomes associated with basic science. As we have previously reported, measuring the performance of science-related projects can be difficult because a wide range of factors determine if and how a particular research and development project will result in a commercial application or have other benefits. Efforts to cure diseases or pursue space exploration are difficult to quantify and break down into meaningful quarterly performance measures. Crosscutting Issues Hinder Successful GPRA Implementation: Crosscutting issues continue to be a challenge to GPRA implementation. Mission fragmentation and program overlap are widespread across the federal government. Moreover, addressing this challenge is essential to the success of national strategies in areas such as homeland security, drug control, and the environment. We have reported that agencies could use the annual performance planning cycle and subsequent annual performance reports to highlight crosscutting program efforts and to provide evidence of the coordination of those efforts. Our review of six agencies' strategic and annual performance plans showed some improvement in addressing their crosscutting program efforts, but a great deal of improvement is still necessary. Few of the plans we reviewed attempted the more challenging task of discussing planned strategies for coordination and establishing complementary performance goals and complementary or common performance measures. For example, SSA's 2004 performance plan makes some mention of the agency's efforts to coordinate with other agencies to preserve the integrity of the Social Security number as a personal identifier, but there are very few details about this important component of its mission. Previous GAO reports and agency managers identified several barriers to interagency coordination. First, missions may not be mutually reinforcing or may even conflict, making reaching a consensus on strategies and priorities difficult. In 1998 and 1999, we found that mission fragmentation and program overlap existed in 12 federal mission areas, ranging from agriculture to natural resources and the environment. Implementation of federal crosscutting programs is often characterized by numerous individual agency efforts that are implemented with little apparent regard for the presence of related activities. Second, we reported on agencies' interest in protecting jurisdiction over missions and control over resources.[Footnote 49] Focus group participants echoed this concern, noting that there can be "turf battles" between agencies, where jurisdictional boundaries, as well as control over resources, are hotly contested. Finally, incompatible procedures, processes, data, and computer systems pose difficulties for agencies to work across agency boundaries. For example, we reported how the lack of consistent data on federal wetlands programs implemented by different agencies prevented the government from measuring progress toward achieving the governmentwide goal of no net loss of the nation's wetlands.[Footnote 50] We have previously reported and testified that GPRA could provide OMB, agencies, and Congress with a structured framework for addressing crosscutting program efforts.[Footnote 51] OMB, for example, could use the provision of GPRA that calls for OMB to develop a governmentwide performance plan to integrate expected agency-level performance. Unfortunately, this provision has not been fully implemented. OMB issued the first and only such plan in February 1998 for fiscal year 1999. In our review of the plan,[Footnote 52] we found that it included a broad range of governmentwide management objectives and a mission- based presentation of key performance goals based on agency performance plans and the plan's framework should ultimately allow for a cohesive presentation of governmentwide performance. However, the specific contents of this initial plan did not always deliver an integrated, consistent, and results-oriented picture of fiscal year 1999 federal government performance goals. OMB officials we interviewed at the time stressed that developing the governmentwide plan was viewed as an essential and integral component of the President's budget and planning process. From OMB's perspective, both the plan and the budget submission were intended to serve as communication tools for a range of possible users. In their opinion, the plan added value by reflecting a governmentwide perspective on policy choices made throughout the budget formulation process. OMB acknowledged that the plan itself did not serve to change the process through which decisions on government priorities were made, but enhanced it by placing a greater emphasis on results. As one official described it, the governmentwide performance plan was a derivative document, reflecting the budget and management decisions made throughout the process of formulating the President's budget submission. However, we found that focusing broadly on governmentwide outcomes should be a central and distinguishing feature of the federal government performance plan. To be most effective and supportive of the purposes of GPRA, the governmentwide plan must be more than a compilation of agency-level plans; integration, rather than repetition, must be its guiding principle. OMB has not issued a distinct governmentwide performance plan since fiscal year 1999. Most recently, the President's fiscal year 2004 budget focused on describing agencies' progress in addressing the PMA and the results of PART reviews of agency programs. Although such information is important and useful, it does not provide a broader and more integrated perspective of planned performance on governmentwide outcomes. Additionally, the fiscal year 2004 budget identified budget requests and performance objectives by agency, such as the U.S. Department of Defense, as opposed to crosscutting governmentwide themes. From this presentation, one could assume that the only activities the U.S. government planned to carry out in support of national defense were those listed under the chapter "Department of Defense." However, the chapter of the fiscal year 2004 budget discussing "the Department of State and International Assistance Programs," contains a heading titled, "Countering the Threat from Weapons of Mass Destruction." And while OMB may have a technical reason for not classifying this task as being related to national defense or homeland security, it is unclear that a lay reader could make that distinction. The fiscal year 2005 budget also identified budget requests by agency, not by crosscutting theme. Without such a governmentwide focus, OMB is missing an opportunity to assess and communicate the relationship between individual agency goals and outcomes that cut across federal agencies and more clearly relate and address the contributions of alternative federal strategies. The governmentwide performance plan also could help Congress and the executive branch address critical federal performance and management issues, including redundancy and other inefficiencies in how we do business. It could also provide a framework for any restructuring efforts. A strategic plan for the federal government, supported by key national indicators to assess the government's performance, position, and progress, could provide an additional tool for governmentwide reexamination of existing programs, as well as proposals for new programs. If fully developed, a governmentwide strategic plan could potentially provide a cohesive perspective on the long-term goals of the federal government and provide a much needed basis for fully integrating, rather than merely coordinating, a wide array of federal activities. Successful strategic planning requires the involvement of key stakeholders. Thus, it could serve as a mechanism for building consensus. Further, it could provide a vehicle for the President to articulate long-term goals and a road map for achieving them. In addition, a strategic plan could provide a more comprehensive framework for considering organizational changes and making resource decisions. Developing a strategic plan for the federal government would be an important first step in articulating the role, goals, and objectives of the federal government. It could help provide critical horizontal and vertical linkages. Horizontally, it could integrate and foster synergies among components of the federal government as well as help to clarify the role of the federal government vis-à-vis other sectors of our society. Vertically, it could provide a framework of federal missions and goals within which individual federal agencies could align their own missions and goals that would cascade down to individual employees. The development of a set of key national indicators could be used as a basis to inform the development of the governmentwide strategic and annual performance plans. The indicators could also link to and provide information to support outcome-oriented goals and objectives in agency-level strategic and annual performance plans. Managers View Congress' Use of Performance Information as Limited: Focus group members believed that one of the main challenges to GPRA implementation was the reluctance of Congress to use that information when making decisions, especially appropriations decisions. This concern was cited as a significant challenge in each of the focus groups, and was one of the top three "challenges" in five of the eight focus groups. In some cases, managers in our focus groups noted that this lack of usage was a significant disincentive to doing a good job in preparing GPRA plans and reports. Agency managers made the following criticisms regarding the perceived lack of congressional use of performance information: * appropriators have not bought into GPRA, so there is no incentive to do this well, * failure of congressional leadership in developing and using performance measures, * appropriators do not use performance data or tools to make decisions, and: * GPRA does not drive public policy decisions. Results from our survey provide some further information in support of this view. On our 2003 survey, when we asked federal managers about the extent to which they thought congressional committees paid attention to agency efforts under GPRA, only 22 percent of federal managers responded in the great to very great categories. This result was not significantly different from the results we observed on our 2000 survey when we asked this question about three specific types of congressional committees--authorization, appropriation, and oversight. On the 2000 survey, only 18 percent of federal managers held a similar view concerning authorizing committees, 19 percent for appropriations committees, and 20 percent for oversight committees. As we noted earlier, when this item was asked in relation to OMB, there was a significant increase in the percentage of managers responding to a great or very great extent from 2000 to 2003. The 31 percent of managers who viewed OMB as paying attention to a great or very great extent in 2003 was significantly higher than the 22 percent holding a comparable view of congressional committees. Although managers expressed these concerns about the use of this information, a recent review by the CRS suggested that Congress uses performance information to some extent, as evidenced by citations in legislation and committee reports.[Footnote 53] For example, in the 106th Congress (1999-2000), 42 public laws contained statutory language relating to GPRA and performance measures, and 118 legislative reports[Footnote 54] contained GPRA-associated passages. As shown in figure 18, across all three of our surveys, only a minority of federal managers governmentwide viewed the lack of ongoing congressional commitment for using performance information as a hindrance to a great or very great extent. Figure 18: Percentage of Federal Managers Reporting to a Great or Very Great Extent That a Lack of Ongoing Congressional Commitment or Support for Using Performance Information in Making Program/Funding Decisions Is a Hindrance: [See PDF for image] Note: Percentages are based on those respondents answering on the extent scale. [End of figure] While there is concern regarding Congress' use of performance information, it is important to make sure that this information is initially useful. One of GPRA's purposes is to respond to a need for accurate, reliable information for congressional decision making. In 2000, we reported that congressional staffs stated that they were looking for recurring information on spending priorities within programs; the quality, quantity, and efficiency of program operations; the populations served or regulated; as well as programs' progress in meeting their objectives.[Footnote 55] For example, learning who benefits from a program can help in addressing questions about how well services are targeted to those most in need. Some of these recurring needs were met through formal agency documents, such as annual performance plans. However, some information the agencies provided did not fully meet the congressional staffs' needs because the presentation was not clear, directly relevant, or sufficiently detailed. For example, congressional staffs wanted to see more direct linkages among the agencies' resources, strategies, and goals. In other cases, the information was not readily available to the congressional staffs, either because it had not been requested or reported, or because staff were not informed that it was available. As a key user of performance information, Congress also needs to be considered a partner in shaping agency goals at the outset. For example, through the strategic planning requirement, GPRA requires federal agencies to consult with Congress and key stakeholders to reassess their missions and long-term goals as well as the strategies and resources they will need to achieve their goals. GPRA also provides a vehicle for Congress to explicitly state its performance expectations in outcome-oriented terms when establishing new programs or in exercising oversight of existing programs that are not achieving desired results. Congress could use authorizing and appropriations hearings to determine if agency programs have clear performance goals, measures, and data with which to track progress and whether the programs are achieving their goals. If goals and objectives are unclear or not results oriented, Congress could use legislation to articulate the program outcomes it expects agencies to achieve. This would provide important guidance to agencies that could then be incorporated in agency strategic and annual performance plans. [End of section] Chapter 5: Conclusions and Recommendations: Agenda for Achieving a Sustainable, Governmentwide Focus on Results: As we have shown in this report, in the 10 years since the enactment of GPRA, significant progress has been made in instilling a focus on results in the federal government. First, GPRA statutory requirements laid a foundation for results-oriented management in federal agencies. Expert and agency focus group participants cited the creation of this statutory foundation as one of the key accomplishments of GPRA. Since GPRA began to be implemented governmentwide in fiscal year 1997, we have observed significant increases in the percentage of federal managers who reported having results-oriented performance measures for their programs. Focus group participants' views on whether GPRA has had a positive effect on the federal government's ability to deliver results to the American public were mixed. For example, the information gathered and reported for GPRA allows agencies to make better-informed decisions, which improves their ability to achieve results. In addition, GPRA has made the results of federal programs more transparent to the public. Other participants stated that while certain aspects of GPRA-related work have been positive, agencies' ability to deliver results and public awareness of their activities cannot be exclusively attributed to GPRA. Second, GPRA has increased the connection between resources and results by creating more formal linkages between agency performance goals and objectives and the program activities in the budget. Over the first 4 years of agency efforts to implement GPRA, we observed that agencies continued to tighten the required linkage between their performance plans and budget requests. However, much remains to be done in this area. For example, we have not observed notable increases in federal managers' perceptions about their personal use of plans or performance information when allocating resources, or about the use of performance information when funding decisions are made about their programs. However, it should be noted that we estimate a majority have positive perceptions about the use of performance information to allocate resources. Third, GPRA has provided a foundation for examining agency missions, performance goals and objectives, and the results achieved. We have seen improvements in the quality of agency strategic plans, annual performance plans, and performance reports since initial efforts. However, few of the six agencies we reviewed in this report produced GPRA planning and reporting documents that met all of our criteria for the highest level of quality. Most of these agencies continued to miss opportunities to present clear pictures of their intended and actual performance results in their GPRA plans and reports and to show how resources are aligned with actual performance results. Furthermore, most of the agencies we reviewed did not provide a full level of confidence in the credibility of their performance data. Performance-based management, as envisioned by GPRA, requires transforming organizational cultures to improve decision making, maximize performance, and assure accountability. This transformation is not an easy one and requires investments of time and resources as well as sustained leadership commitment and attention. Challenges to successful implementation of GPRA include inconsistent top leadership commitment to creating a focus on results; an approach to setting goals and developing strategies for achieving critical outcomes that creates individual agency stovepipes rather than an integrated, holistic governmentwide approach; getting federal managers to make greater use of performance information to manage their programs and providing them authority to act that is commensurate with their accountability for results; difficulty in establishing meaningful measures of outcomes and assessing results of federal programs that are carried out by nonfederal entities; and untimely performance data. The challenges identified in this report are not new--most have not changed significantly since we first reported on governmentwide implementation of GPRA. However, we have frequently reported on approaches that agencies, OMB, and Congress could use to address the challenges. These approaches include strengthening the commitment of top leadership to creating and sustaining a focus on results; taking a governmentwide approach to achieving outcomes that are crosscutting in nature; improving the usefulness of performance information to managers, Congress, and the public; and improving the quality of performance measures and data. Collectively, these approaches form the agenda that federal agencies, OMB, and Congress will need to follow to bring about a more sustainable, governmentwide focus on results. Strengthening Top Leadership Commitment to Creating and Sustaining Results-Oriented Cultures: Successfully addressing the challenges that federal agencies face requires leaders who are committed to achieving results, who recognize the importance of using results-oriented goals and quantifiable measures, and who integrate performance-based management into the culture and day-to-day activities of their organizations. Top leadership must play a critical role in creating and sustaining high- performing organizations. Without the clear and demonstrated commitment of agency top leadership--both political and career--organizational cultures will not be transformed, and new visions and ways of doing business will not take root. To be positioned to address the array of challenges faced by our national government, federal agencies will need to transform their organizational cultures so that they are more results oriented, customer-focused, and collaborative. Leading public organizations here in the United States and abroad have found that strategic human capital management must be the centerpiece of any serious change management initiative and efforts to transform the cultures of government agencies. Performance management systems are integral to strategic human capital management. Such systems can be key tools to maximizing performance by aligning institutional performance measures with individual performance and creating a "line of sight" between individual and organizational goals. Leading organizations use their performance management systems as a key tool for aligning institutional, unit, and employee performance; achieving results; accelerating change; managing the organization day to day; and facilitating communication throughout the year so that discussions about individual and organizational performance are integrated and ongoing.[Footnote 56] Furthermore, achieving this cultural transformation requires people to have the knowledge and skills to develop and use performance information to improve program performance. Our survey data indicated a significant relationship between those managers who reported they received training on setting performance goals and those who used performance information when setting or revising performance goals. However, federal agencies have not consistently showed a commitment to investing in needed training and development opportunities to help ensure that managers and employees have the requisite skills and competencies to achieve agency goals. The commitment to focusing on and using performance information needs to extend to OMB and Congress as well. Through the administration's PMA and PART initiatives, OMB has clearly placed greater emphasis on management issues over the past several years. However, the focus of such oversight needs to extend beyond the emphasis on formulating the President's Budget to include an examination of the many challenges agencies face that may be contributing to poor performance. In spite of the persistent weaknesses we found in agencies' strategic plans and annual performance plans and reports, OMB significantly reduced the scope of its guidance to agencies on how to prepare these documents. By emphasizing a focus on resource allocation through its PART exercise and providing less information on how to comply with GPRA, OMB may be sending a message to agencies that compliance with GPRA is not important. Without strong leadership from OMB, the foundation of performance information that has been built could deteriorate. OMB leadership is critical to addressing the continuing challenges presented in GPRA implementation and the transformation of the federal government to an increasingly results-oriented culture. OMB, as the primary focal point for overall management in the federal government, can provide the needed impetus by providing guidance, fostering communication among agencies, and forming intragovernmental councils and work groups tasked with identifying potential approaches and solutions to overcoming the persistent challenges to results-oriented management. Congress can also play a decisive role in fostering results-oriented cultures in the federal government by using information on agency goals and results at confirmation, oversight, authorization, and appropriation hearings. Consistent congressional interest in the status of an agency's GPRA efforts, performance measures, and uses of performance information to make decisions, will send an unmistakable message to agencies that Congress expects GPRA to be thoroughly implemented. We also found that timing issues may affect the development of agency strategic plans that are meaningful and useful to top leadership. The commitment of top leadership within agencies, OMB, and Congress is critical to the success of strategic planning efforts. A strategic plan should reflect the policy priorities of an organization's leaders and the input of key stakeholders if it is to be an effective management tool. However, GPRA specifies time frames for updating strategic plans that do not correspond to presidential or congressional terms. As a result, an agency may be required to update its strategic plan a year before a presidential election and without input from a new Congress. If a new president is elected, the updated plan is essentially moot and agencies must spend additional time and effort revising it to reflect new priorities. Our focus group participants, including GPRA experts, strongly agreed that this timing issue should be addressed by adjusting time frames to correspond better with presidential and congressional terms. Addressing Governmentwide Needs: Mission fragmentation and program overlap are widespread throughout the federal government.[Footnote 57] We have noted that interagency coordination is important for ensuring that crosscutting program efforts are mutually reinforcing and efficiently implemented. Our review of six agencies' strategic and annual performance plans along with our previous work on crosscutting issues has demonstrated that agencies' still present their goals and strategies in a mostly stovepiped manner. They have generally not used their plans to communicate the nature of their coordination with other agencies, in terms of the development of common or complementary goals and objectives or strategies jointly undertaken to achieve those goals. We have also reported that GPRA could provide a tool to reexamine federal government roles and structures governmentwide. GPRA requires the President to include in his annual budget submission a federal government performance plan. Congress intended that this plan provide a "single cohesive picture of the annual performance goals for the fiscal year." The governmentwide performance plan could help Congress and the executive branch address critical federal performance and management issues, including redundancy and other inefficiencies in how we do business. It could also provide a framework for any restructuring efforts. Unfortunately, this provision has not been fully implemented. Instead, OMB has used the President's Budget to present high-level information about agencies and certain program performance issues. The agency-by-agency focus of the budget does not provide the integrated perspective of government performance envisioned by GPRA. If the governmentwide performance plan were fully implemented, it could provide a framework for such congressional oversight. For example, in recent years, OMB has begun to develop common measures for similar programs, such as job training. By focusing on broad goals and objectives, oversight could more effectively cut across organization, program, and other traditional boundaries. Such oversight might also cut across existing committee boundaries, which suggests that Congress may benefit from using specialized mechanisms to perform oversight (i.e., joint hearings and special committees). A strategic plan for the federal government, supported by key national indicators to assess the government's performance, position, and progress, could provide an additional tool for governmentwide reexamination of existing programs, as well as proposals for new programs. If fully developed, a governmentwide strategic plan can potentially provide a cohesive perspective on the long-term goals of the federal government and provide a much needed basis for fully integrating, rather than merely coordinating, a wide array of federal activities. Successful strategic planning requires the involvement of key stakeholders. Thus, it could serve as a mechanism for building consensus. Further, it could provide a vehicle for the President to articulate long-term goals and a road map for achieving them. In addition, a strategic plan can provide a more comprehensive framework for considering organizational changes and making resource decisions. Developing a strategic plan for the federal government would be an important first step in articulating the role, goals, and objectives of the federal government. It could help provide critical horizontal and vertical linkages. Horizontally, it could integrate and foster synergies among components of the federal government as well as help to clarify the role of the federal government vis-à-vis other sectors of our society. Vertically, it could provide a framework of federal missions and goals within which individual federal agencies could align their own missions and goals that would cascade down to individual employees. The development of a set of key national indicators could be used as a basis to inform the development of governmentwide strategic and annual performance plans. The indicators could also link to and provide information to support outcome-oriented goals and objectives in agency-level strategic and annual performance plans. Improving Usefulness of Performance Information: We have found that leading organizations that progressed the farthest to results-oriented management did not stop after strategic planning and performance measurement. They applied their acquired knowledge and data to identify gaps in their performance, report on that performance, and finally use that information to improve their performance to better support their missions. Performance data can have real value only if they are used to identify the gap between an organization's actual performance level and the performance level it has identified as its goal. Once the performance gaps are identified for different program areas, managers can determine where to target their resources to improve overall mission accomplishment. When managers are forced to reduce their resources, the same analysis can help them target reductions to keep to a minimum the impact on their organization's overall mission.[Footnote 58] Under GPRA, agencies produce a single strategic plan, annual performance plan, and annual performance report. However, there are many potential consumers of agencies' performance information-- Congress, the public, and the agency itself. One size need not fit all. Clearly, an agency will need more detailed information on its programs for operational purposes than would be suitable for external audiences. Of the six agencies' performance reports we reviewed, some of them provided useful summary tables or information that provided overall snapshots of performance or highlighted progress in achieving key goals. Other reports that lacked such a summary made it difficult to assess the progress achieved. To improve the prospect that agency performance information will be useful to and used by these different users, agencies need to consider the different information needs and how to best tailor their performance information to meet those needs. For example, we have reported that, although many information needs were met, congressional staff also identified gaps in meeting their information needs.[Footnote 59] Key to addressing these information gaps was improving communication between congressional staff and agency officials to help ensure that congressional information needs are understood, and that, where feasible, arrangements are made to meet them. Improved two-way communication might also make clear what information is and is not available, as well as what is needed and what is not needed. This might entail the preparation of simplified and streamlined plans and reports for Congress and other external users. Another challenge that limits the usefulness of agency performance reports is the lack of timely data on performance. For the six performance reports we reviewed we continued to observe a significant number of goals for which performance data were unavailable. Policy decisions made when designing federal programs, particularly intergovernmental programs, may make it difficult to collect timely and consistent national data. In administering programs that are the joint responsibility of state and local governments, Congress and the executive branch continually balance the competing objectives of collecting uniform program information to assess performance with giving states and localities the flexibility needed to effectively implement intergovernmental programs. Improving Performance Measures and Data Quality: Another key challenge to achieving a governmentwide focus on results is that of developing meaningful, outcome-oriented performance goals and collecting performance data that can be used to assess results. Performance measurement under GPRA is the ongoing monitoring and reporting of program accomplishments, particularly progress toward preestablished goals. It tends to focus on regularly collected data on the level and type of program activities, the direct products and services delivered by the programs, and the results of those activities. For programs that have readily observable results or outcomes, performance measurement may provide sufficient information to demonstrate program results. In some programs, however, outcomes are not quickly achieved or readily observed, or their relationship to the program is uncertain. In such cases, more in-depth program evaluations may be needed, in addition to performance measurement, to examine the extent to which a program is achieving its objectives. Given the difficult measurement challenges we have identified, it is all the more important that agency strategic planning efforts include the identification of the most critical evaluations that need to take place to address those challenges. However, our previous work has raised concerns about the capacity of federal agencies to produce evaluations of program effectiveness.[Footnote 60] Few of the agencies we reviewed deployed the rigorous research methods required to attribute changes underlying outcomes to program activities. Yet we have also seen how some agencies have profitably drawn on systematic program evaluations to improve their measurement of program performance or understanding of performance and how it might be improved.[Footnote 61] Our review of six agencies' strategic plans and performance reports in this report revealed weaknesses in their discussions of program evaluation. Most of the strategic plans lacked critical information required by GPRA, such as a discussion of how evaluations were used to establish strategic goals or a schedule of future evaluations. Furthermore, two of the six performance reports did not summarize the results of program evaluations completed that year, as required. Our work has also identified substantial, long-standing limitations in agencies' abilities to produce credible data and identify performance improvement opportunities that will not be quickly or easily resolved.[Footnote 62] According to our review, five of six agencies' annual performance plans showed meaningful improvements in how they discussed the quality of performance data. However, only DOT's performance plan and report contained information that provided a full level of confidence in the credibility of its performance data. In particular, the plans and reports did not always provide detailed information on how the agencies verified and validated their performance data. Recommendations for Executive Action: To provide a broader perspective and more cohesive picture of the federal government's goals and strategies to address issues that cut across executive branch agencies, we recommend that the Director of OMB fully implement GPRA's requirement to develop a governmentwide performance plan. To achieve the greatest benefit from both GPRA and PART, we recommend that the Director of OMB articulate and implement an integrated and complementary relationship between the two. GPRA is a broad legislative framework that was designed to be consultative with Congress and other stakeholders, and allows for varying uses of performance information. PART looks through a particular lens for a particular use--the executive budget formulation process. To improve the quality of agencies' strategic plans, annual performance plans, and performance reports and help agencies meet the requirements of GPRA, we recommend that the Director of OMB provide clearer and more consistent guidance to executive branch agencies on how to implement GPRA. Such guidance should include standards for communicating key performance information in concise as well as longer formats to better meet the needs of external users who lack the time or expertise to analyze lengthy, detailed documents. To help address agencies' performance measurement challenges, we recommend that the Director of OMB engage in a continuing dialogue with agencies about their performance measurement practices with a particular focus on grant-making, research and development, and regulatory functions to identify and replicate successful approaches agencies are using to measure and report on their outcomes, including the use of program evaluation tools. Additionally, we recommend that the Director of OMB work with executive branch agencies to identify the barriers to obtaining timely data to show progress against performance goals and the best ways to report information where there are unavoidable lags in data availability. Interagency councils, such as the President's Management Council and the Chief Financial Officers' Council, may be effective vehicles for working on these issues. To facilitate the transformation of agencies' management cultures to be more results-oriented, we recommend that the Director of OMB work with agencies to ensure they are making adequate investments in training on performance planning and measurement, with a particular emphasis on how to use performance information to improve program performance. Matters for Congressional Consideration: To ensure that agency strategic plans more closely align with changes in the federal government leadership, Congress should consider amending GPRA to require that updates to agency strategic plans be submitted at least once every 4 years, 12-18 months after a new administration begins its term. Additionally, consultations with congressional stakeholders should be held at least once every new Congress and interim updates made to strategic and performance plans as warranted. Congress should consider using these consultations along with its traditional oversight role and legislation as opportunities to clarify its performance expectations for agencies. This process may provide an opportunity for Congress to develop a more structured oversight agenda. To provide a framework to identify long-term goals and strategies to address issues that cut across federal agencies, Congress also should consider amending GPRA to require the President to develop a governmentwide strategic plan. Agency Comments: We provided a copy of the draft report to OMB for comment. OMB's written comments are reprinted in appendix VIII. In general, OMB agreed with our findings and conclusions. OMB agreed to implement most of our recommendations, noting that these recommendations will enhance its efforts to make the government more results oriented. OMB agreed to (1) work with agencies to ensure they are provided adequate training in performance management, (2) revise its guidance to clarify the integrated and complementary relationship between GPRA and PART, and (3) continue to use PART to improve agency performance measurement practices and share those practices across government. In response to our recommendation that OMB fully implement GPRA's requirement to develop a governmentwide performance plan, OMB stated that the President's Budget represents the executive branch's governmentwide performance plan. However, the agency-by-agency focus of the budget over the past few years does not provide an integrated perspective of government performance, and thus does not meet GPRA's requirement to provide a "single cohesive picture of the annual performance goals for the fiscal year." In response to our matter for congressional consideration that Congress should consider amending GPRA to require the President to develop a governmentwide strategic plan, OMB noted that the budget serves as the governmentwide strategic plan. However, in our opinion, the President's Budget focuses on establishing agency budgets for the upcoming fiscal year. Unlike a strategic plan, it provides neither a long-term nor an integrated perspective on the federal government's activities. A governmentwide strategic plan should provide a cohesive perspective on the long-term goals of the federal government and provide a basis for fully integrating, rather than primarily coordinating, a wide array of federal activities. We provided relevant sections of the draft report to Education, DOE, HUD, SBA, SSA, and DOT. Education and SBA did not provide any comments, while DOT provided minor technical comments. Written comments from DOE, HUD, and SSA are reprinted in appendixes IX, X, and XI, respectively, along with our responses. DOE disagreed with portions of our analyses of its 2004 Annual Performance Plan and its 2002 Performance and Accountability Report. Our analysis of DOE's documents was based on specific criteria (see appendixes IV and V for details) and was characterized in relation to our reviews of the other five agencies' documents. We modified or clarified certain characterizations in response to DOE comments, but for the most part found that our characterizations were appropriate. SSA generally agreed with our observations and agreed to incorporate them in its future planning efforts. SSA made several points of clarification and disagreed with our observation that its performance and accountability report does not clearly state how program evaluations were used to answer questions about program performance and results and how they can be improved. SSA noted that its evaluations rely on surveys, and these surveys form the basis for its efforts to deliver high-quality service. SSA also noted that it lists other evaluations that are of great importance to its ongoing operations. We do not discount the usefulness of SSA's surveys in assessing its day- to-day management of programs. Rather, we believe that it would be helpful for SSA to clearly identify the range of evaluations conducted and how each of them contributed to improved program performance. HUD noted that all of the areas we suggested for further improvement are under consideration for improvement. However, they disagreed with us on two observations related to the strategic plan: (1) that the link between its long-term and intermediate goals is difficult to discern and (2) that it did not explain how it used the results of program evaluations to update the current plan and did not include a schedule for future evaluations. On the basis of OMB guidance for preparing strategic plans and the criteria we used to evaluate all six agencies' strategic plans (see app. III for more detail), we maintain that these two observations are valid and require further attention. HUD also disagreed with how we presented the performance information in its summary report cards (see fig. 22). HUD noted that many of the results were explained in the individual indicator write-ups that followed the summary information. Our analysis of HUD's information included qualitative aspects of how the information was presented, such as its usefulness to inform the average reader with little or no exposure to the subject matter, and the extent to which HUD presented a complete summary of performance information in a user-friendly format. Technical comments from DOE, HUD, and SSA were incorporated, as appropriate. [End of section] Appendixes: Appendix I: Objectives, Scope, and Methodology: As agreed with your offices, our objectives for this report were to determine (1) the effect of the Government Performance and Results Act (GPRA) over the last 10 years in creating a governmentwide focus on results and the government's ability to deliver results to the American public, including an assessment of changes in the overall quality of agencies' strategic plans, annual performance plans, and annual performance reports; (2) the challenges that agencies face in measuring performance and using performance information in management decisions; and (3) how the federal government can continue to shift toward a more results-oriented focus. To meet our objectives, we collected governmentwide data to assess the government's overall focus on results. We conducted a governmentwide survey of federal managers, focus groups with federal managers and GPRA experts, and interviews with top appointed officials. We identified and reviewed previously published reports on GPRA. Finally, we selected a sample of agencies to review for changes in the quality of their strategic plans, performance plans, and performance reports since their initial efforts. We conducted our work from January through November 2003 in Washington, D.C. in accordance with generally accepted government auditing standards. We provided drafts of the relevant sections of this report to officials from each of the agencies whose GPRA reports we reviewed. We also provided a draft of this report to OMB. Methodology for Governmentwide Survey: A Web-based questionnaire on performance and management issues was administered to a stratified random probability sample of 800 persons from a population of approximately 98,000 mid-level and upper-level civilian managers and supervisors working in the 24 executive branch agencies covered by the Chief Financial Officers Act of 1990 (CFO). The sample was drawn from the Office of Personnel Management's (OPM) Civilian Personnel Data File as of December 31, 2002, using file designators indicating performance of managerial and supervisory functions. The questionnaire was designed to obtain the observations and perceptions of respondents on various aspects of GPRA as well as such results-oriented management topics as the presence, use, and usefulness of performance measures, hindrances to measuring and using performance information, and agency climate. Most of the items on the questionnaire were closed-ended, meaning that depending on the particular item, respondents could choose one or more response categories or rate the strength of their perception on a 5-point extent scale. Almost all the items on this questionnaire were asked in two earlier mail-out surveys. One survey was conducted between November 1996 and January 1997 as part of the work we did in response to a GPRA requirement that we report on implementation of the act. The other survey was conducted between January and August 2000.[Footnote 63] This survey covered the same CFO Act agencies and was designed to update the results from the two earlier surveys. Similar to the two earlier surveys, the sample was stratified by whether the manager or supervisor was Senior Executive Service (SES) or non-SES. The management levels covered general schedule (GS), general management (GM), or equivalent schedules at levels comparable to GS/GM-13 through career SES or equivalent levels of executive service. The sample also included the same or equivalent special pay plans that were covered in our 2000 survey, e.g., Senior Foreign Service executives. We sent an e-mail to members of the sample that notified them of the survey's availability on the GAO Web site and included instructions on how to access and complete the survey. Members of the sample who did not respond to the initial notice were sent up to two subsequent reminders asking them to participate in the survey. The survey was administered from June through August 2003. During the course of the survey, we deleted 26 persons from our sample who had either retired, separated, died, or otherwise left the agency or had some other reason that excluded them from the population of interest. We received useable questionnaires from 503 sample respondents, about 65 percent of the eligible sample. The eligible sample includes 39 persons that we were unable to locate and therefore unable to request that they participate in the survey. To assess whether the views of those individuals who chose not to participate in our survey might be different than those who did, we made an effort to administer a brief survey over the telephone to those individuals who still had not responded about a month or more after the survey had been available to them despite being contacted twice after the initial notification e-mail had been sent out. This telephone survey consisted of four items from the full survey. There were 58 persons who agreed to answer these questions over the telephone. This was 41 percent of those individuals who had not responded at the time we attempted to contact them for the purpose of asking these four questions. We analyzed the responses of this group on the four selected items compared to the responses received from all other respondents. Our analyses of the items showed very few differences between nonresponders and responders. There was no sufficient or consistent pattern of responding that would warrant a conclusion that the views of nonresponders were notably different than responders. The responses of each eligible sample member who provided a useable questionnaire were subsequently weighted in the analysis to account statistically for all the members of the population. The overall survey results are generalizable to the population of managers as described above at the CFO Act agencies. All results are subject to some uncertainty or sampling error as well as nonsampling error. As part of our effort to reduce nonsampling sources of error in survey results, we checked and edited (1) the survey data for responses that failed to follow instructions and (2) the programs used in our analyses. In general, percentage estimates in this report for the entire sample have confidence intervals ranging from about ± 4 to ±11 percentage points at the 95 percent confidence interval. In other words, if all CFO Act agency managers and supervisors in our population had been surveyed, the chances are 95 out of 100 that the result obtained would not differ from our sample estimate in the more extreme cases by more than ±11 percent. Appendix VI shows the questions asked with the weighted percentage of managers responding to each item. Because a complex sample design was used in the current survey as well as the two previous surveys and different types of statistical analyses are being done, the magnitude of sampling error will vary across the particular surveys, groups, or items being compared due to differences in the underlying sample sizes and associated variances. The number of participants in the current survey is only about one fifth of the number in the 2000 survey (2,510) and slightly more than half of those in the first survey (905). The 2000 survey was designed with a larger sample than the other two surveys in order to provide estimates for each individual agency as well as all the CFO Act agencies collectively. Consequently, in some instances, a difference of a certain magnitude may be statistically significant. In other instances, depending on the nature of the comparison being made, a difference of equal or even greater magnitude may not achieve statistical significance. For example, when comparing a result from the current survey to the larger 2000 survey with its relatively smaller confidence interval, a difference of a certain magnitude may be significant. However, when comparing the current survey with the first survey, that difference may not be significant given the greater imprecision in the estimates due to both surveys' smaller sample sizes. We note throughout the report when differences are significant at the .05 probability level. Methodology for Focus Groups: We held a series of focus groups as one of our methods for obtaining information about the accomplishments and challenges agencies face in implementing and overseeing GPRA-related activities. Focus groups are a form of qualitative research in which a specially trained leader, a moderator, meets with a small group of people (usually 8 to 10) who are knowledgeable about the topics to be discussed. In all, we conducted eight focus groups--one with experts on GPRA and performance management and seven with federal managers. For our focus group with experts, we invited individuals from the private sector, academia, the National Academy of Public Administration, and OMB. These individuals were involved either in drafting GPRA, overseeing its implementation, or studying and critiquing implementation, from outside government. Out of the 14 experts we invited, a total of 11 attended the focus group. For our focus groups with agency managers, we obtained a list of potential participants for our focus groups by contacting all 24 CFO Act agencies and requesting that they submit a list of candidates and their profiles based on the following criteria: federal managers (1) in the GS-13 pay grade and above, including members of the SES; (2) having at least 3 years of managerial experience; (3) currently located in the Washington, D.C., metropolitan area; (4) having hands-on experience with GPRA or performance management;[Footnote 64] and (5) representing both departments and their component bureaus. We received profiles of candidates from all agencies; however, no managers from the OPM chose to participate in the focus groups. To select focus group participants, we reviewed the profiles submitted by agencies and selected candidates with diverse experience who held a variety of different positions within the agency in order to capture a broad range of perspectives. For example, we invited a comptroller; a deputy director of management, administration, and planning; budget directors; budget officers; management analysts; and program managers; among others. We contacted the candidates and provided them with the list of questions to be discussed at the focus group in advance so they would be aware of our interests and be better able to provide us, where possible, with examples to illustrate their responses to our questions. Out of 104 agency officials we invited, 70 participated in the focus groups.[Footnote 65] During each session, the moderator explained the scope of our work and elaborated on how the focus groups were one of several methods we were using to collect information relevant to our objectives. As part of the focus group process, the moderator asked participants at each session to identify the main accomplishments and challenges that, in their view, could be attributed to GPRA and to mention possible solutions to these challenges. During the sessions, we created lists of the accomplishments, challenges, and solutions identified by group participants and posted these lists around the meeting room. Participants were then asked to review the lists and vote on the three most important accomplishments and the top three challenges.[Footnote 66] To organize the information collected during the focus groups, we reviewed the statements made by participants in response to our questions. We identified related sets of statements and summarized them as a general theme. We noted how often a theme was expressed both within and across each focus group. We also examined the number of votes each posted statement obtained. Our analysis focused on those themes that were supported by statements that obtained a high number of votes and were mentioned frequently within and across the majority of focus groups. The focus group results discussed in this report are summary descriptions reflecting the range of views and perceptions held by employees, supervisors, or project managers who participated in the focus groups. Although we cannot assume all federal managers share these views, the extent to which certain opinions or perceptions were repeatedly expressed or endorsed by many participants from multiple focus groups provides a rough gauge of the significance of these views. Methodology for Interviews with Political Appointees: To obtain an additional perspective from top political managers of federal agencies on GPRA, we held telephone or in-person interviews with 10 high-level officials serving under political appointments with CFO Act agencies. Five former officials from the Clinton administration and five serving under the current Bush administration were interviewed. For example, we interviewed deputy secretaries, chief financial officers, and deputy assistant secretaries for management. We asked them to provide their perspective on the main accomplishments or other effects of GPRA, the key challenges to implementation, and possible improvements to GPRA. We summarized the interviewees' answers and identified recurring themes or observations for our analysis. Methodology for Selecting Agencies to Review for Changes in the Quality of Their Strategic Plans, Annual Performance Plans, and Annual Performance Reports: To address how the quality of agency strategic plans, performance plans, and performance reports have changed since their initial efforts, we reviewed a sample of six agencies' current strategic plans, annual performance plans, and annual performance reports and compared the results to our findings from prior reviews of the agencies' initial efforts in producing these documents. We did not independently verify or assess the information we obtained from agency plans and reports. If an agency chose not to discuss its efforts concerning elements in the plans and reports, it does not necessarily mean that the agency is not implementing those elements. We selected the departments and agencies to review based on the extent to which they collectively represented the full range of characteristics in the following four areas: (1) agency size (small, medium, large); (2) primary program types (direct service, research, regulatory, transfer payments, and contracts or grants); (3) quality of fiscal year 2000 performance plan based on our previous review (low, medium, high);[Footnote 67] and (4) type of agency (cabinet department and independent agency). Based on these characteristics, we selected the following departments and agencies: * Department of Education (Education), * Department of Energy (DOE), * Department of Housing and Urban Development (HUD), * Small Business Administration (SBA), * Social Security Administration (SSA), and: * Department of Transportation (DOT). Table 4 shows the characteristics represented by each of these agencies. Table 4: Summary of Characteristics of Agencies Selected for Review of Strategic Plans, Annual Performance Plans, and Annual Performance Reports: Departments: Education; Size[A]: Full time equivalent positions: 4,756; Size[A]: Small; Functions: Research: Yes; Functions: Direct service: Yes; Functions: Regulatory: Yes; Functions: Transfer payments: Yes; Functions: Grants/contracts: Yes; Quality of fiscal year 2000 performance plans: Medium. Departments: DOE; Size[A]: Full time equivalent positions: 16,067; Size[A]: Medium; Functions: Research: Yes; Functions: Direct service: No; Functions: Regulatory: Yes; Functions: Transfer payments: No; Functions: Grants/contracts: Yes; Quality of fiscal year 2000 performance plans: Low. Departments: HUD; Size[A]: Full time equivalent positions: 10,752; Size[A]: Medium; Functions: Research: No; Functions: Direct service: Yes; Functions: Regulatory: Yes; Functions: Transfer payments: Yes; Functions: Grants/contracts: Yes; Quality of fiscal year 2000 performance plans: Medium. Departments: DOT; Size[A]: Full time equivalent positions: 135,022[B]; Size[A]: Large; Functions: Research: Yes; Functions: Direct service: Yes; Functions: Regulatory: Yes; Functions: Transfer payments: No; Functions: Grants/contracts: Yes; Quality of fiscal year 2000 performance plans: H. Agencies: SBA; Size[A]: Full time equivalent positions: 4,005; Size[A]: Small; Functions: Research: No; Functions: Direct service: Yes; Functions: Regulatory: Yes; Functions: Transfer payments: No; Functions: Grants/contracts: Yes; Quality of fiscal year 2000 performance plans: Low. Agencies: SSA; Size[A]: Full time equivalent positions: 64,418; Size[A]: Large; Functions: Research: No; Functions: Direct service: Yes; Functions: Regulatory: No; Functions: Transfer payments: Yes; Functions: Grants/contracts: No; Quality of fiscal year 2000 performance plans: High. Source: GAO. [A] The size of the agencies is based on data from December 2002. [B] In March 2003, the U.S. Coast Guard and the Transportation Security Administration (TSA) were transferred from DOT to the Department of Homeland Security. According to the fiscal year 2005 President's Budget, in fiscal year 2003, the Coast Guard and TSA had 43,702 and 57,324 full-time equivalent positions, respectively. [End of table] A more detailed discussion of the criteria we used to assess the quality of the agencies' planning and reporting documents and the results of our review is contained in appendixes III, IV, and V. [End of section] Appendix II: Focus Group Participants Agreed GPRA Provides a Framework for Federal Agencies to Become More Results Oriented: While GPRA's goal is to make the federal government more results oriented, work carried out in support of this effort, such as planning activities, implementing programs, reporting on outcomes, and evaluating performance, generally lies in the hands of federal managers. To get a better appreciation for the main accomplishments and challenges agencies face in implementing and overseeing GPRA-related activities, we conducted a total of seven focus groups comprised of federal managers from 23 CFO Act agencies, and an eighth focus group comprised of 11 experts on GPRA and performance management and budgeting. For our focus groups, we asked participants to discuss (1) the key accomplishments of GPRA to date, (2) whether GPRA has created a focus on achieving results across the federal government, (3) the effect of GPRA on the government's ability to deliver results to the American public, (4) the persistent and prevalent challenges agencies face in implementing and overseeing GPRA-related activities, and (5) suggestions to address these challenges. We recorded the views expressed by participants and categorized them into themes that were most commonly expressed or endorsed both within and across the groups. The focus group results discussed in this report are organized according to the themes we identified and are summary descriptions reflecting the range of views and perceptions expressed by the experts, supervisors, and project managers. A more complete discussion of our scope and methodology can be found in appendix I. Focus group participants indicated that GPRA has helped to make the federal government more results oriented. However, they noted that a number of obstacles have made GPRA implementation challenging, such as difficulty in establishing results-oriented goals and measuring performance, addressing frequently changing priorities resulting from changes in administration, and lack of top leadership support for GPRA. In all, participants generally perceive the information contained in GPRA reports to be important and useful; however, they do not believe that lawmakers use this information to make resource decisions or conduct oversight. To address these problems and concerns, focus group participants stated that, among other things, Congress should provide guidance to agencies on how to make GPRA reports more useful, OMB should reinforce its value as a management tool, and agencies need to commit the resources necessary to carry out GPRA-related activities. GPRA Accomplishments: Overall, focus group participants stated that GPRA has had a positive effect on federal agencies' efforts to become more results oriented. Based in statute, GPRA has created a framework for agencies to focus on achieving results by requiring them to establish program goals and objectives, develop performance indicators, and measure the extent to which they have made progress towards achieving program goals. As a result, federal managers noted that they have been increasingly able to view their programs in terms of outcomes, not outputs, and have been generally learning how to use this framework as a management tool. Participants also attributed a series of cultural changes within federal agencies to GPRA, where problem solving, creative thinking, and agencywide discussions on budget and performance have become more common. The strategic and annual performance plan and performance reports that federal agencies are required to submit to OMB and Congress under GPRA have increased the transparency of government activities. These documents have also helped agencies justify their budget requests based on their performance. Creating a Framework in Statute and a Management Tool for Agency Leadership: Participants agreed that GPRA created a framework in statute for federal agencies to plan their activities in order to become more results oriented and provided a managerial tool for program accountability. Using this framework, agencies can develop and focus on strategies to carry out the programs they administer; set goals and identify performance indicators that will inform them whether or not they achieved the performance they expected; and determine what impact, if any, their programs have had on the American public. According to the experts in one of our focus groups, comparing federal agencies' current mission statements contained in their strategic plans to what they were in the past demonstrates that agencies have done some "soul searching" to get a better sense of what their role is (or should be) and how they can achieve it. Given that GPRA is in statute, the use of this planning framework is likely to be sustained within agencies. Participants also mentioned that GPRA has encouraged federal managers to view their programs in terms of results, not just inputs and outputs. Such a change is important, as it has encouraged federal managers to reflect on the statutory intent of their programs and use this framework as a management tool for establishing accountability within their programs. Cultural Change within Federal Agencies: Participants in the majority of focus groups agreed that GPRA has been a driving force behind many cultural changes that have occurred within federal agencies. Highlighting the focus on results, participants stated that GPRA has stimulated a problem-solving approach within federal agencies and encouraged them to think creatively when developing performance indicators for their programs. GPRA has also changed the dialogue within federal agencies; front-line managers and staff at lower levels of the organization now discuss budget issues in connection with performance. Similarly, experts noted that performance management and resource investments are more frequently communicated between agency officials and Congress than in the past. Within agencies, GPRA documents can provide a context of missions, goals, and strategies that political appointees can use to articulate agencies' priorities. Increased Transparency of Government Results: Some participants agreed that GPRA has increased federal agencies' ability to present their results to the American public, benefiting both stakeholders and agency staff. On the one hand, GPRA reports enable federal agencies to communicate the results of government programs and activities to a broad public. For example, GPRA reports are available on agencies' Web sites and provide information to OMB, Congress, and the American public on what agencies plan to do, how they plan to do it, and, as summarized in the performance and accountability reports, what agencies have accomplished and how much money it cost them to do it. Similarly, some participants agreed that GPRA allows federal employees to see exactly how their work can produce a positive outcome, increasing employee morale. Using information contained in GPRA reports, agency employees can compare department goals to the results of their activities, and see how their work contributes to these goals. For example, a focus group participant from the Indian Health Service in California stated that he was pleased to learn that health care indicators of some Native American tribes had already exceeded levels originally projected by his agency to be reached by the year 2010. Link between Budget and Performance: Participants also agreed that the GPRA framework has had a positive effect on agencies' ability to link their performance to their budget. By focusing on their mission and outcomes, agencies are learning to prioritize activities and align their resources to ensure that they will be able to achieve results. For example, a participant stated that the National Wild Horse and Burro Program, managed by the Bureau of Land Management in the U.S. Department of the Interior, recently developed a model for its GPRA-related work which divided their program into discrete components, identifying the results it could accomplish in the short-and long-term and specifying what additional resources were needed. According to the participant, the program received additional funding based on this needs assessment. In addition, a few managers noted that the link between budget and performance has given agencies an incentive to commit the resources necessary to modernize information systems. Having the right information on time enables agencies to integrate budget requests and performance information in ways that are more meaningful. This information also increases Congress's ability to make informed budget decisions based on agency performance. Views on Delivering Results to the American Public Were Mixed: Participants' views on whether GPRA has helped agencies deliver results to the American public were generally mixed. Some federal managers in our focus groups agreed that GPRA has had a positive effect on raising awareness on many issues, and that in and of itself is a way of delivering results. The information gathered and reported for GPRA allows agencies to make better-informed decisions, which improves their ability to achieve results. Of note, GPRA has allowed agencies to move towards performance-based budgeting, which helps agencies identify resources available to use in programs where outcomes can be achieved. For example, programs and expenses that do not add value to the agency's mission could be eliminated. Having performance data readily available is another key area where GPRA has enabled agencies to deliver results to the American public. Other participants stated that while certain aspects of GPRA-related work have been positive, agencies' ability to deliver results and public awareness of their activities cannot always be exclusively attributed to GPRA. For example, while measuring performance is a move in the right direction, GPRA reports provide too many indicators and it is unclear how this has led to better performance among federal agencies. A few participants also stated that agencies deliver results in ways that the American public does not fully recognize. For example, a participant stated that agencies working in the area of international relations generally lack recognition for the results of their activities, as the American public is generally unfamiliar with the nuances of foreign policy and all the programs the U.S. government finances overseas. And while GPRA has helped these agencies prioritize their work to achieve results, it is unclear that GPRA has improved the visibility and understanding in the public eye of what these agencies do because the American public does not see the results of their work. A participant stated that research-based agencies have also used GPRA to plan activities that benefit the American public in ways they are not fully aware of. For example, National Aeronautics and Space Administration's (NASA) space program includes, among other things, predicting whether or not an asteroid will strike the earth, although on a daily basis the American public is probably not worried about such a rarity. For other participants, the link between GPRA and agencies' service delivery was not clear. Participants characterized GPRA-related activities as time consuming, and there is no real evidence that this work has improved their ability to deliver results to the American public. Other participants stated that many agencies rely on grant recipients to carry out their work, and delivering results to the American public depends, to a large extent, on the diligence of these organizations to implement their programs. Overall, they held the view that performance would not change dramatically if GPRA were no longer a requirement for federal agencies. Alternate Views on GPRA's Effect: Participants in one of our focus groups stated that GPRA, per se, had not led federal agencies to achieve specific accomplishments. These participants believed that managers' initiative, not the framework established by GPRA, has been key to improving agencies' planning efforts and focus on results. A few participants also mentioned that the results framework established by GPRA is somewhat intangible unless managers can use it effectively; without the adequate infrastructure to implement GPRA, an agency's compliance with the law is simply paperwork, as GPRA does not allow for a systematic and thorough approach to performance management. For example, while agencies can develop performance indicators to gauge progress towards a program goal, they often encounter problems in collecting relevant data and measuring the agencies' contribution to a specific outcome. In addition, agencies are not able to make changes to the programs they manage, limiting their ability to deliver results. Challenges in Implementing and Overseeing GPRA Activities: Participants stated that agencies face significant challenges in complying with GPRA-related activities. Focus group participants also agreed Congress does not appear to use agencies' performance information when making budget decisions. In carrying out GPRA-related activities, managers find it difficult to identify performance indicators and establish program goals that are results oriented, as required by GPRA. In addition, a few participants stated that they lack the support from senior management to work on GPRA activities. Changes in administration also tend to disrupt progress made by agencies in support of GPRA. Performance Information Not Used: Participants strongly believed that Congress does not take into account agencies' performance information when overseeing agency activities and making budget decisions. As a result, there is a negative effect on how agencies view their efforts in producing GPRA plans and reports--many believe that they are less worthwhile than the effort and resources invested. On the other hand, participants perceive the budget process strictly as a political exercise, and it is unclear how useful performance information can be in this context. Complexity of Establishing Results-Oriented Goals and Measuring Performance: Participants stated that establishing results-oriented goals and identifying performance indicators are generally complex undertakings. Participants agreed that they often feel as if they were trying to measure the immeasurable, not having a clear understanding of which performance indicators could accurately inform the agency how it is carrying out a specific activity. And while agencies generally try to improve the indicators from one year to another, in doing so, they generally lose their ability to develop trend information to track progress made over time. Participants also mentioned that there appears to be a disconnect between some federal programs that generally produce results over a longer time period and GPRA's requirement that agencies report annually on their progress towards achieving their goals. Participants stated that federal programs, especially those that are research-based, often take years to achieve the full scope of their work. Consequently, and for reasons that range from lack of performance data to an absence of short-term outcomes, it could appear as though resources invested in carrying out their activities led to no results in the short run. Focus group participants generally agreed that in cases where third parties, such as states or localities, implement federal programs, some agencies face challenges in obtaining timely performance data from relevant partner organizations. In some instances, federal managers have trouble identifying what the government's contribution has been to a specific outcome. While the experts generally attributed this to agencies not doing a good job at setting appropriate goals and the corresponding measures and objectives not being clear enough, managers stated that their lack of control over grantees and the strict reporting deadlines imposed by GPRA were factors that worked against their efforts to deliver results. Managing Strategically with Frequently Changing Priorities: Some participants stated that it is difficult for them to manage strategically, given the frequently changing priorities that come with changes in administrations. While GPRA requires an agency to develop a strategic plan at least every 3 years to cover the following 5-year period, participants agreed that it makes little sense to update it shortly before a new administration is scheduled to take office. In addition, changes in political leadership generally result in a new agenda with new objectives. These changes force agencies to revise their plans, management initiatives, and strategies, which translate into additional GPRA-related work, generally characterized by focus group participants as a burden agency staff must add to their normal work load. Lack of Top Leadership Support: Some participants stated that they often encounter resistance from agency leadership to endorse GPRA-related activities. In some instances, senior-level support for GPRA falters or is nonexistent. Participants attributed this to the reluctance within some federal agencies to think about outcomes and performance. Some focus group participants stated that in some instances high-level managers are somewhat averse to being held accountable for the results of programs they run. Suggestions to Address GPRA Challenges: To address these challenges, focus group participants made the following suggestions. Congress Should Clarify How Performance Information Could Be More Useful: Congressional staff should provide guidance on agencies' GPRA submissions--specifically, how information could be presented in the reports to make it more useful in the decision-making process. They could also communicate to agencies how the performance information is being used, so that agencies do not perceive their data gathering efforts as inconsequential. Additionally, in using performance information to make budget decisions, Congress should consider the results achieved by agencies in addition to results not achieved. The Administration and OMB Need to Reinforce the Value of GPRA as a Management Tool: Agencies should embrace GPRA as a management tool, not just an external requirement that is separate from their day-to-day activities. To this end, the administration and OMB need make sure agency officials understand how GPRA can be further integrated with other management initiatives. Agency Guidance on GPRA Should Recognize Diversity of Federal Agencies: OMB's guidance to agencies on how to implement GPRA should recognize that one-size-fits-all approaches are unlikely to be useful. OMB should also afford agencies some flexibility by simplifying the reporting process. For example, some participants believed that not everything needed to be measured. OMB should make exceptions for unique situations and programs, e.g., science programs, and it could consider providing multiyear funding for them. OMB Should Publish a Governmentwide Performance Report: OMB should commit to regularly publishing a governmentwide performance report that would articulate the government's accomplishments to the public. It would also be useful if it singled out higher-performing programs so agencies could use them as models to guide their planning efforts. Agencies Need to Obtain and Commit Resources to Carry Out GPRA-Related Activities: Agency leadership needs to ensure that staff members have the necessary resources to work on GPRA-related activities. In addition, they need to invest resources to train staff, including political appointees, on GPRA and the benefits of linking budget to performance. Timing of GPRA Reports Should Take into Account Changes of Administration: Under GPRA, agencies are to update their strategic plans every 3 years. However, this effort can be wasted. Given that federal administrations generally span at least 4 years, participants suggested that the required update be changed to every 4 years to correspond with new presidential terms. Agencies Should Share Experiences on How to Address Common Problems: Agencies should collaborate more to develop strategies to address difficult issues, such as how to identify performance indicators and measure agency contributions to specific outcomes. It would be useful if more agencies created structured forums for managers to share experiences, talk about effective practices, and share solutions. [End of section] Appendix III: Observations on Agencies' Strategic Plans: Under GPRA, strategic plans are the starting point and basic underpinning for results-oriented management. One of our objectives was to assess the changes in the overall quality of agencies' goals, strategies, and data articulated in their strategic plans. To meet this objective, we judgmentally selected six agencies--Education, DOE, HUD, SBA, SSA, and DOT--using criteria such as agency size, primary program types, and previous GAO reviews. To assess the overall quality and improvements made to the agencies' strategic plans, we relied on requirements contained in GPRA and accompanying committee report language,[Footnote 68] guidance to agencies from OMB for developing strategic plans,[Footnote 69] previous GAO reports and evaluations,[Footnote 70] and our knowledge of agencies' operations and programs. In conducting our reviews, we compared our assessments of agencies' current strategic plans to our assessments of draft plans from fiscal year 1997.[Footnote 71] A more detailed discussion of our scope and methodology can be found in appendix I. Required Elements of Agency Strategic Plans: GPRA requires an agency's strategic plan to contain six key elements: 1. A comprehensive agency mission statement. The agency mission statement should concisely summarize what the agency does, as required by law, presenting the main purposes for all its major functions and operations. According to OMB guidance issued in 2002, a mission statement is brief, defining the basic purpose of the agency, and corresponds directly with an agency's core programs and activities. The program goals should flow from the mission statement as well. The federal government's adaptive responses over time to new needs and problems have contributed to fragmentation and overlap in a host of program areas, such as food safety, employment training, early childhood development, and rural development, which could limit the overall effectiveness of the federal effort. The mission statement helps to distinguish agencies' roles from one another and reduce the overlap and identify areas needing coordination and collaboration. 2. Agencywide long-term goals and objectives. General goals and objectives--or strategic goals--explain what results are expected from the agency's major functions and when to expect those results. Thus, such goals are an outgrowth of the mission and are very often results oriented. OMB guidance states that the goals should be defined in a manner that allows a future assessment to be made on whether the goal was or is being achieved. General goals should predominately be outcomes, and are long-term in nature. 3. Approaches or strategies to achieve goals and objectives. Strategies help in aligning an agency's activities, core processes, and resources to support achievement of the agency's strategic goals and mission. Under GPRA, strategies are to briefly describe the operational processes, staff skills, and technologies, as well as the human, capital, information, and other resources needed. According to OMB guidance, descriptions should be brief, but more detailed data should be provided if a significant change in a particular means or strategy would be essential for goal achievement. In addition, the plan should summarize agencies' efforts to provide high-quality and efficient training and skill improvement opportunities for employees. As we have reported previously, agencies' planning processes should support making intelligent resource allocation decisions that minimize, to the extent possible, the effect of funding reductions on mission accomplishment. 4. A description of the relationship between long-term and annual goals. Under GPRA, agencies' long-term strategic goals and objectives are to be linked to their annual performance plans and the day-to-day activities of their managers and staff. Without this linkage, Congress may not be able to judge whether an agency is making progress toward achieving its long-term goals. OMB guidance states that an updated and revised strategic plan should briefly outline (1) the type, nature, and scope of the performance goals being included in annual performance plans and (2) how these annual performance goals relate to the long- term, general goals and their use in helping determine the achievement of the general goals. 5. An identification of key external factors. Identification of key factors, external to the agency and beyond its control that could significantly affect the achievement of the strategic goals, are important for Congress or the agencies to judge the likelihood of achieving the strategic goals and actions needed to better meet those goals. Such external factors could include economic, demographic, social, technological, or environmental factors. Information on these factors can be useful for goal setting and also for explaining results in the agency's annual performance reports, including, when applicable, the reasons annual performance goals were not met. According to OMB guidance, if key factors cannot be identified, a statement of explanation should be included in the plan. 6. A description of program evaluations. Finally, strategic plans should include a description of completed program evaluations that were used in developing the strategic plan, and a schedule for future program evaluations. Program evaluations can be a potentially critical source of information for Congress and others in ensuring the validity and reasonableness of goals and strategies, as well as for identifying factors likely to affect performance. Such information can also be useful in explaining results in an agency's annual performance report, including, when applicable, the reasons annual performance goals were not met, and identifying appropriate strategies to meet unmet goals. Program evaluations are defined in the act as objective and formal assessments of the results, impact, or effects of a program or policy. The evaluations include assessments of the implementation and results of programs, operating policies, and practices. In addition to the six key elements, OMB guidance also states that agencies participating in crosscutting programs should describe in their strategic plans how the programs are related and how coordination will occur to support common efforts. Uncoordinated program efforts can waste scarce funds, confuse and frustrate program customers, and limit the overall effectiveness of the federal effort. OMB guidance also states that the strategic plan should include a brief description of any steps being taken to resolve mission-critical management problems. One purpose of GPRA is to improve the management of federal agencies. Therefore, it is particularly important that agencies develop strategies to address management challenges that threaten their ability to meet long-term strategic goals as well as this purpose of GPRA. As shown in table 5, the majority of agencies have made progress in addressing the required elements of strategic planning under GPRA. Table 5: Agencies' Progress in Addressing Required Elements of Strategic Planning under GPRA: Agency strategic plans: Department of Education; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Education; Plan year: 2002; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Energy; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Energy; Plan year: 2003[A]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Housing and Urban Development; Plan year: 1997; Element included in agency strategic plan? Mission statement: No; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: No; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Department of Housing and Urban Development; Plan year: 2003; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Small Business Administration; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Small Business Administration; Plan year: 2001[B]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Agency strategic plans: Social Security Administration; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Social Security Administration; Plan year: 2003; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long- term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Transportation; Plan year: 1997; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: No; Element included in agency strategic plan? Relationship between long-term goals and annual goals: No; Element included in agency strategic plan? External factors: No; Element included in agency strategic plan? Evaluations: Yes. Agency strategic plans: Department of Transportation; Plan year: 2003[A]; Element included in agency strategic plan? Mission statement: Yes; Element included in agency strategic plan? Long-term goals: Yes; Element included in agency strategic plan? Strategies: Yes; Element included in agency strategic plan? Relationship between long-term goals and annual goals: Yes; Element included in agency strategic plan? External factors: Yes; Element included in agency strategic plan? Evaluations: No. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD- 10.1.16]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R]; and GAO analysis of U.S. Department of Education, Office of the Deputy Secretary, Planning and Performance Management Service, U.S. Department of Education Strategic Plan 2002-2007; (Washington, D.C. 2002); Department of Energy, 2003 Strategic Plan (Draft); (Washington, D.C. 2003); Department of Transportation, U.S. Department of Transportation Draft Strategic Plan for Fiscal Years 2003-2008; (Washington, D.C. 2003); Social Security Administration, Social Security Administration Strategic Plan 2003-2008; (Washington, D.C. 2003); U.S. Department of Housing and Urban Development, Strategic Plan FY 2003-FY 2008; (Washington, D.C. 2003); and Small Business Administration, SBA Strategic Plan, FY 2001-FY 2006; (Washington, D.C. 2000). [A] The 2003 plans for DOE and DOT were in draft form during the time of our review. [B] At the time of our review, the most recent SBA strategic plan was for fiscal years 2001-2008. SBA released a new strategic plan for fiscal years 2003-2008 in October 2003. [End of table] The remainder of this appendix discusses our observations on how the quality of each of the agencies' strategic plans we reviewed has changed since the agencies submitted their first draft strategic plans in 1997. We did not independently verify or assess the information we obtained from agency strategic plans. If an agency chose not to discuss its efforts concerning elements in the strategic plan, it does not necessarily mean that the agency is not implementing those elements. Observations on Changes in the Quality of Education's Strategic Plan: In our review of Education's June 1997 draft strategic plan,[Footnote 72] we found that the plan generally complied with GPRA and included all but one of the six elements required by GPRA; it did not discuss how Education's long-term goals and objectives would be related to its annual performance goals. Also, we observed that the plan presented a logical, fairly complete description of how Education intended to achieve its mission, but a few areas could have been improved. In comparison, Education's 2002-2007 strategic plan has improved on several areas we identified in our 1997 review. However, we still found areas where Education could improve. Table 6 summarizes these findings. Table 6: Education's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Included in initial draft strategic plan: Yes. Mission statement clearly and briefly explained why the agency exists, what the agency does, and how it performs its work; Included in current strategic plan: Yes. Mission statement is outcome oriented, comprehensive, covers all of the agency's functions and activities, and is the same as in the 1997 draft plan. Element of strategic planning: Long-term goals; Included in initial draft strategic plan: Yes. These goals were related to the mission and were results oriented, but did not appear to reflect civil rights enforcement and monitoring responsibilities; Included in current strategic plan: Yes. The goals have changed, but are related to each other and the mission and are results oriented. They now additionally reflect civil rights enforcement and monitoring responsibilities. Element of strategic planning: Strategies; Included in initial draft strategic plan: Yes. The plan outlined strategies to achieve goals overall and to hold managers accountable for achieving objectives, and generally described some of its resource requirements throughout the plan; Included in current strategic plan: Yes. The plan includes strategies that are linked to the goals. Several strategies relate to resource alignment to achieve outcomes, but the actual resources required are not always specified. Element of strategic planning: Relationship between long-term goals and annual goals; Included in initial draft strategic plan: No. Education did not discuss the relationship between its strategic plan goals and those to be included in its annual plan, but indicated this would be done once its annual plan was prepared; Included in current strategic plan: Yes. Plan contains the annual performance measures and targets, which represent the annual performance goals, and aligns them with long-term goals, with which they have a logical relationship. Element of strategic planning: External factors; Included in initial draft strategic plan: Yes. The plan generally described factors outside program scope and responsibilities that could negatively affect Education's ability to achieve goals, but factors were not directly linked to particular goals; Included in current strategic plan: Yes. The plan adequately describes external factors that could affect achieving its goals and they are linked to the particular goals. The plan also briefly discusses strategies to ameliorate the effects of a number of these factors. Element of strategic planning: Evaluations; Included in initial draft strategic plan: Yes. Education said it would provide detailed descriptions of supporting evaluations once it consulted with Congress, completed the strategic plan, and agreed on performance indicators. The plan indicated a commitment to using evaluations, listing evaluations it intended to use to develop sound measures, but did not describe the evaluations; Included in current strategic plan: No. Because of the comprehensive revamping of Education's strategic plan in 2002, its program evaluation plan was completely restructured and was not released until it was included in the 2004 annual plan,[A] which contains information on its new directions for program evaluation. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R]; U.S. Department of Education, Office of the Deputy Secretary, Planning and Performance Management Service, U.S. Department of Education Strategic Plan 2002-2007; (Washington, D.C. 2002). [A] According to OMB Circular No. A-11, Part 6, Preparation and Submission of Strategic Plans, Annual Performance Plans, and Annual Program Performance Reports, June 2002, general goals, which are multiyear and long term, are synonymous with general objectives, and either term can be used. The objectives in Education's strategic plan are such multiyear, long-term objectives, and are referred to in our report as the agency's "long-term goals." OMB's Circular A-11 also indicates that some agencies include strategic goals in their strategic plan, which represent overarching statements of aim or purpose whose achievement cannot be determined and which can be used to group outcome or output goals. Education's strategic goals meet this description. [End of table] Current Strategic Plan Strengths and Improvements from Fiscal Year 1997 Draft Strategic Plan: Education's current mission, "to ensure equal access to education and to promote educational excellence throughout the nation," is the same comprehensive, outcome-oriented mission that was included in its 1997 draft plan. All of Education's functions and activities are covered by it. The plan's long-term goals[Footnote 73] are expressed so as to allow Education and Congress to assess whether they are being achieved. Moreover, in contrast to findings in our review of Education's 1997 Draft Strategic Plan, Education's civil rights responsibilities-- including enforcing five civil right statutes that ensure equal educational opportunity for all students, regardless of race, color, national origin, sex, disability, or age--appear to be reflected, at least in part, in the current plan's goals. For example, one long-term goal is to reduce the gaps in college access and completion among student populations differing by race/ethnicity, socioeconomic status, and disability while increasing the educational attainment of all. Another is to enhance the literacy and employment skills of American adults. Under the latter, the plan includes a strategy to work with state vocational rehabilitation agencies to ensure implementation of standards that will assist individuals with disabilities in obtaining high-quality employment outcomes. As in the past, some goals are targeted at results for which Education has limited direct control and are, instead, greatly influenced by third parties. However, Education recognizes this situation and shows its intent to work closely with its partners. Education's current plan provides some information on linking results and day-to-day activities within the department. For example, the plan says employees will be held accountable for implementation and success from top to bottom and senior officers will have performance contracts linked to the plan and be recognized for achieving results. In addition, the strategy to foster a customer service orientation by ensuring that states, districts, and other partners receive timely responses to inquiries; to assign senior officers to develop relationships with individual states; and to create a customer support team to respond to issues, seems logically linked to the day-to-day activities of managers and staff. However, the link between results and day-to-day activities is not apparent for most of the goals and their strategies. The current plan includes several annual performance measures that are at least related to how well information technology is supporting strategic and program goals, as required by the Clinger-Cohen Act. For example, for its long-term goal to modernize the Federal Student Assistance (FSA) Programs and reduce their high-risk status, the plan contains a measure on the integration of FSA processes and systems that work together to support FSA program delivery functions. Commendably, Education includes some goals related to reducing agency program unintended negative effects, such as a goal and related measure to reduce the data collection and reporting burden. Education's current plan shows great improvement on recognizing and addressing external factors. Beyond adequately describing external factors that could affect achieving its goals and directly linking these factors to particular goals, the plan also briefly describes strategies to ameliorate the effects of a number of factors. For example, for an external factor on teacher certification under a goal on reading, the plan says that Education "will work with the states and national accreditation bodies to encourage the incorporation of research-based reading instruction into teacher certification requirements." In addition, the plan includes strategies indicating that Education monitors some internal factors that could affect achievement of long term goals. Moreover, the plan recognizes the overarching critical external factor for Education--that it depends greatly on third parties who often control the results the department is trying to achieve--and its strategies related to the goal-specific external factors often reflect this. In our 1997 assessment of Education's draft plan, we commented that although the plan identified several management challenges the department would face in the coming years, it provided little detail about them and how they would be addressed. In our January 2001 performance and accountability series, we identified four department- specific and two governmentwide challenges that we said Education needed to meet.[Footnote 74] Education's current strategic plan includes some goals, measures, and strategies that could be used to address these challenges. For example, Education's goal to "develop and maintain financial integrity and management and internal controls" and this goal's strategies and measures are directly related to the management challenge we identified on improving financial management. One of the measures under this goal is "the achievement of an unqualified audit opinion," an important issue in our identification of financial management weaknesses as a challenge in 2001, and Education received an unqualified audit opinion in early 2003. Moreover, the current strategic plan includes goals and strategies meant to improve Education's strategic human capital management and strengthen information technology security, two important governmentwide high- risk areas we identified in 2001. Critical Strategic Planning Issues Needing Further Improvement: In our report on Education's 1997 draft strategic plan, we found that, although the department had numerous crosscutting programs and activities, the plan had identified key interagency activities for some programs but not for others. For example, we said that the plan did not identify or discuss activities for postsecondary programs for which the Higher Education Act of 1965 required coordination. Education's current plan includes an appendix, which highlights collaborative initiatives under each of the department's strategic goals, with some activities related to postsecondary education, including most of those mentioned in our 1997 report. However, as the plan states, the appendix presents a brief overview of the highlights of some of its collaborative initiatives with other agencies. In our 1997 review, we stated that some resource requirements were described throughout the plan. In the current plan, while a number of strategies under the department's strategic goal to establish management excellence are related to aligning resources to achieve outcomes, the actual resources required--such as human, capital, and information--are not always specifically indicated. The exception is for information resources, for which a number of strategies discuss the information resources that will be required to address the related goal. For example, under the goal to develop and maintain financial integrity and management and internal controls, the plan says that Education will "implement a new financial system capable of producing timely and reliable financial data and reconcile systems to the general ledger." Moreover, while the plan stresses accountability throughout, it only refers to providing the authority needed to achieve results once.[Footnote 75] In addition, consideration of alternate strategies for achieving goals is not discussed. In our 1997 review, we reported that Education said it would provide detailed descriptions of supporting evaluations once it consulted with Congress, completed the strategic plan, and agreed on performance indicators. The draft plan indicated Education's commitment to using evaluations and listed evaluations and strategies it intended to use to develop sound measures, but did not describe the evaluations. The current plan does not include descriptions of supporting evaluations or a future program evaluation schedule. According to Education officials, this was not done because, with the complete revamping of the strategic plan based on passage of the No Child Left Behind Act of 2001,[Footnote 76] Education was set on a course to totally restructure its evaluation program, but could not do so in time to include it in the strategic plan. Consequently, Education instead included information about its new directions for program evaluation studies and a schedule of evaluations in its 2004 annual plan. The schedule, however, lacked a timetable, except for stating whether the evaluations were new or continuing. The current strategic plan indicates, in some cases, the use or planned use of program evaluation findings to develop or revise components of the plan. For example, for long-term goals on various types of academic achievement, data from the National Assessment of Educational Progress were identified as having been used to set related targets. Observations on Changes in the Quality of DOE's Strategic Plan: Overall, DOE's draft 2003 strategic plan meets the required elements of GPRA and has improved greatly over its 1997 draft strategic plan, as shown in table 7. DOE made improvements to its plan by establishing results-oriented and measurable objectives, and addressing elements that were not included in the department's 1997 plan, such as reporting on external factors. Although DOE has shown improvement, a few elements in the plan could still be enhanced. For instance, further improvement could be made to the mission statement so that it better addresses the department's major activities, and additional information could be included in DOE's strategies to achieve its goals, such as management accountability. Table 7: DOE's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Included in initial draft strategic plan: Yes. DOE's mission was results oriented, met a public need, and covered the agency's major activities; Included in current draft strategic plan: Yes. DOE's mission is results oriented and meets a public need, but it does not address the department's major activities related to energy supply and conservation. Element of strategic planning: Long-term goals; Included in initial draft strategic plan: Yes. Long-term goals covered mission and major functions of the agency. Goals and objectives were results oriented but not measurable; Included in current draft strategic plan: Yes. Long- term goals cover the mission and major functions of the agency. Objectives, referred to as intermediate goals, are results oriented and measurable. Element of strategic planning: Strategies; Included in initial draft strategic plan: Yes. The plan included strategies and measures to evaluate the results of the strategies, but was missing information on linkages to day-to-day activities, and the extent to which managers have the knowledge, skills, and abilities to implement the strategies; Included in current draft strategic plan: Yes. Strategies to achieve each goal are included in the plan, along with intermediate goals to measure success, but information on linkages to day-to-day activities, and the extent to which managers have the knowledge, skills, and abilities to implement the strategies, is not included. Element of strategic planning: Relationship between long-term goals and annual goals; Included in initial draft strategic plan: No. Relationship between the long-term goals and annual performance goals was missing in the draft plan; Included in current draft strategic plan: Yes. The draft plan provides a brief description of the relationship between the long-term strategic goals and the annual performance goals. Element of strategic planning: External factors; Included in initial draft strategic plan: No. Key external factors were not addressed in the draft plan; Included in current draft strategic plan: Yes. Key external factors and how they could affect the ability to achieve each goal are identified in the draft plan. Element of strategic planning: Evaluations; Included in initial draft strategic plan: No. The impact of program evaluations on the development of strategic goals was not included in the draft plan; Included in current draft strategic plan: No. The impact of program evaluations on the development of strategic goals is not discussed thoroughly. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R]; and Department of Energy, 2003 Strategic Plan; (Draft); (Washington, D.C. 2003). [End of table] Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan: In its 2003 draft strategic plan, DOE has made significant improvements in developing measurable objectives, referred to in its plan as intermediate goals. GAO's review of DOE's 1997 draft strategic plan found that objectives related to DOE's long-term goals were stated in ways that would make it difficult to measure whether they were being achieved. In the 2003 draft plan, the objectives are stated in ways that will enable DOE to measure its progress in achieving goals. For example, to meet the goal of enhancing energy security through various means, one of DOE's affiliated objectives is to ensure that throughout DOE's 25-year planning period, the Strategic Petroleum Reserve is ready to supply oil at a sustained rate of 4.3 million barrels per day for 90 days within 15 days notice by the President. In addition, DOE has improved its draft 2003 strategic plan by including elements that were not included in its 1997 draft plan. These elements consisted of (1) identifying external factors and (2) describing the relationship between long-term and annual goals. For each of its long-term goals, DOE identified external factors, such as reduced funding, lack of scientific talent, and unpredictable technological developments, that could affect its ability to achieve its goals. The strategic plan also included a description of the relationship between the long-term strategic goals and the annual performance goals. The plan included a diagram that showed an example of a strategic goal, its associated objectives, and how they are related to the annual performance goals and targets. However, the plan could be improved if all annual performance goals were discussed in the plan and linked to each strategic goal so that it would be clear how annual performance goals would be used to gauge performance. DOE staff stated that a description of all actual annual performance goals was not something that they thought should be included in the strategic plan because the annual goals differ each year. Finally, our past and current reviews of DOE's 1997 and 2003 draft strategic plans found that DOE addressed performance and accountability challenges that we had previously identified. In January 2003, we identified six areas where DOE's management attention was needed: (1) addressing security concerns, (2) revitalizing infrastructure, (3) improving contract management, (4) managing the nuclear weapons stockpile, (5) cleaning up radioactive and hazardous wastes, and (6) enhancing leadership in meeting energy needs. All areas were addressed in the 2003 draft strategic plan, with the exception of improving contract management. For example, for the challenge of enhancing leadership in meeting energy needs, one of the intermediate goals requires DOE to develop and demonstrate technologies that can reduce emissions by more than 70 metric tons of carbon and equivalent greenhouse gases by 2012. Critical Strategic Planning Issues Needing Further Improvement: There are three elements in DOE's draft 2003 strategic plan that requires further improvement. To begin with, although DOE's mission is results oriented, it was revised from the 1997 draft strategic plan and does not address the department's major activities related to energy supply and conservation. These activities account for approximately 10 percent of DOE's $23.4 billion fiscal year 2004 budget request. Our review of the 1997 draft strategic plan found that DOE's mission addressed all of its major activities. In addition, the impact of program evaluations on the development of strategic goals could be discussed in greater detail. The strategic plan stated that internal, GAO, and the Inspector General (IG) evaluations were used as resources to develop the draft strategic plan, but specific program evaluations were not identified. A schedule of future program evaluations was also not discussed in the strategic plan. According to DOE, there is no plan to include a table of reports and evaluations that were used to develop the goals for the strategic plan because the evaluations were from a prior administration, and when the administration changes, it usually does not use evaluations from past administrations. DOE could also enhance the strategies included in its plan by providing information on how goals are linked to the department's day-to-day activities, and the extent to which managers have the knowledge, skills, and abilities to implement the strategies. None of this information was included in the 1997 and 2003 draft strategic plans. According to DOE officials, in drafting the plan, their goal was to keep the plan at a high level and this additional information would require more detail than what is needed for a strategic plan. For example, one official stated that a description of linkages to day-to- day activities was not discussed in the strategic plan because it would conflict with the performance plan and budget justification, which is where the information can be found. As we have stated previously, without this information, it is difficult to judge DOE's likelihood of success in achieving the goals or the appropriateness of the strategies.[Footnote 77] Finally, DOE's draft strategic plan could be improved by identifying programs and activities that are crosscutting or similar to those of other agencies. In the 2003 draft plan, crosscutting activities are only identified for one goal related to science. According to one DOE official, as the strategic goals were being developed, DOE staff took crosscutting activities into consideration. As we stated in 1997, unless DOE addresses crosscutting issues in its plan, Congress cannot be assured that federal programs are working effectively. Observations on Changes in the Quality of HUD's Strategic Plan: Overall, HUD has made progress in developing its strategic plan for fiscal years 2003 through 2008 as required under GPRA. In 1997, we stated that HUD's draft strategic plan did not cover the six components required under GPRA.[Footnote 78] HUD's fiscal year 2003-2008 strategic plan addressed several issues we had previously identified, such as making sure the mission statement is linked to the department's major operations and functions; including outcome-related goals and objectives for the department's functions and operations; generally providing a description of how it will achieve its goals and objectives; identifying key external factors affecting the achievement of departmental goals; and explaining how HUD will coordinate with other agencies to address crosscutting issues. However, HUD could improve its strategic plan by explaining the relationship between the long-term and intermediate performance measures listed for each strategic goal and discussing how it used program evaluations to develop the plan. Table 8 summarizes these findings. Table 8: HUD's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Included in initial draft strategic plan: No. HUD's mission statement did not cover the major program functions and operations of the agency; Included in current strategic plan: Yes. HUD's mission statement covers the agency's major program functions and operations and relevant statutes. Element of strategic planning: Long-term goals; Included in initial draft strategic plan: Yes. The plan included eight strategic objectives, and they generally covered the department's major functions and operations; Included in current strategic plan: Yes. The plan includes long-term goals that cover the major programs and functions of the agency. Element of strategic planning: Strategies; Included in initial draft strategic plan: No. The strategic plan lacked an adequate description of how its strategic objectives would be achieved; Included in current strategic plan: Yes. The strategic plan discusses the means and strategies to address each strategic goal, including plans to address human capital issues critical to carrying out its mission. Element of strategic planning: Relationship between long-term goals and annual goals; Included in initial draft strategic plan: No. HUD's strategic plan provided limited examples of annual performance goals under each of its strategic objectives, but it did not describe the relationship between them; Included in current strategic plan: Yes. However, some long-term performance measures do not appear to have corresponding intermediate measures, and in other instances it is not clear how HUD will measure progress towards its goals. Element of strategic planning: External factors; Included in initial draft strategic plan: No. HUD briefly discussed the external factors in its draft strategic plan without linking them to specific strategic objectives; Included in current strategic plan: Yes. HUD describes the external factors that could affect achieving its strategic objectives. Element of strategic planning: Evaluations; Included in initial draft strategic plan: No. HUD's strategic plan did not include information on program evaluations; Included in current strategic plan: No. HUD does not describe how program evaluations were used to prepare its strategic plan. Although it mentions that program evaluations will be used to advance key policy objectives, it states that there are no fixed timetables for when these evaluations will take place. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R]; and U.S. Department of Housing and Urban Development, Strategic Plan FY 2003-FY 2008.; (Washington, D.C. 2003). [End of table] Strategic Plan Strengths and Improvements from Fiscal Year 1997 Draft Plan: In its most recent strategic plan for fiscal years 2003-2008, HUD has made progress in crafting a mission statement that generally covers its major program functions, operations, and relevant statutes that authorize its programs. In addition, the current plan builds upon the strength of its first draft strategic plan by including the strategic objectives that cover the department's major functions and operations. In contrast to its draft strategic plan of 1997, HUD's most recent plan provides a general description of how its strategic objectives will be achieved. For example, HUD lists the means and strategies following each strategic goal and supporting objectives to describe how it will carry out its activities. In addition, the strategic plan identifies a few legislative and regulatory changes HUD will pursue to meet its objectives, such as a new tax credit for developers of affordable housing and expanded eligibility for the Assisted Living Conversion Program. To carry out its work, HUD acknowledges in its plan that it needs to recruit and retain its current workforce to ensure the proper skills and abilities needed to carry out its mission. HUD's most recent strategic plan also describes key factors external to the department and beyond its control that could significantly affect the achievement of its objectives. For example, for its goal of Promoting Decent and Affordable Housing, HUD identifies the impact of broad economic factors on opportunities for low-income workers as a factor that will affect the department's ability to assist renters that depend on HUD's programs to make progress towards self sufficiency. HUD's strategic plan also provides a general description of current program evaluations and mentions that the results of these evaluations will support key policy objectives within the department. Addressing a shortcoming of its draft strategic plan in 1997, HUD's current plan generally explains how it will coordinate with other agencies to address crosscutting problems. For example, the Interagency Council on the Homeless (ICH), created by the Secretaries of HUD, the Department of Health and Human Services (HHS), and Veterans Affairs (VA), will continue to work to identify the obstacles homeless people face to enroll in the main service programs and recommend specific changes--legislative, policy, and procedural--that would make federal supportive service programs more accessible to them. Lastly, HUD improved its discussion of how it plans to address performance and accountability challenges we have raised. In January 2003, we reported that programmatic and financial management information systems and human capital issues were HUD's performance and accountability challenges.[Footnote 79] While in the past HUD acknowledged some of these problems and described how they would be addressed, its plans for management reform were not fully integrated into its draft strategic plan. In contrast, one of HUD's strategic goals in its current strategic plan, "Embrace High Standards of Ethics, Management, and Accountability," lists specific actions the department will take to address these challenges. Critical Strategic Planning Issues Needing Further Improvement: HUD has made progress in linking its strategic objectives to both long- term goals and intermediate measures in its strategic plan; however, long-term performance measures are not consistently linked to a corresponding intermediate measure, making it difficult for the reader to understand how the department will measure progress towards its strategic goals. For example, HUD projects that the percentage of architects and builders indicating awareness of the design and construction requirements of the Fair Housing Act will increase through fiscal year 2008. However, HUD does not mention in the plan how many architects or engineers it will survey to establish a baseline of awareness or to gauge progress made towards this goal in subsequent years. HUD officials explained that only those long-term goals that are critical to being achieved within the next 1-2 years have corresponding intermediate performance measures. Therefore, it is somewhat unclear how HUD plans to assess progress made towards its broader goal of "Ensure Equal Opportunity in Housing.": Similarly, the strategic plan does not always provide a clear picture of how HUD will be able to measure progress towards its strategic goals. For example, the plan states that HUD's Community Development Block Grants (CDBG) Program will create or retain 400,000 jobs by fiscal year 2008; HUD's intermediate measure for fiscal year 2004 is 84,000 jobs. These estimates are based on the average jobs created per grant dollar reported by grantees. However, HUD does not mention how it will be able to discern between those jobs created by CDBG and those created by other means. Similar to our previous finding, HUD's current strategic plan does not describe how program evaluations were used to develop its strategic goals or other components of its strategic plan, and does not include a schedule for future evaluations. While the plan mentions that program evaluations will be used to advance key policy objectives, it states that there are no fixed timetables for when these evaluations will take place. Observations on Changes in the Quality of SBA's Strategic Plan: In our review of SBA's March 1997 draft strategic plan,[Footnote 80] we found that the plan did not meet two of the six requirements set forth by GPRA: (1) a discussion of the relationship between the long-term goals and objectives and the annual performance goals and (2) a description of how program evaluations were used to establish or revise strategic goals and a schedule for future program evaluations. In addition, the four elements the plan contained could have better conformed to GPRA's requirements and OMB guidance. In comparison, SBA's 2001-2006 strategic plan has improved on several areas we identified in our 1997 review.[Footnote 81] However, we still found areas that SBA could improve. Table 9 summarizes these findings. Table 9: SBA's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Initial draft strategic plan: Yes. The mission statement was results oriented. However, it did not directly incorporate key aspects of SBA's legislative mandate and it did not encompass one of SBA's significant activities; Current strategic plan: Yes. SBA's mission statement now specifically mentions its mandate and encompasses SBA's disaster loan program to families and businesses. Element of strategic planning: Long-term goals; Initial draft strategic plan: Yes. Generally the strategic goals covered the major functions and operations of SBA. Most of SBA's strategic goals were expressed as processes, not as outcomes; Current strategic plan: Yes. The strategic goals are outcome oriented and cover the major functions and operations of SBA. Element of strategic planning: Strategies; Initial draft strategic plan: Yes. However, the strategies were too vague or general to assess whether or not they would help achieve SBA's goals. Also, the strategies and goals could have been linked more explicitly; Current strategic plan: Yes. The strategies will generally help SBA achieve its goals. The strategies are now listed by strategic goal and related objective. The plan discusses some of the resources SBA needs to achieve its goals. Element of strategic planning: Relationship between long-term goals and annual goals; Initial draft strategic plan: No. The linkage between proposed performance measures, strategies, and objectives was unclear; Current strategic plan: Yes. The plan lists performance measures by objectives for each strategic goal. Logic models show the relationship between measures and outcomes. Element of strategic planning: External factors; Initial draft strategic plan: Yes. The plan listed a number of external factors. A discussion of how the external factors would be taken into account when assessing progress toward goals was not included in the plan; Current strategic plan: Yes. The plan lists several external factors. Generally, strategies to ameliorate the effects of these factors are included. However, SBA does not discuss how external factors will be taken into account when assessing progress toward goals. Element of strategic planning: Evaluations; Initial draft strategic plan: No. The plan did not describe how program evaluations were used to establish or revise strategic goals or include a schedule for future program evaluations; Current strategic plan: No. The plan states that lessons learned from SBA's program evaluations have influenced its strategic direction. Examples of future program evaluations are given, but a specific schedule of when these evaluations are to occur is not included. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R]; and Small Business Administration, SBA Strategic Plan, FY 2001-FY 2006; (Washington, D.C. 2000). [End of table] Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan: SBA's current mission statement is an improvement from the one contained in its 1997 draft strategic plan. In our July 1997 report on SBA's draft strategic plan, we noted that the mission statement could be improved by more directly incorporating key aspects of SBA's legislative mandate to aid, counsel, assist, and protect the interests of small businesses. In addition, the mission statement did not encompass one of SBA's significant activities-making loans to individuals. The mission statement in the 2001-2006 strategic plan now includes both of these items. The long-term, or strategic, goals of SBA's 2001-2006 strategic plan are (1) helping small businesses succeed, (2) helping Americans recover from disasters, and (3) modernizing SBA. These three strategic goals are outcome oriented. We stated in our previous report on SBA's draft strategic plan that only two of SBA's seven strategic goals described outcomes. The rest of the goals were expressed as processes. SBA's 2001-2006 strategic plan shows significant improvement in its strategies. In 1997 we stated that the plan could be improved by making the linkage between the strategies and goals/objectives more explicit. Objectives were listed as a group, followed by the strategies, which were also listed as a group. The 2001-2006 plan describes the strategies, objective by objective, making the linkage clear. The strategies contained in SBA's 1997 plan consisted entirely of one-line statements and most were too vague or general to enable an assessment of whether or not they would help achieve the goals and objectives in the plan. While many of the strategies listed in SBA's 2001-2006 plan are only one-or two-sentence statements, all of the strategies seem to aid in achieving SBA's goals and objectives. For example, one strategy for the objective of providing entrepreneurial development assistance is to train SBA personnel in outreach and business development approaches. This strategy would presumably aid in making staff more competent at helping small businesses develop. The current plan also includes some of the resources needed to achieve SBA's goals. These resources include a table showing SBA's fiscal year 2001 budget request and its plans to modernize its information technology systems. The 2001-2006 strategic plan also makes a clearer connection between strategic goals and annual performance measures that will be used to gauge progress in achieving strategic goals. SBA shows this relationship by linking its performance output measures with the intended outcomes of SBA's programs for two of the strategic goals. For the other strategic goal, "Modernizing SBA," SBA lists the performance measures it will use by objective. For example, the performance measure "personnel trained or retrained" will gauge progress in achieving SBA's human capital investments objective. This shows improvement over the draft strategic plan, which listed the performance measures as a group without showing how they were related to SBA's strategic goals. Similar to its 1997 draft plan, SBA's 2001-2006 plan specifies external factors that could affect the achievement of its strategic goals. The plan lists eight external factors associated with two of its three strategic goals; no external factors were associated with the third goal. The plan includes strategies that seem to be intended to mitigate the effect of six of these external factors, while two of the external factors, congressional support and public support, do not seem to be addressed. An example of an external factor that seems to be addressed by a strategy is economic conditions, listed for the strategic goal "helping small businesses succeed." One of the strategies for this strategic goal is to determine economic trends and conditions. SBA states that it tracks economic trends that affect small businesses and the contribution small businesses make to the economy. SBA then brings these data to the attention of policymakers. We noted in our 1997 report that SBA did not explicitly address the relationship of SBA's activities to similar activities in other agencies and provided no evidence of coordination. In contrast, SBA's current strategic plan includes a section on crosscutting issues. This section contains discussions of innovation and research assistance, international trade assistance, business development assistance, veterans affairs, and disaster assistance. One example of coordination is the U.S. Export Assistance centers, which combine the trade- promotion and export-finance resources of SBA, the Department of Commerce, the Export-Import Bank, and in some locations, the Agency for International Development and the Department of Agriculture. Finally, SBA's 2001-2006 strategic plan generally addresses performance and accountability challenges that we have previously identified. For example, we noted in our January 2001 report on major management challenges and program risks[Footnote 82] that SBA needed to continue to improve oversight of its lending partners to correct oversight weaknesses. The plan identifies lender oversight as a management challenge and states that SBA has developed and implemented a safety and soundness oversight program for Small Business Lending companies, institutionalizing the process through the Office of Lender Oversight. This is an improvement over SBA's draft strategic plan, which we reported did not clearly address previously identified management problems. Critical Strategic Planning Issues Needing Further Improvement: In 1997, we noted that SBA's draft strategic plan did not include a discussion of how the external factors would be taken into account when assessing progress toward goals. This observation holds true for the current strategic plan. For the external factor mentioned above, economic conditions, the plan states that if the economy remains strong, surety bond guaranties will remain constant or decrease, but if the economy deteriorates, demand will increase. However, the plan does not state how SBA will assess success or failure in meeting its goals in relation to this factor. In its 1997 draft plan, SBA did not describe how program evaluations were used to establish or revise strategic goals or include a schedule for future program evaluations. In the 2001-2006 plan, SBA states, "We have used lessons learned in our performance monitoring and program evaluations to influence the strategic direction contained in this plan." The plan includes findings from six completed program evaluations; however, no further detail is given as to how these program evaluations were used to establish or revise the strategic goals. While the current plan gives examples of future program evaluations, such as conducting a benchmark study on the HUBZone program to assess the changes in employment and investment in distressed urban and rural communities, it does not include a schedule of future evaluations. SBA states that for the next several years, the agency plans to systematically review programs that offer the most financial risk to the government and also the programs that can offer tips on how to improve efforts. Observations on Changes in the Quality of SSA's Strategic Plan: SSA's strategic plan for 2003-2008 is well structured and contains all of the required elements under GPRA. In 1997, we noted that SSA's draft strategic plan contained all six required components, but suggested a number of ways it could be strengthened. SSA has addressed some of the issues we previously identified, such as associating specific programs with goals and identifying external factors that may affect goal achievement. SSA could further improve its strategic plan through (1) ensuring that its strategic objectives will assist SSA in achieving its strategic goals, (2) explicitly describing the effect of external factors on goal attainment, (3) providing timetables or schedules for achieving results, (4) providing details on how each performance and accountability challenge will be addressed, (5) clearly explaining how program evaluations were used in formulating the strategic plan, and (6) discussing the manner in which SSA has coordinated with other agencies, especially those that serve the same beneficiaries. Table 10 summarizes SSA's progress in addressing the required elements of GPRA. Table 10: SSA's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Included in initial draft strategic plan: Yes. SSA's mission statement was appropriate and reflective of its new status as an independent agency; Included in current strategic plan: Yes. SSA's mission statement has not changed substantially. Element of strategic planning: Long-term goals; Included in initial draft strategic plan: Yes. Long-term goals were established. However, the relationship between long-term goals and specific programs was unclear and did not identify the results to be achieved; Included in current strategic plan: Yes. The goals' relationship to specific programs is more defined, but SSA's goal for achieving solvency of the social security system is ambitious, given SSA's mission and responsibilities. Key outcomes are identified for each goal. SSA's major management challenges are not all clearly linked to individual goals or objectives. Element of strategic planning: Strategies; Included in initial draft strategic plan: Yes. The strategic plan was generally complete with regard to processes and technology, but did not include timetables or schedules. Some of the success was predicated on changes in processes or technology improvements; Included in current strategic plan: Yes. SSA added some timetables, but the required resources are not specified. Some of the success is still predicated on technological improvements. Element of strategic planning: Relationship between long-term goals and annual goals; Included in initial draft strategic plan: Yes. SSA provided numerous performance measures relating to strategic goals and objectives, and plans for developing new measures were discussed. It was sometimes difficult to link measures with their appropriate objectives. It was also hard to discern which objectives did not yet have performance goals. Some data were expressed by program, while other data were aggregated; Included in current strategic plan: Yes. SSA provided one or more key outcomes for each strategic objective. SSA notes that success in meeting the objectives will be measured in the annual performance plans by progress in achieving the key outcomes. Element of strategic planning: External factors; Included in initial draft strategic plan: Yes. The report mentioned several key external factors, but did not explicitly link factors to general goals and state how they could have affected goal attainment. Also, the report did not discuss needed changes (by Congress) to ensure solvency; Included in current strategic plan: Yes. External (environmental) factors are listed in a separate section. However, the plan does not explicitly link factors to general goals and state how they could affect goal attainment. Specific effects are not discussed--most examples are vague. Element of strategic planning: Evaluations; Included in initial draft strategic plan: Yes. The report contained a broad discussion of program evaluations, but the evaluations were not clearly described. Also, SSA did not describe how the evaluations were used to establish or revise specific goals/objectives. Finally, there was no schedule for completing future evaluations or methodologies; Included in current strategic plan: Yes. Future evaluations (with brief descriptions) are listed, but there is no discussion of how current evaluations are used to establish or revise specific goals/objectives. The plan states that SSA used internal and external (GAO, IG) evaluations to determine strategic plans and objectives. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R]; and Social Security Administration, Social Security Administration Strategic Plan 2003-2008, (Washington, D.C. 2000). [End of table] Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan: SSA's mission statement changed very little between 1997 and 2003. OMB Circular A-11 notes that the mission statement should be brief and define the basic purpose of the agency, with particular focus on its core programs and activities. SSA's statement conforms to this guidance--it reads, "To advance the economic security of the nation's people through compassionate and vigilant leadership in shaping and managing America's social security programs."[Footnote 83] In 1997, the relationship between SSA's long-term goals and specific programs was unclear and did not identify the specific results to be achieved. Since that time, SSA has improved this linkage and better articulated intended results, including quantifiable goals. For example, as part of its strategic goal to ensure superior stewardship of Social Security programs and resources, SSA notes that one of its key outcomes is to increase Supplemental Security Income (SSI) payment accuracy to 96 percent (free of preventable error) by 2008. SSA improved upon its linkage between long-term goals and annual goals in its fiscal year 2003-2008 strategic plan. Under each strategic goal in this plan, SSA provided one or more key outcomes for each strategic objective; in its 1997 draft strategic plan, we found it difficult to link measures with the appropriate objectives and discern which objectives did not yet have performance goals. Critical Strategic Planning Issues Needing Further Improvement: Not all of SSA's strategic objectives and associated performance measures will allow SSA to achieve its related strategic goals. Specifically, the solvency goal in SSA's current strategic plan reads, "To achieve sustainable solvency and ensure Social Security programs meet the needs of current and future generations," but the sole associated objective--through education and research efforts, support reforms to ensure sustainable solvency and more responsive retirement and disability programs--will not allow SSA, on its own, to reach that goal. While SSA's mission is to advance the economic security of the nation's people, it is not unilaterally responsible for achieving solvency in social security programs. An agency's strategic plan is expected to contain strategies for achieving the goals articulated. In 1997, SSA's strategic plan was generally complete with regard to processes and technology, but did not include timetables or schedules for results. While the current strategic plan contains processes, anticipated progress in technology, and some timetables, it does not contain timetables or schedules for all of the results. For example, as part of its strategic objective to "efficiently manage Agency finances and assets and effectively link resources to performance outcomes," SSA's key outcomes include (1) competing or converting 50 percent of commercial positions and (2) to "get to green" on all five President's Management Agenda (PMA) items. SSA has neither identified the required resources to achieve these goals nor has it identified a time frame for achieving them. In our review of SSA's 1997 draft strategic plan, we noted that SSA described several key external factors that may affect its programs, but did not explicitly link such factors to its general goals and state how these factors could affect goal attainment. In the current strategic plan, SSA identifies environmental (external) factors: demographics, health and disability trends, technological advances, and workforce trends. However, as we found in our earlier review, the effects of these factors on specific performance goals are not specified, even though SSA notes that they drive the development of such goals. SSA noted that it considered major management challenges identified by GAO when it determined its strategic goals and objectives, but not all of these challenges are clearly addressed in the plan. While these challenges are not clearly identified, SSA addresses them to some degree throughout the plan. For example, SSA's strategic goal to "Strategically manage and align staff to support SSA's mission" addresses the governmentwide challenge of strategic human capital management.[Footnote 84] SSA includes a list of major strategic process and program evaluations scheduled for the fiscal years 2003-2008 time period, organized by strategic goal. However, SSA does not list ongoing evaluations or mention how the results of these evaluations were used to prepare the current strategic plan. SSA notes that many of the hundreds of process and program evaluations conducted annually were designed to evaluate and improve internal processes falling below the strategic level. However, some of the ongoing evaluations are associated with specific strategic goals; thus, their outcomes could be discussed in the context of the strategic goals with which they are affiliated. SSA's strategic plan contains a very limited discussion of its interactions with other agencies that have similar goals or serve the same beneficiaries. In an interview, SSA officials noted that SSA has extensive interactions with other agencies on such issues as earnings accuracy and medical information, but the level of interaction varies by initiative. SSA's strategic plan would benefit from a broader discussion of these interactions, especially if they were broken down by initiative. For example, as part of the objective to increase the accuracy of earnings records, SSA notes that it will collaborate with the Internal Revenue Service (IRS) to achieve more accurate wage reporting as part of its means and strategies to reduce the size of the suspense file.[Footnote 85] It would be helpful if SSA offered more details as to the nature and extent of its collaboration with IRS. Observations on Changes in the Quality of DOT's Strategic Plan: In our review of DOT's July 1997 draft strategic plan, we found that the plan only met three of the six elements required by GPRA.[Footnote 86] The plan did not meet GPRA's requirements to describe (1) strategies for achieving the goals, (2) a linkage between DOT's long- term goals and annual performance goals, and (3) the external factors that could significantly affect DOT's ability to achieve its goals. Further, for the three elements that the plan did meet, each had weaknesses. In comparison, DOT's 2003-2008 draft strategic plan has improved on several areas we identified in our 1997 review.[Footnote 87] However, we still found areas where DOT could improve. Table 11 summarizes these findings. Table 11: DOT's Progress in Addressing Required Elements of Strategic Planning under GPRA: Element of strategic planning: Mission statement; Included in initial draft strategic plan: Yes. DOT's mission statement was comprehensive and covers its major functions and operations; Included in current draft strategic plan: Yes. The mission statement continues to cover its major functions and operations and more explicitly states DOT's statutory authority. Element of strategic planning: Long-term goals; Included in initial draft strategic plan: Yes. Five long-term goals encompassed DOT's major functions and operations. However, it was not clear as to how DOT would measure success for most of its goals; Included in current draft strategic plan: Yes. Five strategic objectives encompass DOT's major functions and operations. Outcome goals and candidate performance measures for each strategic goal help show how DOT will measure success. Element of strategic planning: Strategies; Included in initial draft strategic plan: No. While the plan listed six corporate management strategies for achieving its long-term goals, it did not describe the operational processes, the skills, the technology, and the resources required to meet them; Included in current draft strategic plan: Yes. Strategies are listed by strategic objective and include discussions on leadership, building expertise, and technology. Element of strategic planning: Relationship between long term-goals and annual goals; Included in initial draft strategic plan: No. The plan did not describe how performance goals would be related to the long- term goals; Included in current draft strategic plan: Yes. The plan includes performance measures and refers to the performance plan for further information on annual performance goals. Element of strategic planning: External factors; Included in initial draft strategic plan: No. Four external factors were identified, but other key factors were not included in the plan. Only one external factor was discussed in terms of how it could have affected DOT's ability to accomplish its goals; Included in current draft strategic plan: Yes. Several external factors are listed for each strategic objective. Generally the plan gives descriptions of how these factors could affect the achievement of goals. Element of strategic planning: Evaluations; Included in initial draft strategic plan: Yes. Program evaluations used in establishing goals and a schedule of future evaluations were discussed. However, the plan did not provide enough information to determine the scope and methodology or the key issues to be addressed in future evaluations; Included in current draft strategic plan: No. An extensive table describes the scope and methodology and the completion date of program evaluations for fiscal years 2003-2008. However, the plan does not specifically mention which or how previous evaluations were used in developing the plan. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R]; and U.S. Department of Transportation, U.S. Department of Transportation Draft Strategic Plan for Fiscal Years 2003-2008; (Washington, D.C. 2003). [End of table] Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan: The mission statement contained in DOT's 2003-2008 draft strategic plan is an improvement over the one contained in the 1997 draft plan. DOT's mission, as stated in the 2003-2008 draft strategic plan, is "to develop and administer policies and programs that contribute to providing fast, safe, efficient, and convenient transportation at the lowest cost consistent with the national objectives of general welfare, economic growth and stability, the security of the United States, and the efficient use and conservation of the resources of the United States." The mission statement covers the major functions and operations of the department. In our July 1997 report on DOT's 1997 draft strategic plan, we noted that the mission statement could be improved by including language from the department's enabling legislation to focus the mission statement more directly on DOT's core activities. We gave an example of adding the department's purpose to develop transportation policies and programs that "contribute to providing fast, safe, efficient, and convenient transportation at the lowest cost" from DOT's enabling legislation. The mission statement in the 2003-2008 plan includes such language. As in its 1997 draft strategic plan, DOT's 2003-2008 draft strategic plan meets the requirement of GPRA to include long-term goals and objectives for the major functions and operations of the department. The 2003-2008 draft strategic plan contains five strategic objectives (long-term goals) that cover the major functions and activities of the department and are results oriented. Besides the strategic objectives of "safety," "mobility," "global connectivity," "environmental stewardship" and "security," the current draft also contains an "organizational excellence" objective to "advance the department's ability to manage for results and achieve the goals of the PMA.": Each strategic objective section in the 2003-2008 draft plan contains strategies for attaining DOT's outcomes and objectives. The strategies for each strategic objective are listed in the categories of "leadership," "building expertise," and "technology." For example, a "technology" strategy for the "mobility" strategic objective is to "examine ways to encourage cargo movements by water through the development of barge and fast vessel technologies to bring new capacity to our intermodal transportation system." The plan states that this strategy supports DOT's outcomes of reduced congestion in all modes of transportation and increased reliability throughout the system. The strategies for each strategic objective generally describe the operational processes, the skills, and the technology required to meet DOT's goals and objectives. The current draft strategic plan also states that the resources and programs listed in DOT's annual performance plans and budgets are necessary to achieve DOT's outcomes and to execute the strategies. In contrast, the 1997 draft strategic plan provided insufficient information to describe the operational processes, the skills, the technology, and the resources required to meet DOT's long-term goals, as required by GPRA. Also, each strategic objective section in the current draft plan contains crosswalks between outcomes in the strategic plan and performance measures in the annual performance plans and reports. These crosswalks show the measures that will be used to measure progress in achieving most of DOT's outcomes and strategic objectives. For example, the performance measure "number of passengers in international markets with open skies aviation agreements" is related in a crosswalk to the outcome "reduced barriers to trade in transportation goods and services." Together, the measure and outcome will show progress toward DOT's global connectivity strategic objective. This is an improvement from DOT's 1997 draft strategic plan when we noted that although supporting documents showed that DOT had developed information on how to measure each outcome goal, this information was not included in the draft. In contrast to DOT's 1997 draft strategic plan, the 2003-2008 draft plan lists several external factors for each strategic objective and generally discusses how these factors could affect the department's ability to achieve its outcomes and objectives. For example, one of the external factors for DOT's environmental stewardship strategic objective is that DOT faces a significant challenge to control and minimize air, water, and noise pollution. The plan states that if DOT cannot control and minimize this pollution, the department may encounter a public backlash that may impede system improvement. For the external factors relating to the safety and mobility strategic objectives, the plan lists both positive and negative consequences the factors could have on achieving goals. One example would be the possible effects the expansion and integration of the telecommunications and e-commerce industry sectors could have upon transportation safety. The plan states that this could affect the achievement of DOT's safety objective by leading to unsafe practices, such as the use of cell phones and other personal devices while driving. On the other hand, these technologies could also contribute to safety by alerting responders to the location of crashes and vehicles in distress. The 1997 draft plan identified four external factors and only discussed how one of those factors could affect DOT's ability to accomplish its goals. The current draft strategic plan includes an extensive table listing program evaluations to be completed during fiscal years 2003-2008. The table includes the name of the program to be evaluated, which strategic goal(s) the program supports, the scope and methodology of the evaluation, and the fiscal year during which the evaluation will be completed. DOT's 1997 draft strategic plan only listed the titles for the evaluations scheduled for 1997 and 1998, which was insufficient to determine the scope and methodology. DOT's current draft plan lists crosscutting programs by strategic objective. The discussions of crosscutting programs include the goal of each program, which of DOT's outcomes each supports, and the agencies involved. For example, the goal of aviation security is to prevent explosives, weapons, and other dangerous items from being placed aboard aircraft. This program supports DOT's security outcome of rapid recovery of transportation in all modes from intentional harm and natural disasters. DOT, through the Federal Aviation Administration, leads this program, which involves the Transportation Security Administration, Federal Bureau of Investigation, U.S. Customs Service, and U.S. Postal Service, among others. Previously, the 1997 draft did not provide evidence that DOT coordinated with other agencies that had programs and activities that were crosscutting or similar to DOT's. DOT's major management challenges, which we identified, are generally discussed in the 2003-2008 draft strategic plan, organized by related strategic objective. For example, we noted in our January 2003 report that one major management challenge that DOT faces is building human capital strategies.[Footnote 88] The current draft strategic plan includes a discussion of human capital in DOT's organizational excellence objective. A separate section of this discussion addresses our concerns regarding human capital strategies and includes several milestones to address these concerns. These milestones include conducting workforce planning for mission-critical occupations in fiscal year 2003, and implementing a departmentwide performance management system beginning in fiscal year 2003, and a uniform branding and marketing approach to attract, acquire, and retain diverse high- quality talent. DOT's 1997 draft strategic plan did not adequately address the major management challenges we had previously identified. Critical Strategic Planning Issues Needing Further Improvement: As stated above, the 2003-2008 plan provides a clear picture of how success will be measured for most of DOT's strategic objectives and outcomes. However, this clarity is not provided for a few strategic objectives and outcomes. We noted the same issue in our review of DOT's 1997 draft strategic plan. For example, in the current plan three of the outcome goals for the global connectivity strategic objective lack corresponding performance measures. These outcomes are enhanced international competitiveness of U.S. transport providers and manufacturers, harmonized and standardized regulatory and facilitation requirements, and the most competitive, cost-effective and efficient environments for passenger travel. The plan states that the measures are to be determined. However, without these measures it is unclear how progress will be measured because the outcomes themselves do not lend themselves to measurement. While the current strategic plan shows improvement in the schedule for future program evaluations, it does not sufficiently describe the evaluations used in establishing or revising DOT's strategic objectives. DOT states that detailed descriptions of completed program evaluations are presented in its 2002 performance and accountability report. Further, the plan states that DOT considered the results of completed program evaluations, as well as reports from DOT's Inspector General and GAO, in writing the strategies to achieve its strategic objectives and outcomes. The plan does not describe which or how program evaluations were used to write the strategies. [End of section] Appendix IV: Observations on Agencies' Annual Performance Plans: Under GPRA, agencies are to prepare annual performance plans after the development of their strategic plans. These annual plans are to establish the connections between the long-term strategic goals outlined in the strategic plans and the day-to-day activities of managers and staff. One of our objectives was to assess the changes in the overall quality of agencies' goals, strategies, and data articulated in their annual performance plans. To meet this objective, we judgmentally selected six agencies--Education, DOE, HUD, SBA, SSA, and DOT--using criteria such as agency size, primary program types, and previous GAO reviews. To assess the overall quality and improvements made to the agencies' performance plans, we relied on requirements contained in GPRA and accompanying committee report language,[Footnote 89] guidance to agencies from the OMB for developing performance plans,[Footnote 90] best practices identified in our published work,[Footnote 91] previous GAO evaluations,[Footnote 92] interviews with agency officials, and our knowledge of agencies' operations and programs. Key Elements of Information for Annual Performance Plans: Although GPRA does not require a specific format for the performance plan, it does require the plan to (1) identify annual performance goals and measures for each of an agency's program activities, (2) discuss the strategies and resources needed to achieve annual performance goals, and (3) provide an explanation of the procedures the agency will use to verify and validate its performance data. We categorized each agency's plan based on the degree to which it collectively addressed these three characterizations. To assess the degree to which an agency's plan provides a clear picture of intended performance across the agency, we examined whether it included (1) sets of performance goals and measures that address program results, (2) baseline and trend data for past performance, (3) performance goals or strategies to resolve mission-critical management problems, and (4) identification of crosscutting programs (i.e., those programs that contribute to the same or similar results), common or complementary performance goals and measures to show how differing program strategies are mutually reinforcing, and planned coordination strategies. To assess the degree to which an agency's plan provides a specific discussion of strategies and resources the agency will use to achieve performance goals, we examined whether it included (1) budgetary resources related to the achievement of performance goals, (2) strategies and programs linked to specific performance goals and descriptions of how the strategies and programs will contribute to the achievement of those goals, (3) a brief description or reference to a separate document of the human capital, information, and other resources required to achieve results,[Footnote 93] and (4) strategies to leverage or mitigate the effects of external factors on the accomplishment of performance goals. Finally, to assess the degree to which an agency provides confidence that its performance information will be credible, we examined how each report discussed the quality of the data presented. To help improve the quality of agencies' performance data, Congress included a requirement in the Reports Consolidation Act of 2000 that agencies assess the completeness and reliability of their performance data. Under the Act, agencies were to include this assessment in the transmittal letter with their fiscal year 2000 performance reports. Agencies were also required to discuss in their report any material inadequacies in the completeness and reliability of their performance data and discuss actions to address these inadequacies. For each of these elements, we characterized each agency's fiscal year 1999 and fiscal year 2004 plan in one of four ways, based on the degree to which the plan contained informative practices associated with that element. Thus, to address the first element concerning the degree to which the plan provided a clear picture of performance, we characterized each plan in one of four ways: (1) clear, (2) general, (3) limited, or (4) unclear. To address the second element, on the extent to which a plan includes specific discussions of strategies and resources, we characterized each plan as (1) containing specific discussions of strategies and resources, (2) general discussions, (3) limited discussions, or (4) no discussions. Finally, to address the third element on the extent to which a plan provides confidence that performance information will be credible, we characterized each plan as providing (1) full confidence, (2) general confidence, (3) limited confidence, or (4) no confidence. In conducting our reviews, we compared our assessments of agencies' fiscal year 2004 plans to our assessments of plans from fiscal year 1999 using similar criteria.[Footnote 94] A more detailed discussion of our scope and methodology and the criteria we used can be found in appendix I. Table 12 summarizes our characterizations of the six agencies' annual performance plans based on our current review of fiscal year 2004 plans and our previously published reviews of 1999 plans. Although the characterization of agency performance plans did not change significantly between the 1999 and the 2004 plans, the majority of agencies' plans showed some improvement. Table 12: Characterizations of Agencies' Annual Performance Plans: Agency: Department of Education; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: General; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: Limited; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: General; Characterizations: Data credible: (no, limited, general, full): 1999: Limited; Characterizations: Data credible: (no, limited, general, full): 2004: General. Agency: Department of Energy; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: Limited; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: General; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: General; Characterizations: Data credible: (no, limited, general, full): 1999: Limited; Characterizations: Data credible: (no, limited, general, full): 2004: Limited. Agency: Department of Housing and Urban Development; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: General; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: Limited; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: General; Characterizations: Data credible: (no, limited, general, full): 1999: Limited; Characterizations: Data credible: (no, limited, general, full): 2004: General. Agency: Small Business Administration; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: General; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: Limited; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: General; Characterizations: Data credible: (no, limited, general, full): 1999: Limited; Characterizations: Data credible: (no, limited, general, full): 2004: General. Agency: Social Security Administration; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: Limited; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: Clear; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: Limited; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: General; Characterizations: Data credible: (no, limited, general, full): 1999: No; Characterizations: Data credible: (no, limited, general, full): 2004: General. Agency: Department of Transportation; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 1999: General; Characterizations: Picture of intended performance: (unclear, limited, general, clear): 2004: Clear; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 1999: General; Characterizations: Strategies and resources: (no discussions, limited, general, specific): 2004: Specific; Characterizations: Data credible: (no, limited, general, full): 1999: Limited; Characterizations: Data credible: (no, limited, general, full): 2004: Full. Sources: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-172R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-194R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-159R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-180R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-200R]; [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-178R]; Department of Education, FY 2004 Annual Performance Plan; (Washington, D.C. 2003); Department of Energy, Annual Performance Plan, Fiscal Year 2004; (Washington, D.C. 2003); Housing and Urban Development, Annual Performance Plan, Fiscal Year 2004; (Washington, D.C. 2003); Small Business Administration, Budget Request & Performance Plan: FY 2004 Congressional Submission; (Washington, D.C. 2003); Social Security Administration, Annual Performance Plan, Fiscal Year 2004; (Washington, D.C. 2003); Department of Transportation, Fiscal Year 2004 Performance Plan; (Washington, D.C. 2003). [End of table] The remainder of this appendix discusses our observations on how the quality of each of the agencies' annual performance plans we reviewed has changed since the agencies submitted their first performance plans in 1999. We did not independently verify or assess the information we obtained from agency annual performance plans. If an agency chose not to discuss its efforts concerning elements in the plan, it does not necessarily mean that the agency is not implementing those elements. Observations on Changes in the Quality of Education's Annual Performance Plan: Education's fiscal year 2004 annual plan[Footnote 95] provides a general picture of intended performance across the agency--an improvement over the 1999 plan--because the measures and indicators adequately indicate progress toward meeting annual targets; the measures are objective, measurable, and quantifiable; and baseline and trend data are included where available. However, the relationship between the goals and measures in volume 2 and the long-term goals in volume 1 is not clear and the plan does not make clear whether, and, if so, how, all of the program activities in the department's budget are covered by the annual performance goals.[Footnote 96] In another improvement over the 1999 plan, volume 1 of the 2004 plan provides a general discussion of Education's strategies and resources to achieve its goals by presenting strategies and resources and the projected distribution of fiscal year 2004 funding and staffing for each long- term goal. The plan also adequately recognizes and discusses external factors that could affect the department's performance. However, the resources and many of the strategies are not directly linked to the achievement of individual annual performance goals and no strategies or resources are designated for the goals and measures in the program performance plans in volume 2. Lastly, the 2004 plan provides general confidence that agency performance information will be credible. The 2004 plan contains information on data sources for most of its measures, and for some, identifies limitations. The plan also includes an appendix entitled "Information Quality Guidelines" which recognizes data quality as a major challenge for the agency and says that its improvement is a top priority. Education's Fiscal Year 2004 Performance Plan Provides a General Picture of Intended Performance: Education's 2004 annual plan generally defines expected performance. Measures and indicators are formulated so as to adequately indicate progress towards meeting annual targets, seem to sufficiently cover key performance aspects, and adequately capture important program distinctions. The plan contains about 360 measures between volumes 1 and 2, an improvement over the 860 contained in its 1999 plan, which we judged to be potentially excessive for an annual performance plan and possibly interfering with Education's ability to assess its performance. Unlike in our review of the department's 1999 annual plan, the measures and indicators in the 2004 plan are objective, measurable, and quantifiable. For example, most measures and indicators are set up to measure percentages, cost, counts, or other numerical values with measurable, quantifiable 2004 targets. In most cases where a measurable target is not given, the plan provides a reasonable explanation, such as a new program or the measure being new, or a case where data are not collected or available each year. In most cases, the plan provides trend data for measures, which provides a helpful context for assessing the relevance of the 2004 targets, or an explanation of why such data were not provided (e.g., the baseline has not yet been established because the measure and/or program are new). In our review of Education's 1999 annual plan, we said that greater outcome measure use would make future annual plans more useful. The 2004 plan frequently includes outcome goals and measures, such as a measure for the number of states meeting their eighth-grade mathematics achievement targets under the long-term goal to improve mathematics and science achievement for all students. Volume 1 of Education's 2004 annual plan directly aligns strategies, action steps, measures, and targets with each of Education's long-term goals and six strategic goals,[Footnote 97] containing the same strategic goals, long-term goals, and mission as the 2002-2007 strategic plan. In our review of the 1999 plan, we also found that the plan had performance goals in volume 1 that were directly linked to its mission, strategic goals, and objectives. However, the relationship between the goals and measures in volume 2 and the long-term goals in volume 1 was not made clear in the department's 2004 annual plan, which was similar to what we found in our review of Education's 1999 plan-- that the department could more directly link the goals and measures in volume 2 with the strategic objectives (long-term goals). Education's fiscal year 2002 Performance and Accountability Report includes a table making it clear that these programs and their goals and measures are aligned across Education's strategic and long-term goals. By including such a table in its annual performance plan, Education could clearly show the link between the goals and measures in volume 2 and the long- term goals in volume 1. Although volume 2 of the 2004 annual plan states that it contains individual program performance plans for all major programs and many smaller programs, the annual plan does not make clear whether, and, if so, how all of the program activities in the department's budget are covered by performance goals. In contrast, we found that the 1999 plan provided sufficient information to determine which performance goals and measures in volume 2 covered which program activities in the budget and whether all were covered. For example, the 1999 plan contained tables indicating the funding levels for the program activities in the department's budget and how those activities related to the programs in volume 2. In our review of Education's 1999 annual plan, we gave the agency credit for addressing the need to coordinate with other federal agencies having related strategic goals or performance goals. However, we further noted that Education could build on its foundation by identifying performance goals that reflect activities being undertaken to support programs of a crosscutting nature and specifying the activities each agency would undertake and what it expects to achieve within the fiscal year. While selected action steps in the 2004 plan refer to instances where the department will coordinate or cooperate with other federal agencies, the plan does not include steps or goals for most crosscutting issues identified in Education's strategic plan. For example, for a crosscutting issue identified in the strategic plan on safe and drug-free schools and communities, Education said it partners with the Departments of Justice (Justice) and HHS to promote drug and alcohol education programs and to disseminate information to schools and private organizations. The department also coordinates closely with the Office of National Drug Control Policy, and works closely with the Office of Juvenile Justice and Delinquency Prevention Programs to share innovative ideas and promote prevention strategies and programs. The relevant annual plan sections in both volumes 1 and 2 do not identify goals or action steps related to these interactions. Additionally, according to our report on Education's June 1997 draft strategic plan,[Footnote 98] the department has numerous crosscutting programs and activities, such as those related to early childhood and employment training, and the 2004 annual plan does not address them all. Education's 2004 annual plan discusses applicable goals, measures, and strategies for two governmentwide major management challenges regarding strategic human capital management and information security, as well as three of the four major management challenges we identified for Education in our January 2001 Performance and Accountability Series.[Footnote 99] For example, for its student financial assistance programs, the department has developed performance targets for fiscal year 2004 related to being removed from our high-risk list, increasing the default recovery rate, and decreasing grant overpayments to students. Also, a key strategy under its goal to improve the strategic management of the department's human capital is to develop a 5-year human capital plan, including developing a recruitment plan and relevant training programs. For the fourth of Education's major management challenges--promoting coordination with other federal agencies and school districts to help build a solid foundation of learning for all children--the 2004 plan did not include specific goals or measures, but it did discuss some related strategies and steps, such as using partnerships with other federal programs to promote development of intervention strategies and methods to address the high incidence of learning disabilities and illiteracy among adolescents attending high schools. Education's Fiscal Year 2004 Performance Plan Provides a General Discussion of Strategies and Resources: In our review of Education's 1999 annual plan, we found that the plan had a limited discussion of how the department's strategies and resources would help achieve its annual performance goals. The 2004 plan includes strategies and resources under each of its long-term goals to be used to achieve its annual performance goals in volume 1, including the projected distribution of fiscal year 2004 funding and staffing, in both dollars and full-time-employees (FTE), for each long- term goal, under which the annual performance goals are organized. However, the resources and many of the strategies are not directly linked to the achievement of individual annual performance goals and no strategies or resources are designated for the goals and measures in the program performance plans in volume 2. Overall, the plan does not discuss how resources were allocated to each goal, a rationale for how the resources will contribute to improving performance, or the relationship of capital asset investments, including those for information technology (IT), to the achievement of specific goals. However, the plan does include a performance measure and goal for the cost and schedule of IT investments and a strategy for completing the department's enterprise architecture, which is to be used to guide IT capital decisions. In addition, the department has a plan for human capital management and a new performance appraisal system that is meant to link employee performance standards to the department's strategic priorities, but neither had been fully implemented. In our review of the 1999 plan, we said that external factors that could affect performance were not discussed and that such factors are important for a department like Education because much of what it hopes to achieve depends on others and external events. In its 2004 plan, Education clearly acknowledges that improving support for its state, local, and institutional partners, who have the direct ability to influence outcomes the department seeks, is a major challenge. The plan contains numerous activities to handle this challenge, including, for example, to provide support and technical assistance, improve grant monitoring, and fund an annual survey of states' efforts. Moreover, although not labeled as addressing external factors, the plan has specifically related strategies and/or action steps for most external factors identified in Education's 2002-2007 strategic plan. Education's Fiscal Year 2004 Performance Plan Provides General Confidence That Performance Data Will Be Credible: In our review of Education's 1999 annual plan, we found that it did not provide sufficient confidence that its performance information would be credible. For example, the 1999 plan did not sufficiently recognize limitations in Education's data for its elementary and secondary education programs. In comparison, Education's 2004 plan recognizes limitations in Education's data for many of its elementary and secondary education programs, as well as for other programs. In many of these cases, the plan also discusses plans to address these limitations. Also, the plan includes an appendix containing an abbreviated form of its "Information Quality Guidelines" and a sample checklist for statistical data from its complete guidelines. The appendix recognizes data quality as a major challenge to the department's successful implementation of GPRA and says that the improvement of data quality is a top priority. The checklist includes several steps related to the verification and validation of data, such as evaluating data quality, including known limitations; addressing the reliability of data sources; and ensuring reproducibility of findings using the same data and methods of analysis. In addition, the plan usually identifies the sources of data and, although not in volume 1, includes a column on sources and data quality for each measure in volume 2, usually with an item entitled "Validated By" and, in some cases, "Limitations." In the end, the lack of direct control over the implementation of its programs, including the collection of data, is a significant data quality challenge that Education must face. The 2004 annual plan also contains several action steps on new or changing information systems that relate to improving the collection of information for measuring performance. For example, under its strategy to reduce Education's partners' data reporting burden, the plan includes an action step to develop and implement the Performance-Based Data Management Initiative collection system. This step is directly related to the plan's measure to reduce the burden hours of Education program data collections per year. Observations on Changes in the Quality of DOE's Annual Performance Plan: Compared to the fiscal year 1999 plan we reviewed, DOE's fiscal year 2004 performance plan continued to provide a limited picture of intended performance. Although the plan included more results-oriented annual performance measures, it still provided a limited linkage between its reported annual goals and its mission, strategic plan goals, and program activities within its budget request. Furthermore, the 2004 plan provided a general discussion of strategies and resources, similar to our 1999 findings. Finally, the 2004 plan provided a limited level of confidence that data will be credible by making little progress in reporting on the procedures it uses to ensure data quality or identifying significant data limitations, which is consistent with our 1999 findings. DOE's Fiscal Year 2004 Performance Plan Provides a Limited Picture of Intended Performance: While DOE has improved its development of annual performance measures- -referred to as targets--by making them more results oriented, the overall picture of performance is limited by the lack of alignment between its annual and strategic goals and minimal discussion of coordination with other agencies. Our review of DOE's 1999 plan found that many measures were unclear, appeared limited in scope, or were not very useful indicators of performance. We found these problems in the performance plans for subsequent years as well. For the 2004 plan, however, the majority of the performance measures related to each goal were results oriented and pertained specifically to the performance of fiscal year 2004. An example of one measure requires DOE to train 4,000 federal employees by the end of fiscal year 2004 in energy management best practices that support National Energy Policy education goals. DOE provided a limited link between its reported annual goals and its mission, strategic plan goals, and program activities within its budget request. While the 2004 annual performance plan goals address all of the major program activities in DOE's budget, the goals and mission of the 2004 plan do not align with the mission and goals for the 2003 draft strategic plan. This represents a set back because in our review of DOE's 1999 annual performance plan, we found that DOE clearly linked its annual goals to the agency's mission, strategic plan goals, and its program activities within its budget request. DOE officials told us the lack of linkage between the performance plan and the strategic plan was a matter of timing. According to these officials, the department originally updated its strategic plan at the same time as the annual performance plan, which was finalized in the early months of 2003, and the goals of each plan coincided, but the draft strategic plan goals were revised in the latter part of the year and no longer align with the 2004 performance plan. DOE's ability to show coordination with other agencies is also limited. In 1999, we reported that DOE did not adequately show that it coordinated with other agencies that have related strategic or performance goals. DOE's 1999 plan contained very little evidence of specific goals and measures that addressed crosscutting programs and only briefly described coordination with other agencies. The 2004 plan does not specifically describe how coordination is taking place among crosscutting programs, but does identify groups that it is collaborating with on certain programs. In response, DOE officials told us the plan does not discuss what specific collaboration activities are taking place because it would require reporting too much detail for a performance plan. DOE officials stated that it collaborates at the program level, rather than the agency level, because the program plans pertain to an organizational layer lower than the annual performance plan. Finally, the plan briefly mentions that the department has been identifying challenges and working on ways to address them. According to DOE officials, when developing annual targets for the department, management challenges are considered but not mentioned specifically in the report. Our review of management challenges in 2002 found that DOE had addressed all eight of its challenges in its 2003 annual performance plan.[Footnote 100] In comparing these challenges to the 2004 plan, we found that DOE continues to have goals that address the eight challenges we identified in 2002 and the additional challenges that we identified in our 2003 performance and accountability report.[Footnote 101] DOE's Fiscal Year 2004 Performance Plan Provides a General Discussion of Strategies and Resources: DOE provided a general discussion of the strategies and resources that it will use to achieve its annual performance goals. DOE's 1999 plan partially provided clear and reasonable strategies for achieving performance goals, how strategies would contribute to achieving the performance goals, and key external factors that might affect performance. For each of the 2004 annual performance goals, DOE included a "Means and Strategies" section in the plan that described how each goal will be achieved. For example, one strategy identified to meet its goal of contributing unique, vital facilities to the biological and environmental sciences is to conduct peer reviews of the facilities to assess the scientific output, user satisfaction, and the overall cost-effectiveness of each facility's operations, and their ability to deliver the most advanced scientific capability. The 2004 plan includes a brief discussion of the department's overall needs, particularly in the areas of human capital, financial, and logistical resources. The plan also identified budget amounts for each of its goals. DOE's 1999 plan partially identified the resources needed to accomplish annual performance goals. The plan also provided a general discussion of the external factors that could affect achievement of the goals, but it did not specifically discuss actions on how the external factors will be addressed. For example, the plan states that external factors related to DOE's goal of achieving reliable, affordable, and environmentally sound energy supplies, such as renewable fuels, include program funding, the state of the economy, the availability of conventional supplies, the cost of competing technologies, and the continuation of federal tax incentives and other national-level policies. DOE's Fiscal Year 2004 Performance Plan Provides Limited Confidence That Performance Data Will Be Credible: DOE has made limited progress on reporting the procedures it uses to ensure data quality. Its 1999 plan described how DOE would ensure that its performance information is sufficiently complete, accurate, and consistent, but did not discuss in detail DOE procedures on how to help ensure the quality of data or the process of collecting the data. The plan also did not identify significant data limitations and how they may affect DOE's ability to achieve performance goals. However, the 2004 plan showed some improvement over the 1999 plan by describing credible procedures to verify and validate performance information and specific program evaluations are mentioned for each goal. The plan also discusses that DOE acquired new commercial software for performance tracking through remote data entry, monitoring, and oversight by program offices and managers. The 2004 plan only identifies data limitations and any new or modified systems very briefly for a few relevant goals. According to DOE officials, with a few exceptions, its plans do not discuss data limitations because DOE writes goals that are not affected by data limitations. The goals are written to ensure that the data will be there to meet performance targets. However, our 2003 performance and accountability series identified several DOE management challenges where data quality was a concern, such as further upgrades needed for cyber security to ensure adequate protection of data and information systems and additional information on the results of contractors' performance to keep projects on schedule and within budget. Observations on Changes in the Quality of HUD's Annual Performance Plan: HUD's annual performance plan for fiscal year 2004 improves upon areas where we previously reported shortcomings and generally meets the criteria set forth in GPRA. HUD's 2004 plan provides a general picture of intended performance by covering all the programs contained in HUD's budget and linking program activities to strategic goals and objectives. The plan also improved by providing specific information on HUD's strategies and activities along with performance measures it will use to assess progress toward its goals and discussing relevant external factors that could affect the attainment of certain program objectives. HUD also provides greater confidence that performance data will be credible by thoroughly discussing the data it will use for measuring progress toward its goals. Nevertheless, the plan could be further enhanced if it included more specific information on how funds will be allocated to achieve program objectives, explain how HUD will contribute to crosscutting efforts along with other agencies, and what steps it will take to mitigate the impact of external factors on its programmatic objectives. HUD's Fiscal Year 2004 Performance Plan Provides a General Picture of Intended Performance: Since our review of HUD's annual performance plan for fiscal year 1999,[Footnote 102] HUD has made progress in developing an annual performance plan that generally reflects the department's mission and provides a general picture of intended performance. HUD's most recent performance plan covers the program activities contained in its budget, and generally links program activities to strategic goals and objectives, key items missing from its plan for fiscal year 1999. HUD has also improved the quality of its performance plan by including performance measures that generally indicate how the department will gauge progress toward achieving its goals. For example, the performance plan lists a series of performance measures for each objective that can be used to indicate progress towards the department's goals and expected performance. These measures are also objective and a number of them have been quantified, another key area where HUD has improved since its first performance plan. For example, activities supporting HUD's long-term strategic objective to "Improve the Physical Quality and Management Accountability of Public and Assisted Housing" include, among other things, eliminating 100,000 units of the worst public housing. According to the current plan, the department intends to demolish 10,000 of these units in fiscal year 2004. While HUD's most recent annual performance plan generally identifies other agencies it will coordinate with to address crosscutting efforts, it does not discuss how it plans to work with these agencies to address these crosscutting activities. For example, the plan states that the Interagency Working Group on Limited English Proficiency will ensure that persons with limited English proficiency will have meaningful access to funded and federally conducted programs and activities. However, the plan does not discuss what HUD's contribution to this multiagency effort will be, what strategies it will employ, or how it will measure progress toward achieving the strategies of this multiagency effort. HUD's current performance plan is an improvement compared to its fiscal year 1999 plan as it describes steps HUD will take to address major management challenges. One of HUD's strategic goals, "Embrace High Standards of Ethics, Management, and Accountability," identifies five objectives that cover management challenges, some of which have been raised by GAO and the HUD IG. These objectives discuss plans to rebuild HUD's human capital and diversify its workforce; improve HUD's management, internal controls, and resolve audit issues; improve accountability, service delivery, and customer service; ensure program compliance; and improve internal communication and employee involvement. HUD's Fiscal Year 2004 Performance Plan Provides a General Discussion of Strategies and Resources: HUD's most recent performance plan also improves upon earlier plans we reviewed in providing readers an idea of the strategies that HUD will employ to carry out its goals. Each strategic goal in the annual performance plan contains a section titled "Means and Strategies," which describes activities HUD will pursue to support that goal. For example, to support its goal of "Increasing Homeownership Opportunities," HUD will fund low-income homeowner assistance programs to provide approximately 40,000 families with down payments and closing costs on their homes, 473,199 families with home purchase and homeownership counseling, and about 232,370 families with rental counseling. HUD's performance plan also discusses the strategies it will employ to address the department's human capital issues, such as the upcoming potential wave of employees planning to retire and the need to equip staff with the desired knowledge and skills. For example, HUD completed a staff resource estimation and allocation system in 2002, and it will conduct a comprehensive workforce analysis in 2004 to serve as the basis to fill mission-critical skill gaps through succession planning, hiring, and training initiatives in its Five-Year Human Capital Management Strategy. Although HUD has made progress in linking its resources to strategies, it could improve the discussion by linking funding allocations to specific performance goals, thus making the plan more informative. The plan discusses budget and staff allocations for programs supporting each strategic goal. For instance, portions of HUD's Community Development Block Grants Fund and Home Investment Partnership Program, with a combined budget authority for fiscal year 2004 of more than $2.5 billion and staff of 203, support the strategic goal of promoting "Decent Affordable Housing." However, it is unclear what resources will be used to pursue specific performance targets for each program. Additionally, HUD does not mention in its plan how IT and capital resources will be used to support its programs. Anticipating that some aspects of the department's strategic goals are intertwined with broader phenomena, the performance plan also discusses several external factors relevant to each strategic goal that could affect HUD's ability to meet its objectives. For example, for its strategic goal "Promote Decent and Affordable Housing," HUD states that broad economic factors can affect opportunities for low-income workers relying on the department for rent assistance to make progress towards self-sufficiency. However, it is unclear from the performance plan what actions, if any, HUD has put in place to mitigate the effect of these external factors. HUD's Fiscal Year 2004 Performance Plan Provides General Confidence That Performance Data Will Be Credible: HUD has also made significant progress in providing assurance that the department will be able to use credible data to gauge progress towards achieving its goals. HUD identifies the steps it (or others) will take to verify and validate the performance data to ensure that what is reported on HUD's performance will be credible. For example, for its objective "Increasing Minority Homeownership," HUD will rely on, among other indicators, the rate of minority homeownership from the Current Population Survey conducted monthly by the U.S. Census Bureau. HUD will not verify the data because the Bureau performs that task. Additionally, HUD also includes in its most recent performance plan a discussion of the inherent limitations of the data it will use and generally discusses steps it will take to improve the measure, providing the reader with a clearer expectation of what HUD will be able to report. Observations on Changes in the Quality of SBA's Annual Performance Plan: SBA's 2004 performance plan shows progress made over the agency's 1999 performance plan. In contrast to our review of SBA's 1999 plan,[Footnote 103] the 2004 plan provides a general picture of intended performance by discussing coordination between SBA and other federal agencies on crosscutting activities. Resource analysis sections throughout the plan provide a general discussion of how SBA has previously used its resources to achieve its goals and how it intends to use future resources for the same purposes. The 2004 plan also provides general confidence that SBA's performance data will be credible by including more detail on how SBA verifies and validates its data, as well as by identifying data limitations. However, several areas of the plan could be improved, such as clearly linking SBA's performance indicators, performance goals, and programs. SBA's Fiscal Year 2004 Performance Plan Provides a General Picture of Intended Performance: SBA's fiscal year 2004 performance plan provides a general picture of intended performance. The performance goals and performance indicators in the plan are generally objective, measurable, and quantified. Performance indicators are listed throughout the plan by related programs and strategic programmatic goals. Performance goals and outcome goals are listed for each strategic programmatic goal. In our review of SBA's fiscal year 1999 performance plan, we noted that SBA's performance goals were objective and measurable, its performance measures were generally objective and quantified, and that the performance goals in the plan were clearly linked to SBA's strategic goals and objectives. Like the performance measures contained in its 1999 plan, the 2004 plan's performance indicators will be useful in assessing progress towards SBA's performance goals. For example, the performance indicator "Regulatory Cost Savings to Small Business" will adequately show progress for the corresponding performance goal "Cost savings for small business due to the efforts of the Office of Advocacy." In this example, the performance indicator, which is listed under the Advocacy Program, can be linked to a performance goal because of a crosswalk that relates outcome goals, performance goals, and programs. However, there is not always such a clear link between all of SBA's performance indicators and performance goals because indicators are listed by program instead of by performance goal. The BusinessLaw.gov program is linked to three performance goals: "number of users of BusinessLaw.gov," "reduced cost to businesses and regulatory agencies," and "increased rate of compliance." While the first two performance goals appear related to the first two indicators listed in the BusinessLaw.gov program section, there is no clear relationship between any of the other performance indicators for this program and the third performance goal, "increased rate of compliance.": SBA's 2004 performance plan contains annual performance goals that generally cover the agency's budget activities. The 2004 performance plan contains a budget crosswalk that "shows how the goals relate to specific and general program areas." This is an improvement over the 1999 plan, which we noted contained a budget crosswalk, but the categories in it did not match SBA's budget accounts or activities by name or account number. However, the performance plan does not seem to cover all of SBA's programs. Three "advocacy" programs listed in the crosswalk do not seem to be contained in the plan: Business.gov, Disability Initiative, and National Women's Business Council. Each strategic programmatic goal section in the 2004 plan contains a discussion of crosscutting issues. Several examples of coordination efforts are given, such as SBA working with the Department of Defense to integrate the PRO-Net system with the Central Contractor Registry and SBA partnering with the Federal Acquisition Institute to develop on-line training courses for small business programs. In contrast, SBA's 1999 performance plan provided little information on SBA's coordination efforts with other entities whose programs and activities crosscut those of SBA. SBA's 2004 performance plan generally addresses performance and accountability challenges we have previously identified. For example, we have previously stated that SBA needs to strengthen its performance in human capital management. The 2004 plan includes outcome goals, performance goals, and programs to address SBA's strategic management of human capital in a section on the PMA. SBA's Fiscal Year 2004 Performance Plan Provides a General Discussion of Strategies and Resources: The 2004 performance plan provides a general discussion of the strategies and resources SBA will use to achieve its goals. Each strategic programmatic goal and each of the goals for the PMA contains a discussion of the strategies for accomplishing the goals. These discussions provide a broad overview of the strategies used at the strategic programmatic goal level. For example, the plan includes a strategy for SBA's strategic management of human capital. The strategy lays out SBA's Transformation and Human Capital plans, which will be used to implement a new vision of SBA. In its 1999 performance plan, SBA discussed strategies for most of its performance goals, although for some of the goals the strategies were missing. Each strategic programmatic goal and several of the PMA goals contain brief discussions of external factors that could affect the achievement of SBA's goals. These discussions include actions to address external factors, such as working with an Interagency Acquisition Working Group under the Procurement Executives Council to develop supplemental performance measures to better evaluate the success of its programs. In 1998, we noted that SBA's 1999 plan recognized certain external factors and contained a discussion of actions SBA could take to mitigate the effects of such factors for one of its strategic goals. We stated that it would be useful for SBA to include a similar discussion of external factors and mitigation strategies for its other strategic goals. Each program listed throughout the plan has a resource analysis section that describes how resources were used in fiscal year 2002. Some of these analyses also include planned resources for fiscal year 2004. For example, the resource analysis section for the Small Business Development Centers (SBDC) program states that for fiscal year 2004 SBA requested approximately the same level of funding as in fiscal year 2002. In 2002, 85 percent of the funds were for grants, while the other 15 percent covered field support, program management, and overhead costs such as rent, legal services, human resources, and information technology support. Some of the resource analyses also contained pie charts of the breakdown of costs. This is an improvement over SBA's 1999 performance plan, which we found did not specifically identify the human or technological resources that SBA would need to achieve its performance goals. SBA's Fiscal Year 2004 Performance Plan Provides General Confidence That Performance Data Will Be Credible: SBA's 2004 performance plan provides general confidence that its performance data will be credible. An appendix of the performance plan contains verification and validation information, as well as data limitations and remedies for these limitations for most of SBA's performance indicators. However, the appendix does not include this information for the performance indicators of the disaster loan program, nor are any of the PMA performance indicators discussed in the appendix. Generally, the discussions of SBA's verification and validation processes for its indicators in the 2004 plan are one-or two-sentence statements. For one of the indicators, "number of jobs created and retained by the 7(a) loan program," SBA states that it does not have access to the data for verification purposes. SBA also notes that it does not independently verify some of the external data it gathers, as is stated in the verification discussion of the indicator, "504 loans to emerging market firms." This is an improvement over SBA's 1999 performance plan, which included brief descriptions, often only one or two words, on the means it used to verify and validate its data. We noted in our report on the 1999 plan that these appeared to be sources of data for the measures rather than means to verify and validate the data. The data limitations contained in SBA's 2004 performance plan are generally one-sentence statements and the same limitations are used for multiple indicators. For example, several limitations, such as "the measure is based on the number of approved loans" or "information is derived from loan approval data," are used for multiple indicators. The appendix also lists remedies for the data limitations for each indicator. For example, a limitation of SBA's indicator "small business appointments conducted with procurement officials" is that the indicator may not capture unscheduled appointments. The remedy for this limitation is to keep track of both scheduled and unscheduled appointments. The discussion of data limitations and their remedies in the 2004 plan shows progress over SBA's 1999 plan, which did not contain a discussion of data limitations. Observations on Changes in the Quality of SSA's Annual Performance Plan: Compared to the fiscal year 1999 plan we reviewed, SSA's performance plan provided a clear picture of intended performance by (1) defining expected performance, (2) offering trend data, which helps track progress toward performance goals, and (3) using objective, measurable, and quantifiable performance measures. SSA also provided general information on its strategies and resources, somewhat better than our 1999 findings. Finally, the plan provided a general level of confidence that data will be credible by describing the Inspector General's (IG) involvement in data testing, providing data sources and definitions, and identifying some data weaknesses, an improvement over the 1999 plan. However, SSA's performance plan still does not fully discuss the agency's coordination with other agencies, identify performance goals that clearly cover all the program activities, address how SSA plans to use the information from the evaluations to improve program results, identify the resources needed to address each performance goal, and discuss data verification and validation procedures for its internal systems. SSA's Fiscal Year 2004 Performance Plan Provides a Clear Picture of Intended Performance: Overall, SSA's fiscal year 2004 performance plan has improved over its 1999 plan.[Footnote 104] The 1999 plan only provided a partial picture of SSA's intended performance across the agency. In June 1998, we reported that SSA's 1999 Annual Performance Plan contained performance goals, many of which were measurable and linked to the agency's strategic goals[Footnote 105] and objectives; some of the performance goals related to particular strategic goals were objective, measurable, and quantifiable. However, other goals were not measurable or quantifiable and did not define the level of performance to be achieved, thus making it difficult to see how SSA would assess success. SSA's fiscal year 2004 plan provides a much clearer picture of intended performance through (1) defining expected performance, (2) offering trend data, which helps track progress toward performance goals, and (3) the use of objective, measurable, and quantifiable performance measures. For example, as part of the strategic objective to "Prevent fraudulent and erroneous payments and improve debt management," SSA provided historical data on the outcome measure "Percent of SSI payments free of preventable error (overpayments and underpayments)" from fiscal years 1999-2001 and projected goals for fiscal years 2002- 2004. While we found significant improvements in SSA's 2004 annual performance plan over its 1999 plan, we also found some weaknesses. For example, coordination efforts with other entities, such as federal agencies, state and local entities, and others, are not well identified. According to SSA officials, SSA coordinates with other federal agencies, such as the IRS and the Immigration and Naturalization Service, as well as Veterans Administration, on information-sharing initiatives. However, these types of coordination efforts are mentioned only briefly, if at all, in the 2004 annual performance plan. In its 2004 plan, SSA includes a list of major program evaluations it plans to conduct during 2003-2004, with a brief description of the evaluations, their associated strategic goals, and projected completion dates. However, there is no indication how SSA plans to use the information from the evaluations to improve program results. The plan could be enhanced if the descriptions of these evaluations included the manner in which SSA planned to use the information gathered to improve its programs. SSA's Fiscal Year 2004 Performance Plan Provides a General Discussion of Strategies and Resources: SSA's 1999 plan had little discussion of the relationship between SSA's mission, goals, and budget activities. Throughout the document, the fiscal year 2004 plan included clearer discussions of the linkage between SSA's mission and goals. It also provided performance data dating back to fiscal year 1999, essential to making comparisons between prior and proposed levels of performance. The 2004 performance plan noted that the Limitation on Administrative Expenses account, SSA's basic administrative account, is an annual appropriation that covers everything from salaries and benefits of SSA federal employees (excluding IG) to systems and telecommunications activities. SSA provided information on the funding sources of this account, including some of its budget accounts. In its fiscal year 2004 plan, SSA provided information on the strategies it plans to use in addressing its key strategic objectives. The plan included a summary chart, showing the strategic objectives associated with each strategic goal, as well as the performance measures under each objective. In addition, the "means and strategies" section associated with each strategic objective identified strategies that support items in the PMA, GAO and IG major management challenges, and Social Security Advisory Board recommendations. In our October 2002 report Performance and Accountability: Reported Agency Actions and Plans to Address 2001 Management Challenges and Program Risks,[Footnote 106] we noted that SSA identified directly related goals and measures for five of its six challenges, and had strategies (without goals or measures) for the sixth challenge. It is difficult to determine whether or not the annual performance plan identifies annual performance goals that cover all of the program activities in the agency's budget, as well as the financial, human capital, and information technology resources needed to address each individual goal. General human capital requirements and goals are identified as part of SSA's strategic goal to strategically manage and align staff to support SSA's mission. The plan is neither structured by program activity nor account. SSA noted that it aligned its strategic goals, performance measures, and budget with its major functional responsibilities rather than by program accounts since direct service and support employees provide services linked to these functional responsibilities, as opposed to a specific program. However, SSA does not indicate what it means by "functional responsibilities," nor does it show a clear link between its strategic goals and such responsibilities. As in the fiscal year 1999 plan, the fiscal year 2004 plan included a discussion of external factors that could affect the achievement of its goals.[Footnote 107] SSA identified strategies to alleviate some, but not all of, the factors. For example, SSA plans to mitigate the loss of institutional knowledge through SSA's "retirement wave" through the use of employee development programs, redeploying positions to direct service, hiring Presidential Management Interns, and the increased use of hiring flexibilities. However, the discussion of factors affecting SSA's solvency strategic goal merely notes that Social Security programs must respond to related developments. SSA's Fiscal Year 2004 Performance Plan Provides General Confidence That Performance Data Will Be Credible: SSA's 1999 plan stated that the Office of the Inspector General was responsible for reviewing the data systems underlying its performance measures, but did not provide further details that would assure the reader that SSA is taking the steps necessary to ensure data integrity. In contrast, SSA's fiscal year 2004 plan provided data sources and definitions for each performance measure. SSA's fiscal year 2004 plan identified data limitations related to performance measures, as well as some efforts to correct or address data weaknesses. When performance indicators and goals are not quantified, SSA describes its benchmarks for goal achievement. For example, for the outcome measure "Developing new performance management systems," SSA defines "Implementing the new SES system" as its goal for 2003. As in the fiscal year 1999 plan, SSA notes that the IG's office is involved in the data system reliability process. In the fiscal year 2004 plan, SSA went further to explain the IG's four-point approach to reviewing performance measures, including assessing whether the reported performance measure data are valid. SSA also noted that performance data for its quantifiable measures are generated by automated management information and workload measurement systems, as a by-product of routine operations. However, there is no discussion of verification and validation procedures for data generated by these systems. Observations on Changes in the Quality of DOT's Annual Performance Plan: DOT's annual performance plan for fiscal year 2004 showed evidence of improvements in areas that we previously identified had shortcomings in our 1998 review of DOT's 1999 performance plan.[Footnote 108] The 2004 plan provides a clear picture of intended performance with DOT's measures and performance goals now being clearly linked to the department's strategic objectives. A specific discussion of DOT's strategies and resources in the plan includes numerous and detailed strategies for achieving DOT's performance goals, and the resources needed for those strategies. Procedures to verify and validate data, as well as known data limitations, are described for each performance measure providing full confidence in the credibility of DOT's performance data. Still, the performance plan could be improved by including a discussion of, and performance measures for, each of DOT's program activities and by more consistently describing DOT's role in crosscutting programs. DOT's Fiscal Year 2004 Performance Plan Provides a Clear Picture of Intended Performance: DOT's fiscal year 2004 performance plan shows evidence of many of the same strengths as, and a few improvements over, its fiscal year 1999 performance plan and provides a clear picture of intended performance. The 2004 plan lists outcome goals, performance goals, and measures by strategic objective, all of which are generally objective, quantifiable, and can show progress toward DOT's strategic objectives. For example, the measure "fatalities per 100 million vehicle-miles of travel" will gauge progress toward the performance goal "reduce highway fatalities per 100 million vehicle-miles traveled to no more than 1.0 in 2008, from 1.7 in 1996." The data gathered by the measure will also show progress toward the related outcome of "reduce the number of transportation-related deaths" for DOT's "safety" strategic objective. This is an improvement over DOT's 1999 plan in which we found that DOT's performance goals typically covered only a portion of the strategic goals and the link between annual performance goals and strategic goals could be improved. DOT's plan also presents trend and baseline data for each performance measure and goal. For the example given above, the plan indicates the performance targets from 1999 to 2004 and also presents actual data for these targets for 1999 to 2002. In its 1999 plan, DOT had provided baseline data for most of its performance goals and measures as well. This information, along with the clearly linked performance goals and strategic objectives, helps to show DOT's progress in achieving its goals. As in the 1999 plan, DOT's 2004 performance plan generally covers each program activity in its budget request for fiscal year 2004. An appendix to the performance plan lists DOT's program activities and indicates the proposed funding level for each program by strategic objective. However, as in its 1999 plan, a few programs do not seem to be linked to the strategic objectives elsewhere in the plan. Capital grants to the National Passenger Rail Corporation (Amtrak) and the Bureau of Transportation Statistics' Office of Airline Information are both linked to the "mobility & economic growth" strategic objective in the budget crosswalk, but they do not appear in the discussions contained within that strategic objective section. When the 2004 plan was published in February 2003, DOT had not yet released its new reform strategy for Amtrak, which was made public in July 2003. Still, the inclusion of information on Amtrak, as well as a discussion of the Bureau of Transportation Statistics' Office of Airline Information, would provide a clearer picture of how DOT intends to achieve its goals. The discussions of each performance goal have sections entitled "Other Federal Programs with Common Outcomes." In this section, the plan describes crosscutting programs and other agencies with which DOT works. For example, the plan states that the Research and Special Programs Administration of DOT continues to develop the National Pipeline Mapping System with the Federal Energy Regulatory Commission, the National Oceanic and Atmospheric Administration, the Department of Energy, the U.S. Geological Survey and others, in order to help analyze risks to environmentally sensitive and populated areas. This supports DOT's efforts to reduce pipeline incidents. Yet for several goals, coordination efforts are not described. One example of this is in the highway congestion section where the plan states that the Federal Highway Administration works closely with the Department of the Interior, Department of Agriculture, and Department of Defense agencies to improve mobility on federally owned lands. However, the plan does not describe the specific actions that are being taken to improve mobility. Our 1998 report stated that DOT's contribution or role was not described in many of the crosscutting programs listed in DOT's 1999 performance plan. The 2004 performance plan generally addresses performance and accountability challenges we previously identified. The discussions of these management challenges are included in the plan by the performance goal and programs to which they are related. For example, in discussing highway safety, DOT addresses our concerns on transportation safety, specifically through the use of safety belts. The strategies include continuing the National Highway Traffic Safety Administration's safety belt outreach to high-risk populations and encouraging states to embrace "Click It or Ticket" as the message or theme for their Buckle Up Campaigns. A performance measure related to this management challenge included in the plan is "percentage of front seat occupants using safety belts." However, not all of the management challenges have related measures and goals. For example, we have previously identified building human capital strategies as a management challenge for DOT. A section within the plan focuses on an "organizational excellence" objective to implement the PMA. Strategic management of human capital strategies is discussed in this section but no goals or measures are given to show DOT's progress with these strategies. Still, this shows some improvement over the 1999 plan, which generally covered management challenges, but did so in a separate appendix without explaining how the challenges were related to the rest of the plan. We noted that this area could be improved by including goals and measures related to resolving these challenges. DOT's Fiscal Year 2004 Performance Plan Provides a Specific Discussion of Strategies and Resources: DOT's 2004 performance plan shows several improvements over its 1999 performance plan, providing a specific discussion of strategies and resources. Discussions of each performance goal include a section titled "Strategies and Initiatives to Achieve 2004 Target." These sections include a variety of means by which DOT intends to accomplish its performance goals. One example would be for DOT's performance goal to "reduce pipeline hazmat (hazardous materials) spilled 30 percent by 2006, from the last five years' average spill rate." The strategies for this goal include enforcing operator qualification requirements, expanding monitoring technology that can help prevent construction- related damage to pipelines, and developing regulatory standards for leak detection technology. This shows progress from when we reported that DOT's 1999 performance plan lacked sufficient information to clearly link the strategies to performance goals in many cases. In contrast to its 1999 performance plan, DOT's 2004 performance plan generally discusses the human, capital, information, and other resources needed to meet its performance goals. Each performance goal section of the performance plan includes a graph showing the enacted funding for fiscal year 2002, and the proposed funding for fiscal years 2003 and 2004. The organizational excellence objective for DOT describes DOT's human capital and information technology resources and strategies. For example, one of DOT's strategies for strategic management of human capital is to "establish a corporate approach to target recruitment efforts, with special emphasis on cross-modal, mission-critical occupations," which includes a pilot program for centrally recruiting and training entry-level employees for one or more mission-critical occupations. The 2004 plan also discusses external factors that could hamper DOT's ability to achieve its performance goals. In our review of DOT's 1999 performance plan, we noted that the plan could be improved by recognizing more external factors and by discussing actions that DOT could take to mitigate the effects of these factors. In contrast, external factors are listed for most of the performance goals in the 2004 plan. For its transportation accessibility goals, DOT states that as the population ages, more people will require accessible public transit, for which states and local agencies decide how best to allocate federally provided resources. One of the strategies that addresses this external factor is the Special Needs of Elderly Individuals and Individuals with Disabilities grants, which DOT states will help meet the transportation needs of the elderly and persons with disabilities when regular transportation services are unavailable, insufficient, or inappropriate to meet their needs. DOT's Fiscal Year 2004 Performance Plan Provides Full Confidence That Performance Data Will Be Credible: The 2004 plan provides full confidence that DOT's performance data will be credible. As in the 1999 performance plan, the 2004 performance plan contains a section, entitled "Performance Data and Performance Measurement," that discusses the means that DOT uses to verify and validate its data. But unlike the 1999 plan in which this discussion was broad and not linked to specific goals and measures, the 2004 plan also contains an appendix that provides the following for each of DOT's measures: the source of the data, limitations of the data, observations on the quality of the data, work planned or ongoing to improve data quality, and any known biases. Finally, DOT has compiled source and accuracy statements,[Footnote 109] which provide more detail on the methods used to collect the data, sources of variation and bias in the data, and methods used to verify and validate the data. The presentation of data limitations in DOT's 2004 performance plan also shows progress from its 1999 plan. The Performance Data and Performance Measurement section includes a general discussion of DOT's data limitations. This discussion includes limitations for the internal and external data used by the department. Specific limitations for internal data can be found in the aforementioned source and accuracy compendium, while details on the limitations of external data are given in the appendix on performance measures. In our report on DOT's 1999 performance plan, we stated that information on data limitations was lacking for most measures and that the plan could be improved by more consistently addressing the data limitations throughout the plan. [End of section] Appendix V: Observations on Agencies' Annual Performance and Accountability Reports: To help Congress and the President determine agencies' actual performance and progress in achieving strategic plan goals, GPRA requires each agency to prepare a report on program performance for the previous fiscal year.[Footnote 110] One of our objectives was to assess the overall quality of agencies' annual performance and accountability reports and the extent to which selected elements of agency reporting have improved. To meet this objective, we judgmentally selected six agencies--Education, DOE, HUD, SBA, SSA, and DOT--using criteria, such as agency size, primary program type, and previous GAO reviews. To assess the overall quality and improvements made to the agencies' performance and accountability reports, we relied on requirements and guidance contained in GPRA and accompanying committee report language,[Footnote 111] guidance to agencies from OMB for developing performance reports,[Footnote 112] interviews with agency officials, the Chief Financial Officers Act,[Footnote 113] our previous reports,[Footnote 114] and our knowledge of agencies' operations and programs. To assess the quality of the six agencies' performance and accountability reports, we categorized each report based on the degree to which it addressed three characterizations: (1) picture of performance, (2) link between resources and results, and (3) credibility of performance information. To assess the degree to which an agency's report provided a clear picture of performance across the agency, we reviewed the extent to which the report addressed elements required by GPRA. The annual performance report should: * describe the performance indicators established in the agency's annual performance plan, along with the actual program performance achieved compared with the performance goals expressed in the plan for that fiscal year; * review the success of achieving the performance goals of the fiscal year; * provide actual results for the 3 preceding fiscal years; * evaluate the performance plan for the current fiscal year relative to the performance achieved toward the performance goals in the fiscal year covered by the report; * explain and describe where a performance goal has not been met or a corresponding level of achievement if an alternative form is used, as well as why the goal was not met, plans and schedules for achieving the established performance goal, and if the performance goal is impractical or infeasible; * describe the use and assess the effectiveness of achieving performance goals of any waivers; and: * include the summary findings for those program evaluations completed during the fiscal year covered by the report.[Footnote 115] We also looked at the extent to which the reports clearly discussed progress achieved in addressing the major management challenges previously identified by us or others. For agencies that choose to issue a performance and accountability report, the Reports Consolidation Act of 2000 requires that the report include a summary of the most serious management and performance challenges facing the agency, as identified by their IG, and a brief assessment of the agency's progress in addressing those challenges. In assessing the clarity of the performance information, we also looked at selected qualitative characteristics used by the Association of Government Accountants, in conjunction with the Chief Financial Officers Council, in assessing performance and accountability reports for the Certificate of Excellence in Accountability Reporting.[Footnote 116] These characteristics included (1) whether there was a clear relationship between the performance information in the report and the goals and objectives contained in the strategic and annual performance plans, (2) the extent to which the agency limited the measures it discussed to those that were most significant for its programs, and (3) the extent to which the report was user friendly by being well- organized, concise, readable, and making effective use of graphics to ease understanding of narrative information. We characterized the clarity of each report in one of four ways: (1) clear, (2) general, (3) limited, or (4) unclear, based on the extent to which the 2002 report addressed the elements required by GPRA and the other informative practices we described. Both GPRA and the CFO Act emphasized the importance of linking program performance information with financial information as a key feature of sound management and an important element in presenting to the public a useful and informative perspective on federal spending. Similarly, the current administration's ambitious agenda for performance budgeting, calling for agencies to better align budgets with performance goals and focus on capturing full budgetary costs and matching these costs with output and outcome goals, suggests that agencies need to develop integrated financial and performance management systems that will enable the reporting of the actual costs associated with performance results. Although linking resources to performance goals is not a requirement of GPRA, the committee report for GPRA suggested that developing the capacity to relate the level of program activity with program costs, such as costs per unit of result, costs per unit of service, or costs per unit of output, should be a high priority. We have reported that world-class financial management practices call for enterprisewide systems to integrate financial and operating data to support both management decision making and external reporting requirements. To assess the degree to which an agency's report discussed the relationship between resources and results, we characterized each report as having a (1) clear relationship, (2) general relationship, (3) limited relationship, or (4) no relationship. Finally, to assess the degree to which an agency's plan provided confidence that the agency's performance information would be credible, we examined how each report discussed the quality of the data presented. To help improve the quality of agencies' performance data, Congress included a requirement in the Reports Consolidation Act of 2000 that agencies assess the completeness and reliability of their performance data. Under the act, agencies were to begin including this assessment in the transmittal letter with their fiscal year 2000 performance reports. Agencies were also required to discuss in their report any material inadequacies in the completeness and reliability of their performance data and discuss actions to address these inadequacies. We have previously reported on other practices that enhance the credibility of performance data that are not specifically required by GPRA.[Footnote 117] For instance, discussions of standards and methods used by agencies to assess the quality of their performance data in their performance reports provides decision makers greater insight into the quality and value of the performance data. We also reported on additional practices, in several agencies' performance reports, that would help foster transparency to the public and assist decision makers in understanding the quality of an agency's data. The additional practices we observed included (1) discussions of data quality, including known data limitations and actions to address the limitations and (2) discussions of data verification and validation procedures. To address the extent to which a report provided confidence that performance information was credible, we characterized each report as providing (1) full confidence, (2) general confidence, (3) limited confidence, or (4) no confidence. In conducting our reviews, to the extent information was available in prior assessments, we compared our findings of agencies' fiscal year 2002 reports to our assessments of reports for fiscal year 1999.[Footnote 118] A more detailed discussion of our scope and methodology and the criteria we used can be found in appendix I. Table 13 shows the results of our assessment of the six agencies' reports. Table 13: Characterizations of Agencies' Fiscal Year 2002 Annual Performance and Accountability Reports: Department/agency: Department of Education; Characterizations: Picture of performance: (unclear, limited, general, clear): Limited; Characterizations: Resources linked to results: (no, limited, general, clear): Clear; Characterizations: Data credible: (no, limited, general, full): General. Department/agency: Department of Energy; Characterizations: Picture of performance: (unclear, limited, general, clear): General; Characterizations: Resources linked to results: (no, limited, general, clear): Limited; Characterizations: Data credible: (no, limited, general, full): Limited. Department/agency: Department of Housing and Urban Development; Characterizations: Picture of performance: (unclear, limited, general, clear): General; Characterizations: Resources linked to results: (no, limited, general, clear): No; Characterizations: Data credible: (no, limited, general, full): General. Department/agency: Small Business Administration; Characterizations: Picture of performance: (unclear, limited, general, clear): Limited; Characterizations: Resources linked to results: (no, limited, general, clear): General; Characterizations: Data credible: (no, limited, general, full): General. Department/agency: Social Security Administration; Characterizations: Picture of performance: (unclear, limited, general, clear): General; Characterizations: Resources linked to results: (no, limited, general, clear): Limited; Characterizations: Data credible: (no, limited, general, full): General. Department/agency: Department of Transportation; Characterizations: Picture of performance: (unclear, limited, general, clear): General; Characterizations: Resources linked to results: (no, limited, general, clear): No; Characterizations: Data credible: (no, limited, general, full): Full. Sources: U.S. Department of Education, U.S. Department of Education FY 2002 Performance and Accountability Report; (Washington, D.C. 2003); U.S. Department of Energy, Performance and Accountability Report, Fiscal Year 2002; (Washington, D.C. 2003); U.S. Department of Housing and Urban Development, Fiscal Year 2002 Performance and Accountability Report; (Washington, D.C. 2003); Small Business Administration, Fiscal Year 2002 Performance and Accountability Report; (Washington, D.C. 2003); Social Security Administration, Performance and Accountability Report, Fiscal Year 2002; (Washington, D.C. 2002); U.S. Department of Transportation, Fiscal Year 2002 Performance and Accountability Report; (Washington, D.C. 2003). [End of table] The remainder of this appendix discusses our observations on the quality of the agencies' annual performance and accountability reports we reviewed and, to the extent information was available from our prior reviews, how the quality has changed since the agencies submitted their first reports on fiscal year 1999 performance. We did not independently verify or assess the information we obtained from agency annual performance reports. If an agency chose not to discuss its efforts concerning elements in the report, it does not necessarily mean that the agency is not implementing those elements. Observations on the Quality of Education's Fiscal Year 2002 Performance and Accountability Report: Education's fiscal year 2002 Performance and Accountability Report comprises two volumes--the main volume and a second volume including performance reports for the agency's individual programs. In our assessment, we did not review the second volume, which includes very detailed, discrete, and disaggregated performance information with over 350 individual measures for the Office of Civil Rights, IG, and 117 Education programs in 60 clusters. Although Education's report included many features designed to present its performance information clearly, the overall clarity was limited by the significant amount of performance information that was unavailable to show Education's performance results. In contrast, Education's report very clearly related its performance to its costs by using both graphics and text to provide the agency's estimate of appropriations associated with achieving each of its six strategic goals, 24 objectives (long-term goals), and individual programs. Finally, Education provided a general level of confidence in the quality of its data, primarily because of its recognition of the challenges it faces on the timeliness, reliability, and validity of its data. Education's recent efforts in undertaking a performance-based data management initiative in partnership with state leaders to allow timely and ready access to high-quality achievement and other performance data, which the IG said would address many of the related concerns identified during IG audits, also aided in the level of confidence in the data. Education's Fiscal Year 2002 Report Provided a Limited Picture of Performance: Education's 2002 performance report is directly aligned with the goals and measures in the agency's 2002-2007 strategic plan and its 2002-2003 annual plan. Of the 210 measures included in the agency's strategic plan and annual plan, 120 were identified for measurement in fiscal year 2002, and all of these are addressed in the performance report. The report contains sections on changes planned to enhance performance on the basis of results. For each measure, the performance report includes trend data, with a table showing actual data from fiscal years 1999 through 2002; in some cases, the table also includes data from fiscal year 1998. When data are not provided, the table indicates that they were not applicable or not available. Overall, the report contains clear, succinct figures and a table summarizing the status of all 120 measures, as summarized in figure 19. Figure 19: Summary of Education's Performance Indicators for Fiscal Year 2002: [See PDF for image] [End of figure] However, while Education's 2002 report does review the levels of success for its performance goals[Footnote 119] for fiscal year 2002, there is a critical limitation. As we observed in our review of Education's 1999 report,[Footnote 120] data were not yet available for many measures in the 2002 report. Specifically, 2002 data were available for only 41 of the 120 measures; the rest were characterized as "Pending: Data Not Yet Available" (63) or "Incomplete: Data Not Expected" (16). Despite the numerous strengths in Education's 2002 performance report, the picture of performance for Education presented in this report is limited mainly because of the lack of data for so many of its 2002 targets. However, Education recognizes the challenges created by its limited access to timely, reliable data: We still face significant challenges to meeting our national education goals. Primary among these challenges is access to timely, reliable data on our performance in meeting our goals and implementing our programs. Our efforts to identify effective and ineffective programs in the Department are severely limited by the frequent absence of useful data about them. In FY 2002 we designed a performance-based data management initiative which will provide much more robust information about our programs and strategic objectives, as well as provide a strong foundation for educational research and evaluation. This data management initiative is being undertaken in partnership with state leaders and the software industry and is expected to result in an electronic data system that will allow ready access to high-quality achievement and other performance data in a timely and seamless manner in the future. Because the lack of data for so many of its targets blurs the picture of performance, Education should take every possible step to complete, as quickly as possible, its newly established performance-based data management initiative. While lacking data for so many of its measures, Education's 2002 report provides an explanation for measures with pending or incomplete data. For pending data, the report states that comparisons to targets will be made in the subsequent performance and accountability report, in addition to citing the department's performance-based data management initiative. The report further indicates that measures with incomplete data were so characterized because methods to collect data were not ready in time to measure fiscal year 2002 results, data collection did not occur, or data collection was delayed. The report goes on to say that, for these measures, Education will put methods in place to measure fiscal year 2003 results, develop other data sources, or revise its measures to measure results differently. In addition, for each incomplete measure, the report clearly describes why data are incomplete and what will be done to address the situation. For example, for its measure on the percentage of states with complete school accountability systems in place, as required by the No Child Left Behind Act,[Footnote 121] the report explains that the requirements under this act are more extensive than in the past, that states that had met prior requirements may not yet meet new requirements, that the department had decided regulation would be necessary, and that regulations had not been finalized to define a complete school accountability system. In addition, the report almost always included explanations of performance and sometimes provided information on why targets were not met when that was the case. However, such information was not always easy to find, as the report did not always include it in the same area as the related measure. For most measures, including those that did not meet their targets, the report provided information on the steps Education is taking or plans to take to improve or enhance performance. For example, the 2002 target for a measure on the number of children attending charter schools was 690,000 and the actual result was 575,000. To improve performance on this measure, Education's report says that it is distributing guidance and information to encourage parents to consider charter schools, using both publications and its Web site to promote charter school enrollment, and sponsoring a charter schools Web site with information on federal assistance for charter schools. However, it was not always clear how the steps cited would improve or enhance performance. For example, for four measures on advanced placement (AP) achievement that were not met in 2002, the strategy given for improving performance is to continue to support increasing AP achievement through the Advanced Placement Incentives program, but the report does not include an explanation of, or any information on, this incentives program. Education's report included an appendix entitled "Findings from FY 2002 Evaluations" and included summaries of the findings from nine GAO reports and eight other studies completed in fiscal year 2002. For example, the Education for Homeless Children and Youth Program: Learning to Succeed report comprised two studies that found that homeless students are best served when promising practices are implemented as part of a comprehensive and coordinated education program for the homeless. The 2002 performance and accountability report also contained an appendix consisting of the IG's summary of serious management challenges, including financial management, federal student aid programs, information technology, program performance and accountability, and human capital. For each of these challenges, the IG provided information on Education's progress in addressing them. Under program performance and accountability, for example, the IG pointed out that a significant amount of the data used to measure education programs were provided by state and local education entities and that it is imperative that these data are accurate, so as to provide Congress, OMB, and the public with an objective measure of the success of education programs. The IG said that Education has recognized the importance of improving data quality and addressed this issue in its performance plan. The IG's summary of serious management challenges also included references to GAO's high-risk list with respect to federal student aid programs and human capital management, and Education's report included measures related to both of these areas. Specifically, for its 2002 measure to have Federal Student Aid (FSA) leave the GAO high-risk list by 2003 and not return, the report states that the department "almost met" its 2002 target by achieving 94 percent of its FSA High Risk Plan, and it described the shortfall as not significant or material. In contrast, in our review of Education's 1999 Performance Report,[Footnote 122] we noted that the department did not have goals, objectives, or measures related to problems with its student assistance programs. In addition, for the six measures in Education's 2002 report under its strategic goal to improve the strategic management of its human capital, the department reports that four targets were pending, one was incomplete, and one had set the baseline. With respect to the GPRA requirement[Footnote 123] that agencies' performance reports include a review of the performance goals and evaluation of the performance plan relative to the department's strategic human capital management, Education's report discusses its human capital management strategic goal and related performance goals in the context of its human capital management plan, One-Ed. Education's 2002 Report Showed a Clear Relationship between Resources and Results: Education's 2002 report included information for each of its six strategic goals and 24 objectives that clearly linked the department's resources with its efforts to achieve specific results. While the department was not able to break the costs down by each of its measures and targets, the report used both graphics and text to provide the department's estimate of appropriations associated with achieving each of its six strategic goals, 24 objectives (long-term goals), and individual programs. For example, for each of its objectives, the report used a pie chart to show the percentage of the department's appropriation that supports the objective, the rest of the strategic goal the objective falls under, and the other five strategic goals. An example is shown in figure 20 for the objective to ensure that all students read on grade level by the third grade. Figure 20: Inputs: Allocating Funds for Education's Objective to Ensure That All Students Read on Grade Level by the Third Grade: [See PDF for image] [End of figure] The text accompanying each chart listed the dollar amount supporting the objective's activities, the percentage of the strategic goal's allocation that amount represented, the individual programs that supported the objective, and the dollar amount from salaries and expenses that was included in the dollar amount for the objective. In addition, in the report's appendixes, a table summarizing fiscal year 2002 appropriations and staffing allocated by goal and objective also included the FTEs under staffing for each strategic goal and objective. Another table provided a percentage breakdown of each objective's appropriations by 146 agency programs. Education's Fiscal Year 2002 Report Provided General Confidence That Performance Data Were Credible: Education's 2002 performance report provided general confidence that the agency's data were credible because of its recognition of the challenges it faces on the timeliness, reliability, and validity of its data; its straightforward disclosure of these challenges; and its recent efforts to address them. In Education's transmittal letter, the department secretary said that the information contained in the report is "as complete and reliable as we have available." However, Education's report recognized that one of the agency's significant challenges to meeting its national education goals is access to timely, reliable data on performance and that the lack of useful data severely limits efforts to identify effective and ineffective programs. The report further explained that 97 percent of the department's funding is awarded to third parties, including, for example, state and local agencies, that have an impact on the measurement of results, especially the timing of data collection. Thus, Education recognized in its report that it had limited control over the data it must use to report results. Similarly, in the IG's summary of serious management challenges, the report noted that Education needs to improve its controls over the timeliness, reliability, and validity of data. Moreover, Education's report included information on recent steps it has taken to address its data challenges. In addition to a discussion of its data management initiative to develop an electronic data system providing access to timely, high-quality data, the report included an appendix with the agency's Information Quality Guidelines and draft Data Quality Standards presented in an abbreviated format. The discussion of data quality standards recognizes the importance of data quality concepts to the process of developing high-quality performance measures. The eight standards provided are: validity, accurate definitions, accurate counts, editing, calculation, timeliness, reporting, and burden reduction. To facilitate the use of the standards, Education reported that it created a data quality checklist and regularly held classes to teach staff how to apply the standards. The IG's summary of serious management challenges gave the department credit for this effort, pointing out that these guidelines address many of the concerns identified during IG audits and that the department plans to disseminate these guidelines to the chief state school officers. The report also provided explanations of data sources and data quality for most measures. For example, for the measures on the percentages of 12thgrade students scoring at or above the basic and proficient levels on the National Assessment for Educational Progress (NAEP) reading test, the source is given as: "U.S. Department of Education; National Center for Education Statistics, (NAEP); The Nation's Report Card, Reading." Under data quality, the report states that NAEP data are validated using rigorous National Center for Education Statistics statistical standards. For most of the measures, the explanations of data quality contain similar information on data validation. However, the data quality information only sometimes identifies what limitations are relevant, if any. For example, for a measure on the percentage of managers satisfied with services received from Education's Office of Management when hiring staff, the department relied on an internal survey of managers for its data. Although the response rate for this survey was 22 percent, the report did not say whether this was a limitation to the data collected. For a few of Education's measures, the report stated that no data limitations had been noted. It would be better if the agency clearly stated whether the data for each measure had limitations or not, and, if so, what they were. Observations on the Quality of DOE's Fiscal Year 2002 Annual Performance and Accountability Report: DOE's 2002 Annual Performance and Accountability Report provided a general picture of intended performance by explaining in detail the progress made in meeting performance measures and addressing management challenges. It also provided a limited discussion of the costs incurred to achieve DOE's performance goals by organizing its report by major program activities, the costs of these activities, and their corresponding goals. Finally, the report provided a limited level of confidence that data will be credible because the report did not include a discussion on data limitations. DOE's Fiscal Year 2002 Report Provided a General Picture of Performance: DOE's 2002 Performance and Accountability Report provided a general picture of performance in meeting its goals and measures. The report contained a detailed explanation of progress for each of its performance measures, which were referred to as "targets," by identifying whether each measure was met, not met, or had mixed results, as shown in figure 21. In addition, the report identified the operational processes, technology, human capital, and other resources used to achieve each performance measure. The results for the past 3 years of performance measures related to each goal were also reported so that performance trends could be identified. However, the report did not clearly explain how the results of the performance measures reported contributed to achieving the performance goals in DOE's annual performance plan and strategic plan. Figure 21: Summary of DOE's Performance Indicators for Fiscal Year 2002: [See PDF for image] [End of figure] Each performance measure that was not met or had mixed results contained a plan of action to achieve the measure in the future. In addition, the majority of fiscal year 2002 measures that were not met contained a clear explanation as to why they were not met or had mixed results. For example, a measure that required Southeastern Power Administration to meet its planned repayment of principal of federal investment was not met due to severe drought. DOE's report explained that to achieve the measure in the future, Southeastern plans to change its rate design, propose rate increases to obtain greater revenue, and increase cost recovery from fixed charges. DOE also discussed the progress it made in addressing performance and accountability challenges. The report identifies significant issues for fiscal year 2002 that, as stated in the report, merit a higher level of attention and focus in the department. Each of the challenges were linked to a goal and its related performance measure(s). In addition, actions taken to address each challenge and the progress made on those actions were identified. For example, one of the challenges identified was the need for DOE to meet federal requirements for improved and more cost-effective use of information technology. A related goal to deal with this challenge was for DOE to promote the effective management of information technology resources in the department. To address this challenge, the report stated that DOE realigned its management structure for information technology issues, established an enterprisewide license for Microsoft software, and launched an e- government applications task force to identify high-priority e- government investments, among other actions. In our prior review of DOE's 2001 performance report, we also found that DOE had made progress in addressing all eight of its major management challenges.[Footnote 124] It is unclear, however, how program evaluations were used to assess performance because DOE did not include a summary of program evaluation findings in either the fiscal year 1999 or 2002 reports. According to DOE officials, a section on program evaluations was not included in fiscal year 2002 and one is not planned for fiscal year 2003 in order to limit the amount of detail included in the report. DOE's Fiscal Year 2002 Report Showed a Limited Relationship between Resources and Results: DOE's 2002 performance report provided a limited discussion of how its resources were related to its performance. As in our review of the 1999 report, DOE's 2002 report, which also includes DOE's financial information, continued its practice linking the department's performance information to the costs of its program activities. For the majority of its program activities, the 2002 report included the program activity's net costs and related performance information for fiscal years 1999 through 2002. During our review of DOE's 1999 performance report, we were supportive of DOE's efforts to link its performance goals and measures to the program activities in the President's Budget. However, DOE has not moved beyond presenting its performance and cost information by program activity, instead of by strategic or annual performance goal or objective. A report that presented cost and performance information by performance goals in addition to other presentations would more clearly identify the costs associated with the achievement of each goal. According to DOE officials, the department plans to link its individual performance measures to the costs of program activities in future reports. DOE's Fiscal Year 2002 Report Provides Limited Confidence That Performance Data Are Credible: DOE's reporting on data credibility has improved but is still limited. Based on our review of DOE's 1999 performance report, a key improvement made to the 2002 performance report was the department's ability to report on its data validation and verification processes. DOE's 1999 report did not discuss the implementation of DOE's verification and validation plan or provide any evidence that the data quality was sufficient for assessing the department's performance. The 2002 report met some requirements of the Reports Consolidation Act by including a statement in the report's transmittal letter assessing the completeness and reliability of the data. The letter did not discuss any material inadequacies with DOE's performance data. The report also included a high-level discussion on how DOE will validate and verify its data and refers the reader to its 2003 annual performance plan for further details. For example, the report stated that DOE's end-of-year reporting process includes certification by heads of organizational elements on the accuracy of reported results, followed by a review for quality and completeness by DOE's Office of Program Analysis and Evaluation. Although the department has improved on reporting its data verification and validation processes, it has not improved on reporting any existing data limitations. Neither the 1999 nor the 2002 report included an overall discussion of the limitations to the data or steps DOE would take to address those limitations, although the 2002 performance report did identify minor data limitations for a few specific goals. DOE officials stated that a discussion on data limitations was not included in the 2002 report because the department already reports on this information in the annual performance plan and they thought it was redundant to put it in the report. Observations on the Quality of HUD's Fiscal Year 2002 Annual Performance and Accountability Report: Compared to its fiscal year 1999 performance report, HUD's fiscal year 2002 Performance and Accountability Report provides a general picture of what the department accomplished by including, among other things, a report card listing its performance indicators with the corresponding achievement, a list of program evaluations concluded during the fiscal year, trend information for some of its performance indicators, a discussion of the department's attempts to address its performance and accountability challenges, and visual aids to illustrate information on its performance. In a few instances, the report mentions whether certain performance goals are impractical or unfeasible. However, the report is not as clear as it could be because it does not (1) explain how it plans to address performance targets that were not met during the fiscal year; (2) include an evaluation of the fiscal year 2003 performance plan relative to the performance information presented in the performance report for the fiscal year; or (3) include an evaluation of the fiscal year 2003 performance plan relative to the performance information presented in the performance report for fiscal year 2002. The report does not show the relationship between resources and results by linking expended dollar amounts to specific program objectives. The report provides general confidence to the reader that the data presented are credible by providing background information on each performance indicator and discussing the results and analysis of the most recent data. HUD's 2002 Report Provided a General Picture of Performance: Overall, HUD's fiscal year 2002 Performance and Accountability Report provides a general understanding of what the department's mission is and what it accomplished during the previous fiscal year.[Footnote 125] Since we first reviewed its report for fiscal year 1999, HUD has made progress in developing its performance report to comply with GPRA.[Footnote 126] In reviewing HUD's fiscal year 1999 performance report, we noted that it only contained performance information for three of the department's four outcome measures. HUD's report for fiscal year 2002 includes a report card for each strategic goal listing performance targets that were met. The report card also provides an explanation for many performance targets that were not marked as being met during the fiscal year. Although the report includes visual aids to enhance readers' understanding of progress made toward attaining performance targets, HUD could enhance the report by providing a summary of performance targets for all strategic objectives met (or not) during the fiscal year. HUD's performance report does not meet GPRA's requirement of including an evaluation of its fiscal year 2003 performance plan relative to the performance attained by the department in fiscal year 2002. Including this evaluation could provide readers some assurance that HUD takes into account prior performance information, such as unmet goals, to manage its performance in the fiscal year already under way. The report suggests that some of the performance targets that were not met during the fiscal year were impractical or unfeasible. For example, for its goal of "Increase the Rate of Homeownership," HUD mentions that the indicator can be resistant to increases above an undetermined level because homeownership is not practical or desirable for all households. Broad economic conditions, including employment, incomes and interest rates can affect homeownership rates. Likewise, HUD will no longer track a performance indicator that measures the percentage of low- income housing units containing threats to health and safety, such as exposed wiring, unvented heaters, holes in the floor, and rodents. HUD mentions that this indicator is not included in the fiscal year 2003 annual performance plan because of the difficulty of attributing the results to its programs. In several instances, HUD's annual performance report lacks a discussion of how it plans to address unmet performance targets as required by GPRA. For example, while HUD substantially met almost half of its performance targets in fiscal year 2002, the report does not mention what steps the department will take to address some of its unmet performance targets (see fig. 22). Figure 22: Summary of HUD's Performance Indicators for Fiscal Year 2002: [See PDF for image] [A] Rather than stating if some performance targets were met or not, HUD provided the following explanations: data not available; no performance goal for this fiscal year; third quarter of calendar year (last quarter of fiscal year, not entire fiscal year); calendar year ending in the current fiscal year; calendar year ending the previous fiscal year; other reporting period; results too complex to summarize; and baseline newly established. [End of figure] HUD continued to build upon the strengths of its earlier report by including trend information for some performance indicators it used to measure progress toward its targets during the past fiscal year. While not presented consistently throughout the report, trend information provides a context to understand HUD's performance and helps to show the extent to which HUD exceeded or fell short of expectations set for its performance targets. For instance, for HUD's performance indicator that tracks increases in the share of welfare families residing in public housing that move from welfare to work each year, the report mentions that in fiscal year 2002 the rate was 13.1 percent compared to 19.9 percent in fiscal year 2001. In preparing to implement the Temporary Assistance for Needy Families program, HUD originally estimated this indicator to be around 6.5 percent in fiscal year 1997. HUD's fiscal year 2002 report also mentions that it concluded several program evaluations during the fiscal year, a key requirement absent from its performance report for fiscal year 1999. The report also provides a brief summary of the main findings of these program evaluations. In fiscal year 2002, HUD concluded and published reports on 21 program evaluations covering five of its strategic goals. Similar to our findings on HUD's previous performance reports,[Footnote 127] HUD's fiscal year 2002 report discusses the steps the department took to address decade-long management challenges. For example, while HUD's report mentions that deficiencies remain in its financial management systems, in fiscal year 2002 the department initiated a project to design and implement an integrated financial system. Similarly, to address staffing imbalances and human capital challenges, HUD implemented the last phase of its Resource Estimation and Allocation Process in January 2002 and started to implement the Total Estimation and Allocation Mechanism, a tool that collects actual workload accomplishments and staff usage within the various operating components at HUD. HUD's 2002 Report Showed No Relationship between Resources and Results: While HUD has made some improvements in how it presents cost information in its report, it is still not useful for linking program objectives to specific dollar expenditures. HUD's report provides a summary of the cost of operations by each reporting segment, such as the total amount of money spent by the Federal Housing Authority and the Public and Indian Housing programs, and states that the total cost for fiscal year 2002 operations was $33 billion. However, the report does not reconcile these costs to specific program performance objectives, limiting the reader's ability to understand how HUD used its resources to carry out its objectives during the fiscal year. HUD's 2002 Report Provided General Confidence That Performance Data Are Credible: HUD's fiscal year 2002 performance report generally informs the reader on critical issues about the reliability of its performance data, an issue that was not discussed in detail in its earlier report. In its transmittal letter, HUD briefly discusses that in some instances the data used in the report were either incomplete and/or unreliable. The report includes background information, results, analysis, and a discussion of the data used for each performance indicator during the fiscal year. In discussing the data, in some instances HUD points out issues concerning data validity and accuracy and mentions steps HUD will take to correct problems. For example, to address problems with its indicator on the number of homeowners who have been assisted with the HOME program, HUD has established a team of managers, technical staff, and contractors to make a series of improvements to the Integrated Disbursement and Information System beginning in fiscal year 2003, which should reduce the need for data cleanup. Observations on the Quality of SBA's Fiscal Year 2002 Annual Performance and Accountability Report: SBA's fiscal year 2002 annual performance report shows several improvements over the agency's initial report. The report shows a general relationship between resources and results by including an analysis of resources used by each program in fiscal year 2002. A section on data validation and verification, which includes data limitations and remedies for those limitations, provides a general level of confidence in the credibility of SBA's performance data. While the 2002 report includes a scorecard to show the agency's overall performance, the report provides a limited picture of performance due to a lack of plans to meet unmet goals in the future and data that were unavailable to show progress towards a large share of SBA's performance goals, among other reasons. SBA's 2002 Report Provided a Limited Picture of Performance: SBA's 2002 performance report[Footnote 128] provides a limited picture of its performance. SBA's 2002 report includes a scorecard that shows overall agency performance for its 2002 goals, including trend data (when available) from fiscal year 1999 to fiscal year 2002, the fiscal year 2002 goal, and a column showing the percentage of the fiscal year 2002 goal achieved. However, based on the performance information provided in the fiscal year 2002 performance report, it can be difficult to gauge SBA's progress in achieving its goals. This is similar to our findings on SBA's 1999 report, which we noted was unclear as to how well SBA performed in achieving several of its performance goals for two of the key outcomes addressed in our 2000 report.[Footnote 129] Figure 23 summarizes SBA's progress on its 19 performance goals for fiscal year 2002. Figure 23: Summary of SBA's Performance Goals for Fiscal Year 2002: [See PDF for image] [End of figure] Data were unavailable for 10 of SBA's 19 performance goals in 2002. For the nine goals that had performance data available, SBA met seven. SBA's 2002 performance report included explanations for all of its goals that were unmet, deemed infeasible, or for which data were not available. For example, the report states that "homes restored to pre- disaster conditions" and "businesses restored to pre-disaster conditions" are no longer goals because SBA is reviewing its outcome measures for the disaster loan program. Also, data were not available for the "customer satisfaction" goal in the disaster assistance program because a Customer Service Survey for disaster loan recipients was not issued during the fiscal year due to having not received final clearance from OMB. This contrasts to our findings on SBA's 1999 report when the agency did not provide explanations for not meeting several of its performance goals. However, for the two goals that were unmet in 2002, "start-ups receiving 7(a) and 504 financing" and "jobs created and retained by SBIC clients," the report does not describe any plans for achieving the goals in the future. The lack of information for over half of SBA's performance goals and the absence of plans for achieving unmet goals limits our ability to assess the overall progress the agency made in fiscal year 2002, as well as the likelihood that it will improve its performance in the future. Several other factors limited our ability to evaluate SBA's performance in its fiscal year 2002 report. The report presents performance data in accordance with the goal structure of SBA's 2003-2008 draft strategic plan. The goal structure contained in the fiscal year 2002 performance report does not directly correspond with the goal structure presented in the 2002 performance plan. Only one of the report's strategic goals, "Help Families and Businesses Recover from Disasters," directly corresponds to the 2002 performance plan. Similarly, not all of the performance goals listed in performance scorecards in both documents correspond. For example, the performance goal "start-ups receiving 7(a) and 504 loans viable 3 years after receiving loan," listed in the scorecard in the 2002 report, is not listed in the 2002 performance plan. The report states that based on 2002 results SBA made "substantial modifications" to its fiscal year 2003 goals, but the report does not specifically discuss how the performance achieved in 2002 could affect the achievement of the 2003 goals. Finally, the report does not include the findings of program evaluations completed in fiscal year 2002. SBA states that it was unable to conduct program evaluations in 2002 due to a lack of funding and that the agency has requested funding for program evaluations in fiscal years 2003 and 2004. Similar to our findings on SBA's 1999 performance report, the agency continues to provide information indicating the agency's progress in addressing management challenges that have been previously identified. For example, we have previously observed that SBA needs to improve the quality of the performance measures that it uses for the disaster loan program. SBA states in its 2002 report that the agency is in the process of reevaluating its measures for the disaster loan program, and specifically that the methodology for measuring the number or percentage of homes and businesses restored through the program will be addressed by this review. SBA's 2002 performance report also discusses the agency's strategic management of human capital. One section of the report relating to the PMA describes SBA's transformation plan, which is to realign the agency's "organization, operation, and workforce to better serve its small business customers." An appendix to the report identifies fully developing and implementing the agency's human capital management strategy as one of SBA's major challenges. This section lists the actions that SBA needs to take to address this challenge as well as the progress the agency has made in implementing these actions. However, the report does not include a review of the performance goals and evaluation of the performance plan relative to the agency's strategic human capital management, as required by the Homeland Security Act of 2002. The report also contains two broad overviews and an appendix of GAO audits and recommendations, as well as a description of management challenges identified by the agency's Inspector General. One chart, entitled "Status of GAO Reviews Conducted at SBA in FY 2002," shows the review title, status of the review (open or closed), and the number of recommendations that came from these reviews. Another chart, entitled "Number of Open GAO Recommendations at End of FY 2002," lists the GAO report number and title, the year it was issued, and the number of recommendations remaining open. Further detail is provided in an appendix to the performance report, which lists GAO's outstanding recommendations, the status of the recommendations, and the estimated date of completion. Another appendix includes a report from SBA's Acting IG that describes the most serious management challenges SBA faced in fiscal year 2002. SBA's 2002 Report Showed a General Relationship between Resources and Results: SBA's 2002 performance report contains analyses of resources and results for SBA's programs that show a general relationship between resources and results. In the description of each program's performance, the report includes an analysis of the resources used by each program. For example, the fiscal year 2002 cost of the Advocacy Program was estimated to be $8 million, with 50 percent of the funds going to support the Office of Advocacy, 14 percent funding economic research, 11 percent for SBA's executive direction support, 16 percent for fixed costs, and 9 percent going to human resources, information technology, and procurement. The report contains crosswalks that show the relationship between SBA's strategic goals, outcome goals, performance goals, and programs. The resources used by each program can then be linked through this crosswalk to performance goals to generally show the resources needed for the results achieved towards the goals. However, the connection of resources to results could be more explicitly stated in the report if results and resources were also presented by performance goal. SBA's 2002 Report Provided General Confidence That Performance Data Are Credible: SBA's 2002 performance report provides general confidence that the agency's performance data are credible. In the letter transmitting its 2002 performance report, SBA states that the performance data for its credit and procurement assistance programs are complete and reliable, based on a systematic review of these data. SBA further states that it is "working to improve the completeness and reliability of the performance data for the advice provided to small business through SBA's resource partners." Data for this aspect of SBA's performance are collected through surveys, which the agency notes are neither consistent nor comparable, and from which client responses are difficult to obtain. This could be seen as a material inadequacy, of which a discussion is required by the Reports Consolidation Act. SBA discusses the actions it will take to address the quality of the surveys by stating in the transmittal letter that it is working to improve the survey instruments it uses to obtain performance data. SBA provides a detailed discussion of each performance indicator in a section of the report on data validation and verification. For each indicator, the report provides a definition, source, information on validation, and means for data verification. The verification process for several measures includes audits, independent reviews, and consistency checks. However, for these measures this is the only discussion of verification procedures and no further details are provided. Also, for several measures, such as "number of start-up firms financed by 7(a) & 504" and "regulatory cost savings to small businesses," SBA states that it does not independently verify the data. The report also addresses limitations to its data in the section on data validation and verification. In this section SBA states that it faces many challenges in acquiring high-quality data on both outputs and outcomes, from both internal and external sources. The strategies that SBA will use to address the shortcomings of its data quality are contained in this section as well, which include ensuring the validity of performance measures and data, fostering organizational commitment and capacity for data quality, assessing the quality of existing data, responding to data limitations, and building quality into the development of performance data. The section on data validation and verification in the 2002 report discusses how SBA plans to remedy the limitations for each indicator. For example, for the "customer satisfaction rate" measure of the disaster loan program, the report states that the surveys used for this measure only determine the satisfaction of those who received disaster loans and therefore do not address the satisfaction of those who did not receive the loans. The remedy listed for this limitation is to expand the survey to include all applicants. This is an improvement from SBA's 1999 report, which we noted did not discuss data limitations that could affect the quality of data used by SBA to assess its performance. Observations on the Quality of SSA's Fiscal Year 2002 Annual Performance and Accountability Report: On the whole, SSA's 2002 performance report generally showed the agency's progress towards its annual goals for fiscal year 2002.[Footnote 130] It showed continued emphasis on outcome-oriented goals and identified relevant results that were linked to individual strategic objectives. It also provided trend information, typically back to fiscal year 1999, and contained a brief discussion of the program evaluations completed during fiscal year 2002. SSA's strategic goals were linked to financial resources at a very high level, but none of the performance goals were associated with costs; thus, the cost of achieving (or not achieving) a particular goal was not clear. Additionally, the SSA Commissioner certified that SSA's data presentation was credible, but missing data and a lack of documentation of the methods and data used to measure its performance reduced the overall quality of the document. SSA's 2002 Report Provided a General Picture of Performance: SSA's 2002 performance report exhibited a general description of performance, including the identification of 14 key performance indicators out of a total of 69 indicators. Similar to our review of SSA's fiscal year 1999 report,[Footnote 131] we found that many of SSA's fiscal year 2002 goals and indicators were outcome oriented. In the fiscal year 2002 report, SSA plainly summarized progress on its 69 performance indicators, as shown in figure 24. Figure 24: Summary of SSA's Performance Goals for Fiscal Year 2002: [See PDF for image] [End of figure] In SSA's 1999 annual performance report, performance measures focused on activities rather than results, so it was difficult to determine the agency's real progress in achieving results. For example, the measures directly related to the outcome "long-term disability benefits are reduced because people return to the workplace" did not sufficiently track progress toward this key outcome. One of the measures was to "begin implementation of the 'Ticket to Independence' program, contingent upon enactment of supporting legislation in FY 1998."[Footnote 132] This measure was neither quantifiable nor measurable, and did not measure the number of beneficiaries achieving this outcome. In the 2002 report, SSA identified relevant results that are linked to strategic objectives. For example, one of SSA's objectives related to the strategic goal "deliver citizen-centered, world-class service" was to "maintain the accuracy, timeliness, and efficiency of service to people applying for Old Age and Survivors Insurance (OASI) and Supplemental Security Income (SSI) Aged benefits." SSA reported on the timeliness of OASI and SSI claims, as well as the implementation of software and infrastructure for paperless processing of claims, as the relevant results. In its 1999 performance report, SSA noted that a number of its goals were not met, such as those relating to accurate and timely disability determinations. Also, data on the accuracy of decisions at the initial level were not available, and accuracy at the appellate level was not measured. In its 2002 report, 10 percent of the goals were not met and 10 percent were almost met. SSA's report provided explanations for 17 performance goals SSA did not meet. However, not all of the explanations actually identified the reasons for SSA's not meeting its goals. For example, SSA did not meet its goals for the performance indicators "percent of 800-number calls handled accurately--payment" and "percent of 800-number calls handled accurately--service." The explanation noted that several quality initiatives were implemented, but SSA did not provide explanations as to why the goals were not met. SSA also noted that some of its performance indicators were being eliminated in favor of more focused and outcome-based goals. In some cases, SSA identified future plans to improve performance. In SSA's 2002 performance and accountability report, trend information was generally available for comparison of data back to fiscal year 1999. This information was helpful in making an assessment whether SSA was making progress towards its goals. SSA noted that it addressed all the IG's list of major management challenges in its report, and that it addressed the major management challenges we identified in its annual performance plan. SSA also addresses the progress it made against certain challenges GAO and the IG identified during fiscal year 2002 in its performance and accountability report.[Footnote 133] For example, SSA highlights components of its SSI Corrective Action Plan that are geared to improve the administration of the SSI program and get it removed from our high-risk list. The IG's report noted that SSA needs to have performance goals and measures that address the major management challenges facing SSA, as they are not all addressed. For example, performance measures were not established to address problems with the Earnings Suspense File and the integrity of the representative payee process. Finally, SSA's performance and accountability report contained a discussion of the program evaluations conducted, organized by strategic goal. However, the program evaluations SSA identified were typically surveys of people who did business with SSA or assessments of internal needs, such as a survey of training effectiveness and water/air quality surveys. While this is a slight improvement over its 1999 report, where there was only a brief summary of program evaluations, it would be helpful for SSA to report on whether and how its evaluations have helped answer questions about program performance and results. We have previously reported that evaluations can help agencies improve their measurement of program performance and/or understanding of performance and how it might be improved.[Footnote 134] SSA's 2002 Report Showed a Limited Relationship between Resources and Results: In the fiscal year 2002 performance and accountability report, SSA's performance goals were not aligned by budget account--rather, they were associated with strategic goals, which in turn cross budget accounts and programs. Thus, the monetary, human capital, and technological resources necessary to achieve many performance goals were not adequately described. The financial statements show a schedule of financing and a schedule of budgetary resources for each of SSA's major programs, and operating expenses were associated with four out of the five strategic goals.[Footnote 135] However, these resources were not broken down by performance goal, and were not linked to outcomes. Additionally, as reported by the IG, SSA needs to further develop its cost accounting system, which it began to use in fiscal year 2002; such a system would help to link costs with performance.[Footnote 136] SSA's 2002 Report Provided General Confidence That Performance Data Are Credible: While SSA provides data sources and definitions for the data supporting its performance indicators, some data issues continue to detract from SSA's performance report. SSA's transmittal letter noted that the performance and financial data presented are fundamentally complete and reliable, and that no material inadequacies were identified. Data sources are identified for many of the performance indicators, such as findings from evaluations and quality assurance reports. In certain cases, data limitations are identified; for example, SSA noted that data to support the "Percent of SSNs issued accurately" goal does not include SSNs (social security numbers) assigned via the Enumeration-at- Birth process and major errors identified by the Office of Quality Assurance that do not include these SSNs result in SSN cards being issued erroneously. Some data verification procedures are noted in the report, but verification procedures are not consistently discussed and data reliability and completeness is not ensured. The IG noted that of the 21 measures it reviewed, 16 were reliable; data or documentation of the methods used to measure SSA's performance were not available for the other five measures. The IG went further to say that even for the performance measures found to be reliable, SSA lacks documentation of the methods and data used to measure its performance. Finally, data were not available for 17 percent of the performance goals, so it was difficult to assess whether or not progress had been made in those areas. Observations on the Quality of DOT's Fiscal Year 2002 Annual Performance and Accountability Report: DOT's fiscal year 2002 performance report provided information that generally showed the department's performance and progress toward its goals. Summary tables within the report showed when DOT met or did not meet its targets and the report supplied brief analyses as to whether or not DOT would likely meet its targets for fiscal year 2003 based on actual performance in 2002. A separate section of the report on performance data completeness and reliability, along with an on-line compendium, provides a full level of confidence in DOT's performance data. However, the report does not clearly show the relationship between resources and results. DOT's 2002 Report Provided a General Picture of Performance: DOT's fiscal year 2002 performance report[Footnote 137] provides a general picture of the department's performance. The report includes performance summary tables, which show the progress made toward each strategic objective. These performance summaries include actual performance data from fiscal years 1996 to 2002 when possible, as well as the performance targets for fiscal year 2002 and whether or not the target was met. Similarly, we noted in our 2000 report[Footnote 138] reviewing DOT's 1999 performance report that performance information was clearly articulated, with summary tables listing the fiscal year 1999 goals and trend data, and checkmarks to indicate whether or not goals were met. Figure 25 summarizes DOT's overall performance on its 40 performance targets, as reported in its 2002 report. Figure 25: Summary of DOT's Performance Indicators for Fiscal Year 2002: [See PDF for image] [End of figure] According to the report, DOT met 24 (60 percent) of its performance targets. Fourteen (35 percent) of DOT's performance targets were not met. The report provides explanations for why five of these targets were unmet. For example, the target for the measure "number of passengers (in millions) in international markets with open skies aviation agreements" of the "mobility and economic growth" strategic objective was unmet. The target was set at 59.7 million passengers, while DOT's preliminary estimate for this measure indicated there were 57 million passengers. The report states that this target was unmet because passenger travel diminished in fiscal year 2002 due to the impact that the events of September 11, 2001, had on air travel. However, the report did not provide explanations describing why the nine other targets were not met. A DOT official stated that explanations for these unmet targets were not included in the 2002 report due, in part, to time constraints. In our 2000 report of DOT's 1999 performance report, we stated that for all of its unmet goals except transit fatalities, the department provided explanations for not meeting the goals related to the outcomes we observed. We noted in our report on DOT's 1999 performance report that the department supplied strategies to achieve its unmet goals in the future, for the areas we reviewed. However, of the 14 unmet targets in the fiscal year 2002 report, DOT provided future plans to achieve only two. For example, the report provided a plan for future achievement of the unmet target "percent of environmental justice cases unresolved after one year." The report stated that DOT's External Complaint Tracking System was being revised "to track complaints more closely, in a more timely way, and with a higher level of data quality." DOT is also developing guidance requiring more intensive legal staff involvement in external civil rights complaints, especially environmental justice cases. A DOT official stated that future plans were not included in the 2002 report for the other unmet targets due, in part, to time constraints. Data were unavailable for two of DOT's measures, "cumulative average percent change in transit passenger-miles traveled per transit market" and "employment sites (in thousands) made accessible by Job Access and Reverse Commute (JARC) transportation services." The report explains the reasons why data were unavailable for both of these measures and includes plans to provide these data in the future. Data were unavailable for the JARC program performance measure because DOT had not received data from JARC grantees to verify that fiscal year 2002 program targets had been achieved. DOT states that a new reporting system is being implemented, which should improve data gathering performance. Although the report does not identify any performance goals that were impractical or infeasible, it states that the measure on transit passenger-miles traveled had been changed in fiscal year 2002 because "it placed excessive emphasis on increasing ridership in the Nation's very largest urban areas." However, after using the measure for one year, DOT concluded that the measure should once again be modified to account for changes in the level of employment in each urban area. The report states that a recent study found that changes in the level of employment are a key economic factor related to changes in the level of transit ridership. Another strength of DOT's fiscal year 2002 performance report is an analysis of whether or not DOT will likely meet its planned performance targets for fiscal year 2003. Each discussion of performance goals contains an evaluation of the fiscal year 2003 performance plan target and whether or not it will likely be met based on the fiscal year 2002 data. For example, for its highway fatality rate targets, DOT says that the National Highway Traffic Safety Administration and the Federal Motor Carrier Safety Administration will be challenged to meet the established fiscal year 2003 targets because targets had not been met for fiscal year 2002. In other instances where DOT is sure that it will meet its targets for fiscal year 2003, it simply states so, as in the case of its aviation safety targets. DOT's 2002 performance report also includes information on completed program evaluations. There is a separate section in the report that discusses DOT's fiscal year 2002 program evaluations. Summaries of the findings of these evaluations are discussed in this section. For example, an evaluation of the Noise Set-Aside Portion of the FAA (Federal Aviation Administration) Airport Improvement Program found that funding for the program's noise compatibility projects was variable from year to year, making it difficult to forecast annual population benefits. As we found in the 1999 report, the major management challenges that DOT faces are generally discussed in the 2002 report. These discussions were contained within the program section to which they relate. The report also showed the progress DOT has made in addressing its management challenges. For example, we noted in our January 2003 report[Footnote 139] on DOT's major management challenges that FAA's financial management systems remained at high risk. DOT's 2002 report states that FAA created an Interim Fixed Asset System to centrally control and account for its property and that in fiscal year 2003, FAA will convert to use Delphi, DOT's financial accounting system. The 2002 performance report included a discussion on strategic human capital management as part of DOT's "organizational excellence" strategic objective. This discussion included a brief overview of DOT's human capital plan as well as strategies for strategic human capital management. For example, the report noted that FAA was redirecting 37,300 employees into a results-oriented Air Traffic Organization, "freeing most of the FAA to manage better and modernize more efficiently." However, the report did not include a review of the performance goals and evaluation of the performance plan relative to the agency's strategic human capital management, as required by the Homeland Security Act of 2002. DOT's 2002 Report Showed No Relationship between Resources and Results: DOT's 2002 performance report did not show the relationship between the resources it used and the results it achieved in fiscal year 2002. The financial portion of the report provided a statement of net cost for each of DOT's programs in fiscal year 2002. The report could be improved by providing net cost information for DOT's performance goals in the performance section of the report, similar to the funding information provided in the 2004 performance plan. DOT's 2002 Report Provided Full Confidence That Performance Data Are Credible: DOT's 2002 performance report provided a full level of confidence that the department's performance data were credible. In his transmittal letter, the Secretary of DOT stated that the 2002 report "contains performance and financial data that are substantially complete and reliable." The letter also stated that a section of the report assessed the inadequacies of DOT's performance data and provided plans to remedy those inadequacies. The "Performance Data Completeness and Reliability" section of the 2002 report generally discussed data completeness, reliability, and limitations. This section discussed an overall limitation in DOT's performance data. The report stated that much of DOT's performance data came from external sources, and therefore, the department had no direct control over the quality of these data. The report continues by stating that DOT takes limitations to its external data into account when it uses these data. The 2002 report noted that DOT has compiled source and accuracy statements that provide detail on the methods used to collect performance data, sources of variation and bias in the data, methods used to verify and validate the data, as well as data limitations.[Footnote 140] However, the online Source and Accuracy Compendium does not include this information for the goals and measures related to the department's organizational excellence objective. The compendium states that a small number of source and accuracy statements are not yet completed and that they will be added upon completion. Finally, the 2002 report also described strategies being undertaken to address the quality of data used by DOT. The report stated that a DOT intermodal working group addressed data quality issues by developing departmental statistical standards and by updating source and accuracy statements for all of DOT's data programs. The working group also worked to improve quality assurance procedures, evaluate sampling and nonsampling errors, and develop common definitions for data across modes. [End of section] Appendix VI: GAO Federal Managers' Survey Data: Q1. What is your current grade level? GS/GM-13 or equivalent (percent): 31.4; GS/GM-14 or equivalent (percent): 35.2; GS/GM-15 or equivalent (percent): 24.6; Senior Executive Service or equivalent (percent): 7.9; Other - please specify - Continue with next question (percent): 0.9; Number of respondents: 500. Q1a. If you answered "Other" in question 1 above, please enter your response below. Writing comment (percent): 0.9; Number of respondents: 3. Q2. In total, for how many years have you been a supervisor and/or a manager in the federal government? Mean: 13.1; Median: 13; Minimum: 1; Maximum: 43; Number of respondents: 497. Q2no. Or, if you have never been a supervisor or a manager in the federal government, please check the box below. Percent: 0.6; Number of respondents: 3. Q3. In your current role, approximately how many government employees are you responsible for? (Please answer for your permanent position. Please specify the total number. If none, enter 0.) Enter numeric digits only. Employees: Mean: 77.7; Median: 15; Minimum: 1; Maximum: 11,000; Number of respondents: 475. Q4. Please indicate where you currently work. (If you are currently on temporary assignment or on detail, please answer for your permanent work location.): Headquarters of my department or agency (percent): 28.9; A field office of my department or agency (percent): 60.7; Other - please specify - Continue with next question (percent): 9.5; No answer (percent): 1.0; Number of respondents: 503. Q4a. If you answered "Other" in question 4 above, please enter your response below. Writing comment (percent): 9.5; Number of respondents: 38. Q5. For those program(s)/operation(s)/project(s) that you are involved with, to what extent, if at all, do you consider your agency's strategic goals when participating in the following activities? (Check one box in each row.): a. Setting program priorities; To a very great extent (percent): 33.9; To a great extent (percent): 45.3; To a moderate extent (percent): 11.4; To a small extent (percent): 4.7; To no extent (percent): 1.8; No basis to judge/Not applicable (percent): 2.2; No answer (percent): 0.7; Number of respondents: 503. b. Allocating resources; To a very great extent (percent): 30.3; To a great extent (percent): 39.9; To a moderate extent (percent): 18.9; To a small extent (percent): 4.8; To no extent (percent): 1.8; No basis to judge/Not applicable (percent): 3.1; No answer (percent): 1.3; Number of respondents: 503. c. Adopting new program approaches or changing work processes; To a very great extent (percent): 33.6; To a great extent (percent): 39.2; To a moderate extent (percent): 16.5; To a small extent (percent): 4.6; To no extent (percent): 2.4; No basis to judge/Not applicable (percent): 1.9; No answer (percent): 1.8; Number of respondents: 503. d. Developing or refining program performance measures; To a very great extent (percent): 29.0; To a great extent (percent): 37.0; To a moderate extent (percent): 16.3; To a small extent (percent): 8.3; To no extent (percent): 4.1; No basis to judge/Not applicable (percent): 4.4; No answer (percent): 0.9; Number of respondents: 503. Q6. Are there performance measures for the program(s)/operation(s)/ project(s) that you are involved with? Yes (percent): 89.0; No (percent): 6.2; Do not know (percent): 3.8; No answer (percent): 1.0; Number of respondents: 503. Q7. To what extent, if at all, do you agree with the following statements as they relate to performance measures for the program(s)/ operation(s)/project(s) that you are involved with? (Check one box in each row.) We have performance measures that. a. Tell us how many things we produce or services we provide. (Output measures); To a very great extent (percent): 25.8; To a great extent (percent): 28.7; To a moderate extent (percent): 19.4; To a small extent (percent): 8.4; To no extent (percent): 4.8; No basis to judge/Not applicable (percent): 1.8; No answer (percent): 11.0; Number of respondents: 503. b. Tell us if we are operating efficiently. (Efficiency measures); To a very great extent (percent): 15.2; To a great extent (percent): 27.5; To a moderate extent (percent): 25.8; To a small extent (percent): 13.4; To no extent (percent): 5.5; No basis to judge/Not applicable (percent): 1.5; No answer (percent): 11.0; Number of respondents: 503. c. Tell us whether or not we are satisfying our customers. (Customer service measures); To a very great extent (percent): 16.8; To a great extent (percent): 29.8; To a moderate extent (percent): 21.6; To a small extent (percent): 12.6; To no extent (percent): 6.1; No basis to judge/Not applicable (percent): 1.6; No answer (percent): 11.4; Number of respondents: 503. d. Tell us about the quality of the products or services we provide. (Quality measures); To a very great extent (percent): 16.0; To a great extent (percent): 30.3; To a moderate extent (percent): 23.5; To a small extent (percent): 12.6; To no extent (percent): 5.6; No basis to judge/Not applicable (percent): 0.7; No answer (percent): 11.3; Number of respondents: 503. e. Demonstrate to someone outside of our agency whether or not we are achieving our intended results. (Outcome measures); To a very great extent (percent): 19.1; To a great extent (percent): 35.9; To a moderate extent (percent): 19.0; To a small extent (percent): 10.5; To no extent (percent): 2.6; No basis to judge/Not applicable (percent): 1.5; No answer (percent): 11.3; Number of respondents: 503. f. Link our product or service costs with the results we achieve. (Cost-benefit measures); To a very great extent (percent): 12.5; To a great extent (percent): 18.7; To a moderate extent (percent): 19.7; To a small extent (percent): 20.0; To no extent (percent): 11.7; No basis to judge/Not applicable (percent): 6.0; No answer (percent): 11.3; Number of respondents: 503. Q8. For those program(s)/operation(s)/project(s) that you are involved with, to what extent, if at all, do you use the information obtained from performance measurement when participating in the following activities? a. Setting program priorities; To a very great extent (percent): 13.0; To a great extent (percent): 38.7; To a moderate extent (percent): 21.7; To a small extent (percent): 10.8; To no extent (percent): 3.4; No basis to judge/Not applicable (percent): 1.2; No answer (percent): 11.0; Number of respondents: 503. b. Allocating resources; To a very great extent (percent): 14.9; To a great extent (percent): 36.6; To a moderate extent (percent): 20.3; To a small extent (percent): 10.8; To no extent (percent): 3.7; No basis to judge/Not applicable (percent): 1.8; No answer (percent): 12.0; Number of respondents: 503. c. Adopting new program approaches or changing work processes; To a very great extent (percent): 13.9; To a great extent (percent): 34.8; To a moderate extent (percent): 24.2; To a small extent (percent): 10.2; To no extent (percent): 4.1; No basis to judge/Not applicable (percent): 0.9; No answer (percent): 11.9; Number of respondents: 503. d. Coordinating program efforts with other internal or external organizations; To a very great extent (percent): 10.1; To a great extent (percent): 32.2; To a moderate extent (percent): 29.4; To a small extent (percent): 10.9; To no extent (percent): 3.6; No basis to judge/Not applicable (percent): 2.7; No answer (percent): 11.1; Number of respondents: 503. e. Refining program performance measures; To a very great extent (percent): 14.4; To a great extent (percent): 28.8; To a moderate extent (percent): 24.7; To a small extent (percent): 13.2; To no extent (percent): 3.4; No basis to judge/Not applicable (percent): 4.1; No answer (percent): 11.3; Number of respondents: 503. f. Setting new or revising existing performance goals; To a very great extent (percent): 16.1; To a great extent (percent): 32.4; To a moderate extent (percent): 21.0; To a small extent (percent): 11.8; To no extent (percent): 2.9; No basis to judge/Not applicable (percent): 3.5; No answer (percent): 12.3; Number of respondents: 503. g. Setting individual job expectations for the government employees I manage or supervise; To a very great extent (percent): 18.4; To a great extent (percent): 33.6; To a moderate extent (percent): 22.0; To a small extent (percent): 9.3; To no extent (percent): 2.8; No basis to judge/Not applicable (percent): 2.4; No answer (percent): 11.4; Number of respondents: 503. h. Rewarding government employees I manage or supervise; To a very great extent (percent): 17.7; To a great extent (percent): 33.3; To a moderate extent (percent): 21.8; To a small extent (percent): 8.2; To no extent (percent): 4.0; No basis to judge/Not applicable (percent): 3.0; No answer (percent): 11.9; Number of respondents: 503. i. Developing and managing contracts; To a very great extent (percent): 7.3; To a great extent (percent): 18.8; To a moderate extent (percent): 19.3; To a small extent (percent): 10.4; To no extent (percent): 8.1; No basis to judge/Not applicable (percent): 24.4; No answer (percent): 11.7; Number of respondents: 503. Q9. Based on your experience with the program(s)/operation(s)/ project(s) that you are involved with, to what extent, if at all, have the following factors hindered measuring performance or using the performance information? a. Setting program priorities; To a very great extent (percent): 14.2; To a great extent (percent): 20.3; To a moderate extent (percent): 32.1; To a small extent (percent): 19.7; To no extent (percent): 9.1; No basis to judge/Not applicable (percent): 3.9; No answer (percent): 0.7; Number of respondents: 503. b. Different parties are using different definitions to measure performance; To a very great extent (percent): 10.3; To a great extent (percent): 23.5; To a moderate extent (percent): 24.6; To a small extent (percent): 24.6; To no extent (percent): 10.4; No basis to judge/Not applicable (percent): 5.6; No answer (percent): 1.0; Number of respondents: 503. c. Difficulty obtaining valid or reliable data; To a very great extent (percent): 7.5; To a great extent (percent): 22.1; To a moderate extent (percent): 26.0; To a small extent (percent): 27.2; To no extent (percent): 12.0; No basis to judge/Not applicable (percent): 4.1; No answer (percent): 1.0; Number of respondents: 503. d. Difficulty obtaining data in time to be useful; To a very great extent (percent): 7.6; To a great extent (percent): 17.5; To a moderate extent (percent): 23.9; To a small extent (percent): 26.0; To no extent (percent): 18.3; No basis to judge/Not applicable (percent): 4.4; No answer (percent): 2.2; Number of respondents: 503. e. Lack of incentives (e.g., rewards, positive recognition); To a very great extent (percent): 12.9; To a great extent (percent): 15.8; To a moderate extent (percent): 25.0; To a small extent (percent): 24.1; To no extent (percent): 16.6; No basis to judge/Not applicable (percent): 4.8; No answer (percent): 0.8; Number of respondents: 503. f. Difficulty resolving conflicting interests of stakeholders, either internal or external; To a very great extent (percent): 9.5; To a great extent (percent): 19.3; To a moderate extent (percent): 25.5; To a small extent (percent): 22.0; To no extent (percent): 14.5; No basis to judge/Not applicable (percent): 8.1; No answer (percent): 1.0; Number of respondents: 503. g. Difficulty distinguishing between the results produced by the program and results caused by other factors; To a very great extent (percent): 7.4; To a great extent (percent): 18.7; To a moderate extent (percent): 24.3; To a small extent (percent): 28.6; To no extent (percent): 13.7; No basis to judge/Not applicable (percent): 6.3; No answer (percent): 1.0; Number of respondents: 503. h. Existing information technology and/or systems not capable of providing needed data; To a very great extent (percent): 9.7; To a great extent (percent): 19.4; To a moderate extent (percent): 20.6; To a small extent (percent): 26.6; To no extent (percent): 17.0; No basis to judge/Not applicable (percent): 5.4; No answer (percent): 1.3; Number of respondents: 503. i. Lack of staff who are knowledgeable about gathering and/or analyzing performance information; To a very great extent (percent): 10.1; To a great extent (percent): 17.5; To a moderate extent (percent): 23.4; To a small extent (percent): 26.2; To no extent (percent): 16.5; No basis to judge/Not applicable (percent): 5.0; No answer (percent): 1.3; Number of respondents: 503. j. Lack of ongoing top executive commitment or support for using performance information to make program/funding decisions; To a very great extent (percent): 9.7; To a great extent (percent): 16.0; To a moderate extent (percent): 18.4; To a small extent (percent): 24.0; To no extent (percent): 23.1; No basis to judge/Not applicable (percent): 7.9; No answer (percent): 1.0; Number of respondents: 503. k. Lack of ongoing Congressional commitment or support for using performance information to make program/funding decisions; To a very great extent (percent): 7.1; To a great extent (percent): 16.6; To a moderate extent (percent): 18.1; To a small extent (percent): 16.5; To no extent (percent): 17.1; No basis to judge/Not applicable (percent): 23.8; No answer (percent): 0.7; Number of respondents: 503. l. Difficulty determining how to use performance information to improve the program; To a very great extent (percent): 5.3; To a great extent (percent): 12.8; To a moderate extent (percent): 30.1; To a small extent (percent): 26.6; To no extent (percent): 18.8; No basis to judge/Not applicable (percent): 5.6; No answer (percent): 0.7; Number of respondents: 503. m. Concern that OMB will micromanage programs in my agency; To a very great extent (percent): 7.9; To a great extent (percent): 10.8; To a moderate extent (percent): 14.7; To a small extent (percent): 19.2; To no extent (percent): 26.6; No basis to judge/Not applicable (percent): 19.7; No answer (percent): 1.0; Number of respondents: 503. Q10. To what extent, if at all, do you agree with the following statements? (Check one box in each row.): a. Agency managers/supervisors at my level have the decision making authority they need to help the agency accomplish its strategic goals; To a very great extent (percent): 9.5; To a great extent (percent): 30.1; To a moderate extent (percent): 28.1; To a small extent (percent): 25.2; To no extent (percent): 5.6; No basis to judge/Not applicable (percent): 0.6; No answer (percent): 1.0; Number of respondents: 503. b. Agency managers/supervisors at my level are held accountable for agency accomplishment of its strategic goals; To a very great extent (percent): 14.6; To a great extent (percent): 42.9; To a moderate extent (percent): 24.1; To a small extent (percent): 12.7; To no extent (percent): 3.3; No basis to judge/Not applicable (percent): 1.5; No answer (percent): 1.0; Number of respondents: 503. c. Agency managers/supervisors at my level are held accountable for the results of the program(s)/operation(s)/project(s) they are responsible for; To a very great extent (percent): 24.2; To a great extent (percent): 46.5; To a moderate extent (percent): 17.2; To a small extent (percent): 7.2; To no extent (percent): 3.0; No basis to judge/Not applicable (percent): 0.9; No answer (percent): 1.0; Number of respondents: 503. d. Employees in my agency receive positive recognition for helping the agency accomplish its strategic goals; To a very great extent (percent): 9.7; To a great extent (percent): 27.1; To a moderate extent (percent): 31.2; To a small extent (percent): 22.0; To no extent (percent): 7.2; No basis to judge/Not applicable (percent): 1.5; No answer (percent): 1.3; Number of respondents: 503. e. My agency's top leadership demonstrates a strong commitment to achieving results; To a very great extent (percent): 24.4; To a great extent (percent): 37.1; To a moderate extent (percent): 21.8; To a small extent (percent): 9.9; To no extent (percent): 2.9; No basis to judge/Not applicable (percent): 2.6; No answer (percent): 1.3; Number of respondents: 503. f. My agency is investing the resources needed to ensure that its performance data is of sufficient quality; To a very great extent (percent): 9.5; To a great extent (percent): 21.4; To a moderate extent (percent): 29.3; To a small extent (percent): 21.9; To no extent (percent): 7.4; No basis to judge/Not applicable (percent): 9.3; No answer (percent): 1.0; Number of respondents: 503. Q11. To what extent, if at all, do you agree with the following statements? (Check one box in each row.) The following items focus on the program(s)/operation(s)/project(s) that you are responsible for. a. The individual I report to periodically reviews with me the results or outcomes of the program(s) operation(s)/project(s) that I am responsible for; To a very great extent (percent): 18.8; To a great extent (percent): 36.1; To a moderate extent (percent): 23.1; To a small extent (percent): 15.3; To no extent (percent): 5.3; No basis to judge/Not applicable (percent): 0.4; No answer (percent): 1.0; Number of respondents: 503. b. Funding decisions for the program(s)/operation(s)/project(s) I am responsible for are based on results or outcome-oriented performance information; To a very great extent (percent): 4.5; To a great extent (percent): 20.9; To a moderate extent (percent): 23.1; To a small extent (percent): 26.4; To no extent (percent): 16.5; No basis to judge/Not applicable (percent): 7.5; No answer (percent): 1.0; Number of respondents: 503. c. Staffing and personnel decisions for the program(s)/operation(s)/ project(s) I am responsible for are based on results or outcome- oriented performance information; To a very great extent (percent): 4.2; To a great extent (percent): 20.9; To a moderate extent (percent): 27.8; To a small extent (percent): 23.2; To no extent (percent): 19.0; No basis to judge/Not applicable (percent): 3.7; No answer (percent): 1.3; Number of respondents: 503. d. Changes by management above my level to the program(s)/operation(s)/ project(s) I am responsible for are based on results or outcome- oriented performance information; To a very great extent (percent): 2.6; To a great extent (percent): 20.5; To a moderate extent (percent): 25.2; To a small extent (percent): 26.5; To no extent (percent): 15.3; No basis to judge/Not applicable (percent): 8.2; No answer (percent): 1.6; Number of respondents: 503. e. It is easy to motivate employees to be more results-oriented in the program(s)/operation(s)/project(s) I am responsible for; To a very great extent (percent): 3.5; To a great extent (percent): 22.0; To a moderate extent (percent): 33.6; To a small extent (percent): 29.0; To no extent (percent): 7.8; No basis to judge/Not applicable (percent): 2.8; No answer (percent): 1.3; Number of respondents: 503. f. I have sufficient information on the validity of the performance data I use to make decisions; To a very great extent (percent): 4.7; To a great extent (percent): 32.6; To a moderate extent (percent): 30.2; To a small extent (percent): 20.3; To no extent (percent): 7.1; No basis to judge/Not applicable (percent): 3.2; No answer (percent): 1.8; Number of respondents: 503. Q12. During the past 3 years, has your agency provided, arranged, or paid for training that would help you to accomplish the following tasks? (Check one box in each row.): a. Conduct strategic planning; Yes (percent): 46.7; No (percent): 52.3; No answer (percent): 1.0; Number of respondents: 503. b. Set program performance goals; Yes (percent): 48.8; No (percent): 50.2; No answer (percent): 1.0; Number of respondents: 503. c. Develop program performance measures; Yes (percent): 42.9; No (percent): 55.5; No answer (percent): 1.6; Number of respondents: 503. d. Assess the quality of performance data; Yes (percent): 35.3; No (percent): 63.5; No answer (percent): 1.3; Number of respondents: 503. e. Use program performance information to make decisions; Yes (percent): 40.7; No (percent): 56.8; No answer (percent): 2.4; Number of respondents: 503. f. Link the performance of program(s)/operation(s)/project(s) to the achievement of agency strategic goals; Yes (percent): 40.8; No (percent): 57.0; No answer (percent): 2.2; Number of respondents: 503. g. Implement the requirements of the Government Performance and Results Act (GPRA or the Results Act); Yes (percent): 31.9; No (percent): 66.7; No answer (percent): 1.3; Number of respondents: 503. Q13. In your opinion, to what extent, if at all, do you believe you need training (or additional training) in order to help you to accomplish the following tasks? (Check one box in each row.): a. Conduct strategic planning; To a very great extent (percent): 8.2; To a great extent (percent): 18.3; To a moderate extent (percent): 27.8; To a small extent (percent): 30.8; To no extent (percent): 11.2; No basis to judge/Not applicable (percent): 2.8; No answer (percent): 0.9; Number of respondents: 503. b. Set program performance goals; To a very great extent (percent): 8.6; To a great extent (percent): 19.7; To a moderate extent (percent): 27.6; To a small extent (percent): 27.2; To no extent (percent): 12.7; No basis to judge/Not applicable (percent): 3.0; No answer (percent): 1.2; Number of respondents: 503. c. Develop program performance measures; To a very great extent (percent): 9.6; To a great extent (percent): 21.1; To a moderate extent (percent): 29.2; To a small extent (percent): 23.7; To no extent (percent): 11.2; No basis to judge/Not applicable (percent): 3.6; No answer (percent): 1.6; Number of respondents: 503. d. Assess the quality of performance data; To a very great extent (percent): 8.9; To a great extent (percent): 22.1; To a moderate extent (percent): 25.7; To a small extent (percent): 25.7; To no extent (percent): 13.3; No basis to judge/Not applicable (percent): 2.5; No answer (percent): 1.9; Number of respondents: 503. e. Use program performance information to make decisions; To a very great extent (percent): 9.5; To a great extent (percent): 22.1; To a moderate extent (percent): 23.7; To a small extent (percent): 26.9; To no extent (percent): 13.8; No basis to judge/Not applicable (percent): 2.2; No answer (percent): 1.9; Number of respondents: 503. f. Link the performance of program(s)/operation(s)/project(s) to the achievement of agency strategic goals; To a very great extent (percent): 12.0; To a great extent (percent): 21.6; To a moderate extent (percent): 24.2; To a small extent (percent): 27.3; To no extent (percent): 11.2; No basis to judge/Not applicable (percent): 2.8; No answer (percent): 0.9; Number of respondents: 503. g. Implement the requirements of the Government Performance and Results Act (GPRA or the Results Act); To a very great extent (percent): 12.0; To a great extent (percent): 25.0; To a moderate extent (percent): 26.4; To a small extent (percent): 17.6; To no extent (percent): 11.1; No basis to judge/Not applicable (percent): 6.9; No answer (percent): 1.0; Number of respondents: 503. Q14. What, in your opinion, can the Federal government do to improve its overall focus on managing for results? Writing comment (percent): 65.4; Number of respondents: 503. Q15. Prior to receiving this questionnaire , which of the following statements best describes your awareness of GPRA? I had never heard of GPRA. (percent): 19.5; I had heard of GPRA but had no knowledge of its requirements. (percent): 13.5; I had heard of GPRA and had a low level of knowledge of its requirements. (percent): 24.5; I had heard of GPRA and had moderate knowledge of its requirements. (percent): 35.4; I had heard of GPRA and had extensive knowledge of its requirements. (percent): 5.6; No answer (percent): 1.5; Number of respondents: 503. Q16. For those program(s)/operation(s)/project(s) that you are involved with, to what extent, if at all, do you consider the annual performance goals set forth in your agency's GPRA annual performance plan when participating in the following activities? a. Setting program priorities; To a very great extent (percent): 7.5; To a great extent (percent): 19.8; To a moderate extent (percent): 24.4; To a small extent (percent): 12.7; To no extent (percent): 13.5; No basis to judge/Not applicable (percent): 20.8; No answer (percent): 1.2; Number of respondents: 503. b. Allocating resources; To a very great extent (percent): 5.2; To a great extent (percent): 16.8; To a moderate extent (percent): 23.9; To a small extent (percent): 15.7; To no extent (percent): 14.7; No basis to judge/Not applicable (percent): 22.5; No answer (percent): 1.2; Number of respondents: 503. c. Adopting new program approaches or changing work processes; To a very great extent (percent): 5.9; To a great extent (percent): 21.3; To a moderate extent (percent): 23.8; To a small extent (percent): 13.5; To no extent (percent): 13.3; No basis to judge/Not applicable (percent): 21.1; No answer (percent): 1.2; Number of respondents: 503. d. Coordinating program efforts with other internal or external organizations; To a very great extent (percent): 5.4; To a great extent (percent): 16.2; To a moderate extent (percent): 26.4; To a small extent (percent): 15.7; To no extent (percent): 12.8; No basis to judge/Not applicable (percent): 22.3; No answer (percent): 1.2; Number of respondents: 503. e. Developing or refining program performance measures; To a very great extent (percent): 5.7; To a great extent (percent): 18.1; To a moderate extent (percent): 21.5; To a small extent (percent): 15.9; To no extent (percent): 14.7; No basis to judge/Not applicable (percent): 22.9; No answer (percent): 1.2; Number of respondents: 503. f. Setting individual job expectations for the government employees I manage or supervise; To a very great extent (percent): 5.7; To a great extent (percent): 22.0; To a moderate extent (percent): 19.7; To a small extent (percent): 15.4; To no extent (percent): 15.0; No basis to judge/Not applicable (percent): 20.5; No answer (percent): 1.6; Number of respondents: 503. g. Rewarding government employees I manage or supervise; To a very great extent (percent): 5.9; To a great extent (percent): 19.5; To a moderate extent (percent): 20.4; To a small extent (percent): 16.1; To no extent (percent): 16.2; No basis to judge/Not applicable (percent): 20.5; No answer (percent): 1.5; Number of respondents: 503. h. Developing and managing contracts; To a very great extent (percent): 3.3; To a great extent (percent): 12.3; To a moderate extent (percent): 14.9; To a small extent (percent): 13.3; To no extent (percent): 16.4; No basis to judge/Not applicable (percent): 37.1; No answer (percent): 2.7; Number of respondents: 503. Q17. During the past 3 years, have you been involved in these activities? (Check one box in each row.) GPRA-related activities: a. Developing ways to measure whether program performance goals are being achieved; Yes (percent): 45.8; No (percent): 52.7; No answer (percent): 1.5; Number of respondents: 503. b. Gathering and analyzing data to measure whether programs are meeting their specific performance goals; Yes (percent): 50.6; No (percent): 47.9; No answer (percent): 1.5; Number of respondents: 503. c. Using measures for program performance goals to determine if the agency's strategic goals are being achieved; Yes (percent): 42.6; No (percent): 55.9; No answer (percent): 1.5; Number of respondents: 503. d. Assessing the quality of data used in measuring performance; Yes (percent): 39.7; No (percent): 58.2; No answer (percent): 2.1; Number of respondents: 503. Q18. To what extent, if at all, do you agree with the following statements? (Check one box in each row.) Extent I agree with the following statements: a. The objectives of my program(s)/operation(s)/project(s) are in alignment with my agency's strategic plan under GPRA; To a very great extent (percent): 10.1; To a great extent (percent): 29.4; To a moderate extent (percent): 15.8; To a small extent (percent): 7.2; To no extent (percent): 1.2; No basis to judge/Not applicable (percent): 35.0; No answer (percent): 1.2; Number of respondents: 503. b. The costs associated with implementing GPRA have taken time or funds away from other important activities or projects; To a very great extent (percent): 4.8; To a great extent (percent): 7.7; To a moderate extent (percent): 13.1; To a small extent (percent): 15.1; To no extent (percent): 10.6; No basis to judge/Not applicable (percent): 46.1; No answer (percent): 2.5; Number of respondents: 503. c. The benefits to my agency that are achieved by implementing GPRA are worth the costs incurred in doing so (e.g., in time, money, and effort); To a very great extent (percent): 3.8; To a great extent (percent): 11.3; To a moderate extent (percent): 15.8; To a small extent (percent): 12.4; To no extent (percent): 7.4; No basis to judge/Not applicable (percent): 48.0; No answer (percent): 1.3; Number of respondents: 503. d. GPRA strategic and annual performance plans are mostly a repackaging of goals, measures, and objectives that were already being used within my agency; To a very great extent (percent): 5.7; To a great extent (percent): 22.7; To a moderate extent (percent): 19.1; To a small extent (percent): 7.7; To no extent (percent): 2.8; No basis to judge/Not applicable (percent): 40.1; No answer (percent): 1.9; Number of respondents: 503. e. Managerial effectiveness is impeded by the lack of integration between GPRA and other federal management programs; To a very great extent (percent): 3.8; To a great extent (percent): 7.7; To a moderate extent (percent): 16.3; To a small extent (percent): 13.4; To no extent (percent): 6.8; No basis to judge/Not applicable (percent): 50.5; No answer (percent): 1.5; Number of respondents: 503. f. My agency considers contributions to and comments on GPRA plans or reports from managers/supervisors at my level: ; To a very great extent (percent): 3.5; To a great extent (percent): 10.6; To a moderate extent (percent): 15.6; To a small extent (percent): 15.0; To no extent (percent): 11.7; No basis to judge/Not applicable (percent): 42.1; No answer (percent): 1.5; Number of respondents: 503. g. Agency managers/supervisors at my level use GPRA annual performance plans to manage their program(s)/operation(s)/project(s); To a very great extent (percent): 3.6; To a great extent (percent): 7.8; To a moderate extent (percent): 18.5; To a small extent (percent): 16.4; To no extent (percent): 17.7; No basis to judge/Not applicable (percent): 33.9; No answer (percent): 2.1; Number of respondents: 503. h. GPRA's planning and reporting requirements impose a significant paperwork burden; To a very great extent (percent): 2.9; To a great extent (percent): 7.4; To a moderate extent (percent): 17.1; To a small extent (percent): 15.9; To no extent (percent): 7.3; No basis to judge/Not applicable (percent): 47.7; No answer (percent): 1.8; Number of respondents: 503. i. GPRA has caused agency managers/supervisors at my level to place a greater emphasis on getting input from appropriate stakeholders on their interests and expectations; To a very great extent (percent): 2.5; To a great extent (percent): 7.0; To a moderate extent (percent): 20.8; To a small extent (percent): 13.1; To no extent (percent): 12.2; No basis to judge/Not applicable (percent): 42.2; No answer (percent): 2.1; Number of respondents: 503. Q19. To what extent, if at all, do you believe that GPRA has improved your agency's ability to deliver results to the American public. To a very great extent (Continue with question 19a. ) (percent): 1.6; To a great extent (Continue with question 19a. ) (percent): 6.9; To a moderate extent (Continue with question 19a. ) (percent): 14.5; To a small extent (percent): 25.2; To no extent (percent): 11.6; No basis to judge/Not applicable (percent): 38.0; No answer (percent): 2.1; Number of respondents: 503. Q19a. Please briefly describe how GPRA has improved your agency's ability to deliver results to the American public. Writing comment (percent): 23.0; Number of respondents: 139. Q20. To what extent, if at all, do you believe the following persons or entities pay attention to your agency's efforts under GPRA? (Check one box in each row.): a. Department Secretary (if applicable); To a very great extent (percent): 10.3; To a great extent (percent): 18.3; To a moderate extent (percent): 10.6; To a small extent (percent): 6.2; To no extent (percent): 3.6; No basis to judge/Not applicable (percent): 47.1; No answer (percent): 3.9; Number of respondents: 503. b. Agency head other than Department Secretary (if applicable); To a very great extent (percent): 11.4; To a great extent (percent): 20.6; To a moderate extent (percent): 12.9; To a small extent (percent): 6.9; To no extent (percent): 2.1; No basis to judge/Not applicable (percent): 42.9; No answer (percent): 3.2; Number of respondents: 503. c. The individual I report to; To a very great extent (percent): 8.5; To a great extent (percent): 16.0; To a moderate extent (percent): 19.3; To a small extent (percent): 14.0; To no extent (percent): 13.2; No basis to judge/Not applicable (percent): 26.6; No answer (percent): 2.4; Number of respondents: 503. d. Managers and supervisors at my level; To a very great extent (percent): 5.1; To a great extent (percent): 13.2; To a moderate extent (percent): 19.7; To a small extent (percent): 19.5; To no extent (percent): 16.0; No basis to judge/Not applicable (percent): 24.1; No answer (percent): 2.4; Number of respondents: 503. e. Employees who report to me; To a very great extent (percent): 2.5; To a great extent (percent): 8.3; To a moderate extent (percent): 13.8; To a small extent (percent): 18.1; To no extent (percent): 30.7; No basis to judge/Not applicable (percent): 24.0; No answer (percent): 2.7; Number of respondents: 503. f. Office of Management and Budget; To a very great extent (percent): 14.1; To a great extent (percent): 16.7; To a moderate extent (percent): 11.4; To a small extent (percent): 4.0; To no extent (percent): 1.8; No basis to judge/Not applicable (percent): 49.0; No answer (percent): 2.9; Number of respondents: 503. g. Congressional committees; To a very great extent (percent): 7.8; To a great extent (percent): 14.5; To a moderate extent (percent): 13.0; To a small extent (percent): 8.2; To no extent (percent): 3.5; No basis to judge/Not applicable (percent): 50.3; No answer (percent): 2.6; Number of respondents: 503. h. The audit community (e.g., GAO, Inspectors General); To a very great extent (percent): 11.7; To a great extent (percent): 16.9; To a moderate extent (percent): 11.5; To a small extent (percent): 7.3; To no extent (percent): 1.2; No basis to judge/Not applicable (percent): 48.2; No answer (percent): 3.2; Number of respondents: 503. i. The general public; To a very great extent (percent): 1.5; To a great extent (percent): 2.5; To a moderate extent (percent): 6.1; To a small extent (percent): 17.5; To no extent (percent): 26.4; No basis to judge/Not applicable (percent): 42.2; No answer (percent): 3.9; Number of respondents: 503. Q21. To what extent, if at all, do you believe that efforts to implement GPRA to date have improved the program(s)/operation(s)/ project(s) in which you are involved? I have not been sufficiently involved in GPRA to have an opinion. (percent): 48.7; To a very great extent (percent): 2.9; To a great extent (percent): 5.2; To a moderate extent (percent): 14.7; To a small extent (percent): 16.2; To no extent (percent): 10.0; No answer (percent): 2.4; Number of respondents: 503. Q22. To what extent, if at all, do you believe that efforts to implement GPRA to date have improved your agency's programs/operations/ projects? I have not been sufficiently involved in GPRA to have an opinion. (percent): 47.9; To a very great extent (percent): 2.3; To a great extent (percent): 6.2; To a moderate extent (percent): 17.0; To a small extent (percent): 16.3; To no extent (percent): 8.7; No answer (percent): 1.5; Number of respondents: 503. Q23. To what extent, if at all, do you believe implementing GPRA can improve your agency's programs/operations/projects in the future? To a very great extent (percent): 3.4; To a great extent (percent): 11.9; To a moderate extent (percent): 23.7; To a small extent (percent): 16.7; To no extent (percent): 5.6; No basis to judge (percent): 35.7; No answer (percent): 3.0; Number of respondents: 503. Q24. If you have been involved to any extent in implementing GPRA for the program(s)/operation(s)/project(s) you are involved with, what has been your greatest difficulty, and in what ways, if any, do you think this difficulty could be addressed? Writing comment (percent): 26.0; Number of respondents: 503. Q25. If you have additional comments regarding any previous question or any comments/suggestions concerning GPRA, please use the space provided below. Writing comment (percent): 16.5; Number of respondents: 503. Note: Percents reported are weighted percents based on the population size. Unweighted N reported for each item. [End of table] [End of section] Appendix VII: Agencies Subject to the Chief Financial Officers Act: The Chief Financial Officers Act of 1990 (the CFO Act) created the position of Chief Financial Officer in each executive department and in each major executive agency in the federal government. The agencies covered by the CFO Act are: 1. Agency for International Development: 2. Department of Agriculture: 3. Department of Commerce: 4. Department of Defense: 5. Department of Education: 6. Department of Energy: 7. Department of Health and Human Services: 8. Department of Housing and Urban Development: 9. Department of the Interior: 10. Department of Justice: 11. Department of Labor: 12. Department of State: 13. Department of Transportation: 14. Department of the Treasury: 15. Department of Veterans Affairs: 16. Environmental Protection Agency: 17. Federal Emergency Management Agency[Footnote 141] 18. General Services Administration: 19. National Aeronautics and Space Administration: 20. National Science Foundation: 21. Nuclear Regulatory Commission: 22. Office of Personnel Management: 23. Small Business Administration: 24. Social Security Administration[Footnote 142] [End of section] Appendix VIII: Comments from the Office of Management and Budget: EXECUTIVE OFFICE OF THE PRESIDENT: OFFICE OF MANAGEMENT AND BUDGET WASHINGTON, D.C. 20503: DEPUTY DIRECTOR FOR MANAGEMENT: February 2, 2004: Ms. Pat Dalton: Director: Strategic Issues: General Accounting Office: 441 G Street, NW: Washington, DC 20548: Dear Ms. Dalton: Thank you for the opportunity to comment on the draft GAO report on GPRA implementation (Results-Oriented Government: GPRA has Established a Solid Foundation for Achieving Greater Results, GAO-04-38). We appreciate GAO's extensive review of the Executive Branch's implementation of the Government Performance and Results Act (GPRA). We agree that GPRA provides a foundation for making the Federal government more results-oriented. However, too often, neither Congress nor the Executive Branch was using GPRA performance information to manage or oversee programs. As you know, we are giving enhanced attention to agency and program results and identifying the corrective action necessary to overcome any shortcomings in program management or performance. Our goal should be to implement those corrective actions, which will help us realize the results-orientation envisioned by GPRA. Your recommendations will enhance our efforts to make the government more results-oriented. For instance, we will continue to use the Program Assessment Rating Tool (PART) to improve agency performance measurement practices and share those practices across government. And we will revise our guidance, both for GPRA and the PART, to clarify the integrated and complementary relationship between the two initiatives. We will also work with agencies to ensure they are provided adequate training in performance management, a key objective of the PMA. With respect to your recommendation to produce a government-wide strategic and performance plan, we believe that the President's annual Budget represents the Executive Branch's government-wide strategic and performance plan. The President's Budget provides a strategic cross-cutting view of the President's priorities, and the Budget is providing increasing emphasis on performance. The Budget includes a discussion of agency performance against goals, as well as agency status and progress on the President's Management Agenda. As the Administration works to reach its goal of applying the PART to all programs administered by the Federal government, we will be able to include in the President's Budget outcome-oriented goals for all the government's programs. Thank you for the opportunity to review and comment on your draft report. I appreciate your willingness to take our oral and written comments into consideration in the final draft. I look forward to working with you to improve the ways in which we are creating a results-oriented government. Sincerely, Signed by: Clay Johnson III: [End of section] Appendix IX: Comments from the Department of Energy: Department of Energy: Washington, DC 20585: January 15, 2004: Ms. Patricia A. Dalton, Director: Strategic Issues: U.S. General Accounting Office: Washington, DC 20548: Dear Ms. Dalton: We are forwarding our comments on the draft audit entitled, "Results - Oriented Government, GPRA has Established a Solid Foundation for Achieving Greater Results, GAO-04-38, January 2004." In this report, GAO reviewed the draft strategic plans, annual performance plans, and annual performance reports for the Department of Energy (DOE) and five other agencies. The DOE took a different approach to preparing these documents than some other agencies. Below we have demonstrated how we complied with GPRA guidelines in preparing the required documents, yet in some cases GAO did not give us the proper credit. In regard to Annual Performance Plans (APPS), GAO compared agencies' APPs for 2004 with 1999. The report (pages 8-9) indicates that DOE was the only agency of the six reviewed that did not: * Maintain or improve our efforts to provide a clear picture of intended performance. * Specify the strategies and resources we intended to use to achieve our performance goals. * Report on how we would ensure performance data will be credible. DOE disagrees with all three comments. * In the introduction section of the FY 2004 APP, we provide a clear picture of intended performance. Each of the 17 General Goals contains a description of the program. A good example of providing a clear picture of performance can be found on page 12 of our FY 2004 APP in General Goal 1 where each activity (hydrogen, wind, hydropower, solar, etc.) is separately addressed. * Each General Goal contains a paragraph on the means and strategies that will be used to achieve our goals. In addition, funding is included at the General Goal level. * The introduction contains a section on "Criteria for Assessing our Results." Under the Validation and Verification section on page 4 of our FY 2004 APP, we discuss the Joule software package used for documenting and tracking performance data. On page 1 I of the report, GAO states that "The clarity of Energy's plan remained limited because its annual goals were not clearly linked to its mission, the long-term goals in its strategic plan, or the program activities in its budget request." This is not a accurate statement. As was explained to the GAO team that visited with DOE, when the FY 2004 APP was prepared, the draft strategic plan contained 17 General Goals and matched the published version of the FY 2004 APP. As a result of coordinating our draft strategic plan within DOE and with the Office of Management and Budget, the draft strategic plan used for this report had been reduced to seven General Goals. Had the draft of the strategic plan in its place at the time the FY 2004 APP was prepared been used, there would have been a one-to-one correlation between the goals in the APP and the strategic plan. The portion of the draft GAO report dedicated to Annual Performance Reports for FY 2002 (i.e., Performance and Accountability Report (PAR)) states on page 19, that six agencies (including DOE) did not consistently report the reasons for not meeting their goals. In the DOE FY 2002 PAR, each target that was not met provided a plan of action for addressing the causes of not meeting the target. An example can be found on page 32 of the FY 2002 PAR, under SC3-1, where an action plan is provided for a target that was not met. Also in reference to the DOE FY 2002 PAR, page 105 of the draft GAO report indicates that DOE did not provide a discussion of the relationship between the strategic plan, performance plan, and the performance report. However, page 80 of the FY 2002 PAR under the introductory section of the Detailed Performance Results contains a paragraph dedicated to providing that linkage. We are hopeful that our comments will lead to a fair and accurate presentation of DOE in your final report. Our staff is available to meet with you to discuss these points in further detail. In order to arrange such a meeting, please contact Van Jones on 202-586-4050. Signed by: James G. Powers: Director, Office of Program Assessment and Evaluation/OMBE: The following are our comments on DOE's letter dated January 15, 2004. GAO Comments: The Director of the Office of Program Assessment and Evaluation forwarded written comments from DOE on a draft of this report. DOE disagreed with several of our conclusions concerning its 2004 Annual Performance Plan and 2002 Performance and Accountability Report. We incorporated the additional information and perspectives of DOE into our report as appropriate. 1. We stated that, when compared to the 1999 Annual Performance Plan, DOE's 2004 Annual Performance Plan continued to provide a limited picture of intended performance. DOE disagreed, stating that the plan provided a clear picture of intended performance because a description of each program was contained in each of the general goals. We agree that a description of programs is provided for each goal, and we also stated in our draft report that improvement was made in developing results-oriented performance measures that pertained specifically to fiscal year 2004. However, in our view, describing each program does not sufficiently explain DOE's expectations for intended performance. More specifically, we found the overall picture of intended performance was limited because DOE did not specifically describe how it coordinates with other agencies to accomplish crosscutting programs and did not provide a clear link between its annual goals and its mission and strategic goals. In our draft report we acknowledged that a link did not exist between the performance plan and the strategic plan because the strategic plan was revised after the performance plan was finalized. Nevertheless, DOE did not revise the final 2004 performance plan to reflect its changes in strategic goals. The lack of alignment between the performance plan and strategic plan goals limits the usefulness of the performance plan to support managers and staff in their day-to-day activities in achieving DOE's long-term strategic goals. 2. In response to our observation that DOE provided a general discussion of the strategies and resources needed to achieve its performance goals, DOE stated that its 2004 annual performance plan provided specific strategies and resources that will be used to achieve performance goals. DOE also noted that funding was included at the general goal level. We agree that funding was included at the general strategic goal level. However, better plans relate resources to the achievement of performance goals. DOE did not provide resource information at the performance goal level. Furthermore, while DOE discussed external factors that could affect its ability to achieve its performance goals at a high level, it did not discuss any specific strategies to mitigate those factors. 3. DOE disagreed with our characterization that its 2004 annual performance plan provided limited confidence that performance data will be credible. The department stated that the introduction section of its plan contained specific sections on assessing results and validating and verifying data, as well as discussed a software package used to document and track performance. In our draft report, we stated that DOE's plan showed some improvement over its 1999 plan by describing credible procedures to verify and validate performance information and by mentioning specific program evaluations for each goal. We also noted that DOE acquired new commercial software for performance tracking through remote data entry, monitoring, and oversight by program offices and managers. However, we concluded that DOE's reporting of credible performance data was limited because its plan does not specifically identify data limitations overall or for each of its goals. As we stated in our report, we found this to be of particular concern because, as we mentioned in our 2003 performance and accountability series, DOE has several management challenges where data quality is a concern. 4. Concerning its 2002 Annual Performance and Accountability Report, DOE stated that it provided a plan of action for addressing the causes of targets that were not met. We agree, and in our draft report we state that all targets that were not met or had mixed results contained a plan of action to achieve the target in the future. We also found that the majority of DOE targets that were not met or had mixed results contained clear explanations. We revised our text in the final version of this report to make these findings more evident. 5. Finally, DOE disagreed with our finding that it did not provide a discussion of the relationship between the strategic plan, performance plan, and the performance report in its 2002 Performance and Accountability Report. DOE stated that the introductory section of the report contains a paragraph that discusses the linkages between these three reports. However, although the performance and accountability report links these documents by organizing its results section according to strategic goals, associated program performance goals and targets, it did not succinctly demonstrate how the results relate to the annual and long-term strategic goals. We modified the draft accordingly to clarify this point. [End of section] Appendix X: Comments from the Department of Housing and Urban Development: DEPARTMENT OF HOUSING AND URBAN DEVELOPMENT WASHINGTON, DC 20410-3000: OFFICE OF THE CHIEF FINANCIAL OFFICER: JAN 9 2003: Ms. Patricia A. Dalton: Director: Strategic Issues: U.S. General Accounting Office: 451 G Street NW: Washington, DC 20548: The Department of Housing and Urban Development (HUD) appreciates the recognition of the substantial progress HUD has made in meeting the Government Performance and Results Act requirements in your draft report entitled "Results-Oriented Government: GPRA has Established a Solid Foundation for Achieving Greater Results." All of the areas that GAO has suggested for further improvement are in process or under consideration at HUD. Per your request, we are providing the enclosed comments for your consideration in clarifying or correcting the draft report. Please contact James Martin at (202) 708-0638 ext. 3706 or Kenneth M. Leventhal (202) 708-0614 ext. 6849 if you have any questions. Sincerely yours, Signed by: De W. Ritchie: Deputy Chief Financial Officer: Enclosure: ENCLOSURE: Comments for GAO: Strategic Plan: Comment 1: In the last paragraph on page 4 of the draft report, we question why HUD is omitted from the list of agencies whose strategic plans contained long-term goals that: demonstrated improvements in the quality of their 1997 goals; covered their missions; were results- oriented; and were expressed in a manner that could be used to gauge future success. HUD's March 2003 Strategic Plan addresses all these issues. The omission of HUD from this listing appears inconsistent with the HUD progress recognized by GAO in Table 1 on page 3 of the draft report. In addition, HUD does not believe that the improvements in both long- term numerical and intermediate goals are fully reflected. We would first comment that the nature of a Strategic Plan has certain limitations as to the level of details and that the linkages flow in large part from the corresponding development of the Annual Performance Plan. Furthermore, we do not agree with the assessment on page 41 that the link between long-term and intermediate goals is difficult to discern. For example, in the last paragraph of page 41, the general point is not apparent. The statement that "long-term performance measures are not consistently linked to a corresponding intermediate measure" seems to miss the point that both long-term and intermediate measures are aligned with strategic goals. The need for a direct linkage with each other is not apparent. Second, the example given seems to be characterized as a problem simply because the baseline has not yet been determined. It is not clear how this fact illustrates the general point. Finally, the statement about the number of respondents to be surveyed seems out of place; the appropriate consideration for a Strategic Plan would seem to be whether a statistically reliable measure currently exists or is planned. HUD suggests that extraneous material be edited out of the paragraph to clarify the point, and that GAO further consider the following points. * HUD worked diligently to improve the quality of its long-term strategic goals and substantially increased the number of long-term goals and advanced the quality by committing to many numerical goals that can be reviewed, as was requested during Congressional consultation. The provision of numerical goals makes HUD's goals particularly meaningful since they can then be measured. In addition, many of the long-term performance measures reflect a maturing of HUD's efforts and ability to measure performance accurately. As an example, the PHAS and SEMAP systems have now been in place for sufficient time to provide over the course of the strategic plan accurate measurement of progress in our Public Housing and Section 8 programs which represent close to 70 percent of the Department's overall resources. * HUD also feels strongly that the long-term performance measurements under Strategic Goal: Embrace High Standards of Ethics, Management and Accountability reflect substantial progress in capturing the major management challenges facing the Department as identified by GAO and IG reviews. Comment 2: On page 40, HUD is given credit for explaining how it will coordinate with other agencies but credit should also be recognized on page 7. Paragraph 3, states that none of the agencies provided any details regarding crosscutting activities. Following are citations from HUD's Strategic Plan, for GAO's consideration: * On page 7. HUD cites its intention to "Work with the Department of Treasury to enact a tax credit for developers of affordable single- family housing." * On page 17, under "Increase-housing opportunities for persons with disabilities HUD cites "...a pilot program in 11 states called Project Access, a joint effort between HUD and the Department of HHS designed to ease the transition of non-elderly persons with disabilities from nursing homes into community living and also cites...." * "Signing a Memorandum of Understanding with the Department of Justice and IRS to ensure that multifamily properties within the Low-Income Housing Tax Credit Program, the number one producer of low-income housing, comply with federal accessibility standards." * On page 26, HUD details the interagency approach in fighting homelessness reflected in the reauthorized Interagency Council on the Homeless and the coordination between the Secretaries of HUD, HHS, and Veterans Affairs. * On page 34, HUD cites the cooperative efforts with the Department of Justice to enforce the Fair Housing Act and with the Departments of Treasury and Justice to ensure the Low-Income Housing Tax Credit projects are in compliance with the Fair Housing Act. * On page 50 the Plan also refers to the conducting of interagency events and conferences designed to educate and train faith-based and community organizations on equal access issues and partnership opportunities. * Finally, the Department has a more extensive discussion of Interagency Partnerships on page 53, which highlights 6 additional interagency efforts. Comment 3: On page 42, GAO states that, "HUD's current strategic plan does not describe how program evaluations were used to develop its strategic goals or other components of its plan, and does not include a schedule for future evaluations." On page 3, HUD requests that in Table 1 and in the discussion on page 3 GAO will recognize HUD's fulfillment of the Evaluations element in our Strategic Plan. For example, the closing cost research discussed on page 9 is the only feasible way to address the issue of home purchase transaction costs. The research directly supports the long-term performance measure and the strategic objective, and the discussion specifies a timetable for replicating the evaluation for long-term performance assessment. Similarly, the "awareness of fair housing laws" discussion (page 35, as well as 33-34) shows how research provided a baseline and was used to establish a goal, and also provides a timetable for replicating the study. Numerous similar examples appear. HUD was not recognized for the Plan's extensive discussion of policy development and program evaluation supporting strategic goals (pages 57-60). For example, the "increasing minority homeownership" and "Affordable housing goals for GSEs" discussions on page 57 explain how research helped establish the goals and in turn how research is being shaped to address policy issues that are critical to goal achievement. Program evaluation also is mentioned throughout the document, and plays a central role in the establishment of performance goals; a number of the long-term performance measures and intermediate measures explicitly rely on research efforts to supplement the performance metrics available from Administrative data for the reasons discussed above, HUD requests that GAO modify Table 111. 4 and the discussion on page 39 to recognize the inclusion of evaluations material in HUD's Strategic Plan. The plan explains how evaluations were used to establish and measure goals in the Strategic Plan, and identifies timetables for several evaluations. In light of this material, it seems appropriate to change the "No" in Table 111. 4 to "Yes." HUD believes that the second paragraph of page 42 is factually incorrect for the reasons discussed above. It would seem appropriate for GAO to recognize the ways in which HUD's Strategic Plan fulfills the GPRA evaluations requirement and to identify ways in which it falls short of GAO's criteria for a successful presentation of evaluation information. Comment 4: On page 42 the draft rises the issue of how HUD will "discern between those jobs created by CDBG and those created by other means", this question is more appropriate to a data review or data audit and seems unnecessary for this report. Annual Performance Plans: Comment 1: HUD included a detailed Means and Strategy discussion for each of our Strategic Goals and we believe we should be added to the DOE citation on pages 12-13. Comment 2: On page 75, HUD's goal is misstated: it should read, "Promote decent affordable housing." A strategic objective is misstated on the same page: "Increase minority homeownership." Performance and Accountability Report: Comment 1: On page 13, for the FY 2002 PAR, HUD for the first time provided a table akin to the SSA table with results shown in the front of each Strategic Goal Section. The table/shows whether HUD "substantially met" (a check mark) or did not meet (no check mark) and/ or provided extensive footnotes where necessary to clarify the result. In addition, HUD had approximately 190 data elements to report on compared to the 69 cited for SSA. We recognize that the number of indicators required by GPRA guidance regarding full coverage of program and activities does require a thorough effort to review HUD's inclusive performance results. Thus, the multi-pronged nature of HUD's mission and programs to some extent precludes simple summaries of overall performance. HUD believes that the summary tables for each strategic goal provided the best and most feasible balance of simplicity, transparency and completeness for the Department's strategic framework and programmatic structure. Comment 2: On page 19, the report says that HUD, SBA and DOT did not discuss plans or strategies to achieve unmet goals in the future and that HUD and several other agencies did not report the reasons for not meeting their goals. HUD believes that we consistently tried to report on reasons for not meeting goals as well as discussing strategies to achieve unmeet goals in the future. We would concede that we may not have done this in every case and that is in part because it is often difficult to assess why some goals were not met and therefore what actions might improve performance. As an example of our consistent effort, on the very first goal and indicator in which we did not meet a goal, pages 2-13, indicator 1. 1.f., the first paragraph under "Results" gives a reason for missing the target and the following paragraph describes our planned efforts to improve results. This same approach is on the very next indicator on pages 2-14, indicator 1.1.13 and is replete throughout the report. Comment 3: Regarding the middle paragraph on page 108 of the draft report, HUD requests that GAO consider: 1) the impacts of acceleration of the performance report issuance on an agency's ability to include an evaluation of the new fiscal year's performance plan relative to the performance attained in the just completed fiscal year, and 2) reasonable alternative actions to fulfill this requirement. At HUD, the issuance of the annual performance report was initially completed in six months or 180 days after the end of the fiscal year and this has been accelerated to 120 days last year, 80 days this year and will be in only 45 days next year. As a reasonable alternative, HUD generally considers prior year performance in conjunction with updating the annual performance plan for each upcoming year. The intended point of the last paragraph on page 108 needs to be clarified for HUD and other users of this report. Whereas the previous paragraph contends that HUD's FY 2002 performance report did not include an evaluation of the FY 2003 performance plan relative to performance attained in FY 2002, the paragraph in question cites two examples where HUD's FY 2002 performance report reflects an adjustment of FY 2003 plans based on an assessment of prior year performance. If the examples are to be retained in the final report, HUD requests that they be put in proper context because the current wording incorrectly implies that HUD no longer has an interest in establishing and measuring goals for homeownership or housing quality. HUD did eliminate the two indicators in question because they were too broad and beyond the direct scope or influence of HUD's programs. However, the draft report fails to indicate that HUD's performance plans and reports retained other indicators that capture the direct impact of HUD programs on homeownership and housing quality. Comments 4: On page 110, we disagree with the chart's depiction that 35 percent of HUD's performance targets were "undetermined" and that 15 percent had "no explanation." While it is true that detailed explanations of missed goals were not provided in the summary charts at the beginning of each strategic goal section, many of the missed targets had explanations in the individual indicator write-ups that followed the summaries. We request that GAO reassess the basis for the percentages and labels reflected in this chart to accurately reflect the content of HUD's performance report. We are concerned that GAO has not recognized the complexity in reporting whether indicator performance goals have been substantially met or missed. The estimate of 47 percent is of concern since HUD has not done that measurement and there can be very legitimate differences as to how to score indicators where "results are too complex to summarize" or data is unavailable. We are also concerned how you determined that 2 percent of data is "not reliable." More specifically, on page 110, the footnote to Figure V.4 appears to mischaracterize the purpose of the footnotes in HUD's performance report cards. The footnotes are intended to briefly and transparently disclose attributes of the performance data in the report cards that could mislead a reader who does not read the full discussion of each performance indicator. In a number of instances, the explanations appear along with a statement of whether targets were missed. HUD requests that GAO revise the statement "Rather than stating if some performance targets were met or not, HUD provided the following explanations" to capture our efforts to better inform the reader. In Figure V.4, it is unclear why HUD indicators are characterized as "undetermined," "no explanation," and "not reliable," as GAO has not defined these terms or their criteria. The preponderance of "undetermined" and "no explanation" designations is especially puzzling and suggests that GAO may have misinterpreted the performance report cards. HUD requests that GAO review these indicators for recategorization as "data not available" (as denoted by footnote "a" or "goal not achieved." No indicator should be classified as "undetermined" simply because HUD described it with a footnote. Agencies also appear to be treated inconsistently in the use of categories and the way indicators are assigned to categories. For example, a number of HUD indicators did not have a performance target for fiscal year 2002, usually because data sources were not scheduled to be available. These indicators were denoted by footnote "b" in HUD's performance report card. GAO characterized these indicators as "data not expected" for the Department of Education (Figure V.1), but has not used this category for HUD. HUD requests that GAO review all indicators with footnote "b" for recategorization as "data not expected." Comment 5: On page 93, first paragraph, the first sentence seems to be confusing the annual performance "plan" with the annual performance "report," as it is unclear what the report could do to make the plan more credible. Comment 6: On page 107, second paragraph, it appears that "fiscal" should be changed to "fiscal year." The following are our comments on HUD's letter dated January 9, 2004. GAO Comments: HUD provided written comments and disagreed with several of our observations, which we address below. HUD also mentioned that all of the areas we suggested for further improvement were already in process or being considered. Where appropriate, we incorporated HUD's comments and perspectives to clarify our report. 1. HUD did not agree with our observation that the link between long- term and intermediate goals in its strategic plan is difficult to discern. The department mentioned that the need for a direct link between long-term and intermediate goals is not apparent, as they are aligned with strategic goals. GPRA requires that an agency's strategic plan contain, among other things, a description of the relationship between the long-term goals and objectives and the annual performance goals. In addition, OMB's June 2002 Circular A-11 states that the strategic plan should briefly outline how annual performance goals relate to the long-term, general goals, and how they help determine the achievement of the general goals. Federal agencies can help readers understand how they move from general goals to specific, measurable outcomes by discussing how they plan to measure progress in achieving the long term-goals in their strategic plan. For example, for its strategic goal of "Increase Homeownership Opportunities," HUD mentions that one of its long-term performance measures is to combat predatory lending. Readers can review the intermediate measures listed under that goal to get a sense of how HUD plans to accomplish this objective. For example, HUD mentions that beginning in the third quarter of fiscal year 2003, field offices will report all activities related to predatory lending to headquarters each quarter. However, not all long- term measures listed in the strategic plan have a corresponding intermediate performance measure. 2. HUD disagreed with our observation that it did not explain in its strategic plan how it used the results of program evaluations to update the current plan and did not include a schedule for future evaluations. As we have previously reported, program evaluations are individual, systematic studies that use objective measurement and analysis to answer specific questions about how well a program is working and, thus, may take many forms. Where a program aims to produce changes that result from program activities, outcome or effectiveness evaluations assess the extent to which those results were achieved. Where complex systems or events outside a program's control also influence its outcomes, impact evaluations use scientific research methods to establish the causal connection between outcomes and program activities and isolate the program's contribution to those changes. A program evaluation that also systematically examines how a program was implemented can provide important information about why a program did or did not succeed and suggest ways to improve it.[Footnote 143] In its strategic plan, HUD provides a few examples of how it modified performance measures as a result of program evaluations. However, we found that 38 of the 41 performance measures discussed in the strategic plan did not mention how, if at all, HUD revised and/or updated them as the result of program evaluations. Elsewhere in the plan, HUD discussed program evaluation activities carried out by its Office of Policy Development and Research; however, a significant number of those evaluations will take place in the future and there is no fixed timetable for when HUD will issue reports on its findings. 3. HUD questioned an example we used to show that its strategic plan did not always provide a clear picture of how it will be able to measure progress toward its strategic goals. We chose this example because HUD used the number of jobs created or retained to measure its progress in achieving the results of the Community Development Block Grant (CDBG) program. As HUD discusses in it strategic plan, there are factors external to the CDBG program, such as broad macro-economic trends and HUD's limited control over how grant recipients use the funding, which can significantly affect job creation in a community. Therefore, it is difficult to establish the contribution of the CDBG program--apart from the other factors--to HUD's stated goal. 4. HUD also disagreed with our observation that in its annual performance report it did not state the steps it would take to address unmet performance goals. We recognize that in some instances HUD mentioned how it would address unmet goals. GPRA requires that agencies explain and describe, where a performance goal has not been met, why the goal was not met, schedules for achieving the established performance goal, and whether or not the performance goal is impractical or unfeasible. However, our review of HUD's performance report found that of the 93 unmet performance targets for fiscal year 2002, 74 lacked an explanation of how HUD would address them in fiscal year 2003. 5. In commenting on our observation that HUD did not include an evaluation of its fiscal year 2003 performance plan relative to the performance attained by the department in fiscal year 2002, HUD mentioned that we should consider the impact of the acceleration of reporting deadlines on the department's ability to include an evaluation of the new fiscal year's performance plan relative to the performance attained in the just completed fiscal year and reasonable alternative actions to fulfill this requirement. While we acknowledge that changes in the reporting deadlines can create challenges for federal agencies, these deadlines are governmentwide and not specific to HUD. In our review of agency plans we found that some agencies, such as DOT, were able to collect performance information for 95 percent of their performance indicators and were able to predict future performance, despite not having complete performance information and facing the same deadlines. DOT provided an evaluation of whether or not fiscal year 2003 performance targets would be met for each of its 40 performance goals based on fiscal year 2002 results. These evaluations were included for the two performance goals for which data were unavailable. For example, for the measure "Number of employment sites (in the thousands) that are made accessible by Job Access and Reverse Commute (JARC) transportation services," DOT could not characterize performance since data had not yet been received from JARC grantees. The 2002 performance report stated that a new easier to use reporting system is being implemented that should improve data gathering performance. The report further stated that DOT would meet this target in fiscal year 2003. 6. HUD also disagreed with how we presented the performance information in its summary report cards (see fig. 22). HUD noted that many of the results were explained in the individual indicator write-ups that followed the summary information. Our review of HUD's reports included, among other things, qualitative aspects of how the information was presented, such as its usefulness to inform the average reader with little to no exposure on the subject matter, and the extent to which it presented summarized performance information that was complete and user-friendly. Our analysis of HUD's performance information was largely based on a review of the information and terms used in the performance report cards. We characterized some of HUD's performance indicators as being "undetermined," given that HUD did not clearly indicate whether or not a goal was achieved. Instead, HUD provided footnotes, such as "results too complex to summarize." We also characterized some performance targets as having "no explanation," given that information was missing from the report card to determine whether HUD had reached its desired target. To develop the graphic summarizing HUD's performance information, we compiled the results of HUD's performance indicators across all five report cards contained in the report. [End of section] Appendix XI: Comments from the Social Security Administration: SOCIAL SECURITY: The Commissioner: January 16, 2004: Ms. Patricia A. Dalton: Director, Strategic Issues: U.S. General Accounting Office Room 2440C: 441 G Street, NW Washington, D.C. 20548: Dear Ms. Dalton: Thank you for the opportunity to review and comment on the draft report "Review Results-Oriented Government: Government Performance Results Act (GPRA) has Established a Solid Foundation for Achieving Greater Results" (GAO-04-38). Our comments on the report contents and suggestions are enclosed. If you have any questions, please contact Candace Skurnik, Director, Audit Management and Liaison Staff at (410) 965-4636. Staff questions should be directed to Laura Bell at (410) 965-2636. Sincerely, Signed by: Jo Anne B. Barnhart: Enclosure: SOCIAL SECURITY ADMINISTRATION BALTIMORE MD 21235-0001: COMMENTS ON THE GENERAL ACCOUNTING OFFICE (GAO) DRAFT REPORT "REVIEW RESULTS-ORIENTED GOVERNMENT: GOVERNMENT PERFORMANCE RESULTS ACT (GPRA) HAS ESTABLISHED A SOLID FOUNDATION FOR ACHIEVING GREATER RESULTS" (GAO- 04-38): Thank you for the opportunity to review and comment on the draft report. I am proud to see that the Social Security Administration (SSA) was once again the only Agency, out of the six reviewed, where GAO found that our Strategic Plan (SP) contained all required elements during both the 1997 and the 2002 review. The figure on page 28 clearly demonstrates our commitment and continued progress in our GPRA planning efforts. It should be noted that our performance measures are more outcome oriented and that we have made, and are continuing to make, progress in integrating SSA's budget with performance. For example, we are able to align costs and workyears with overarching performance goals in our SP (e.g., eliminating backlogs) as part of the budget development process. In summary, we agree for the most part with your conclusions and appreciate the suggestions for improvement. We will incorporate them where appropriate in our future planning efforts. Regarding the specific references to our SP, Annual Performance Plan (APP), and Performance and Accountability Report (PAR) please see my comments below. In addition, I am providing some technical comments that should be included in the final report. Strategic Plan: On page 47 the report suggests that we could further improve our SP through: 2) Explicitly describing the effect of external factors on goal attainment. Response: We believe that the relationship between external factors and our goals, objectives and strategies are clear in most cases. Our SP clearly identifies four categories of environmental factors impacting our future operations: Demographics, Health and Disability Trends, Technological Advances and Workforce Trends. Each strategic goal includes an introduction and description of related issues that explains the interrelationship between these factors and our strategic priorities. For example, under Service we explain how changing demographics (page 15), health and disability trends, including the desire of disabled people to return to work (page 16), and improvements in technology (page 18), impact our service delivery strategies. The impact of environmental factors is equally clear in the context of our Solvency and Staff goals. Our Stewardship goal, driven by the need for the effective management of Agency finances and assets, is the only goal not explicitly linked to environmental factors. 3) Providing timetables or schedules for achieving results. Response: Generally speaking, unless specified, we expect to achieve our long- term outcomes within the 5-year period covered by the SP. In some instances, shorter time frames are specified for key outcomes. More detailed plans and timetables for specific activities are featured in our APP. 4) Providing details on how each performance and accountability challenge will be addressed. Response: Presumably this is a reference to the Major Management Challenges (MMC) identified by the GAO and our Office of Inspector General (OIG). We are not aware of a specific requirement in the Office of Management and Budget's (OMB) A-1 I instructions to address every one of these challenges in our long-term SP's. While our current SP does address some of the challenges that were identified at the time of its publication, because the MMC's are updated every year, we believe that they are more appropriately addressed in the context of our APP and our annual PAR. Therefore, we have made more of an effort to address the MMC beginning with our 2004 APP and our 2003 PAR. 6) Discussing the manner in which SSA coordinated with other agencies, especially those that serve the same beneficiaries. Response: While we agree that such a discussion would be useful, we note that it is not one of the OMB requirements for preparing a strategic plan. Our SP does discuss in several places our interaction with other Government agencies including our partnership with the State Disability Determination Services, and collaboration with the Department of Labor on Ticket to Work initiatives. Again, while we believe that this is not an OMB requirement for preparing a strategic plan, we will make a greater effort to provide details regarding such collaborations in future SP's. Annual Performance Plan: This section highlights our improvement in identifying data sources and definitions for each performance measure but observes that there is no discussion of verification and validation procedures for data generated by these systems. Response: We will continue to improve our verification and validation procedures and ensure that there is adequate documentation in the APP. Currently Appendix C (pages 69 to 70) of our APP outlines our validation efforts. We maintain internal controls on our data gathering systems and make significant efforts to validate those controls by use of a variety of audits, both internal and external. Most notably, PricewaterhouseCoopers (PwC) conducts an audit of our financial statements each year. This audit includes a very thorough review of the 15 or so key performance measures highlighted in our annual PAR. We also work closely with our OIG to validate the remaining performance measures through ongoing reviews. Annual Performance Report (Performance and Accountability Report-PAR): Page 118 states that missing data and a lack of documentation of the methods used to measure our performance reduced the overall quality of the document. Response: As mentioned previously, in addition to our own internal controls, we work with PwC and our OIG to validate our data sources. Our annual PAR has received the Certificate of Excellence in Accountability Reporting for five consecutive years, in large part because of this collaboration. Beginning in FY 2003, we began placing greater emphasis on documenting the processes and validation methods used in our data gathering activities. We will do more to report on these efforts in future PAR'S. Also, on page 121 in the program evaluation section, GAO implies the need for us to report on how our evaluations have helped answer questions about program performance and results as opposed to merely describing the surveys we conduct with those who do business with us or based on our internal needs assessment surveys. Response: We disagree with this assessment. Many of the evaluations directed at our efforts to deliver high quality, citizen-centered service rely on these surveys. This assessment seems to minimize the importance of these surveys; however, we feel that they are an extremely important part of our service delivery efforts. Furthermore, the program evaluation section of our report lists about a dozen: other evaluations that are of great importance to our ongoing operations. For example, our continuing disability and redetermination reviews reduce program outlays by billions of dollars every year. In the future, we will make more of an effort to explain how these and other evaluations help improve program performance. Technical Comments: References on pages 13 and 83 state that we did not structure our performance and strategic plans by program activity or account. The references on page 83 then go on to say that "SSA noted that it aligned its strategic goals, performance measures, and budget with its major functional responsibilities rather than by program accounts since direct service and support employees provide services linked to these functional responsibilities, as opposed to a specific program. However, SSA does not indicate what it means by `functional responsibilities,' nor does it show a clear linkage between its strategic goals and such responsibilities." We believe that we do attempt to describe our functional responsibilities in Parts I and II of our FY 2004 APP. We will take under advisement GAO's concern and consider including clearer descriptions in future APPs. On page 21 we believe that footnote 10 should be included as part of the narrative in the first paragraph and that the sentence be modified as follows: "According to the Inspector General, SSA began to implement an improved cost accounting system in FY 2002, which was to be phased in over the next 3 to 4 years." The current language makes it appear that we did not have a cost accounting system prior to 2002. The following are our comments on SSA's letter dated January 16, 2004. GAO Comments: In general, SSA agreed with our conclusions. SSA also agreed to incorporate the suggestions for improvement in its future planning efforts. SSA made several points of clarification and disagreed with our assessment in one area. 1. In our draft report, we noted that SSA did not explicitly link external factors that may affect its programs to its general goals and state how these factors could affect goal attainment. SSA attests that its four categories of environmental factors are discussed under each of the strategic goals, as appropriate, and the relationship between these factors and SSA's strategic priorities is described. This general discussion of the environmental factors is useful in understanding the challenges SSA faces in working toward its broad strategic goals. However, SSA provides little or no discussion of these challenges in its discussion of the agency's performance goals. Thus, the range of challenges facing the agency in meeting each of its performance goals is not fully explained. 2. In our draft report, we noted that SSA does not provide timetables or schedules for achieving all the results in its strategic plan. SSA noted that it expects to achieve its long-term outcomes within the 5- year period covered by the strategic plan; in selected instances, shorter time frames are specified. SSA noted that more detailed plans and timetables are featured in its annual performance plan. GPRA requires agencies to furnish a schedule of significant actions in their strategic plans; however, SSA does not clearly articulate its timetables and schedules for achieving each of its long-term outcomes in its strategic plan. 3. We noted that SSA's strategic plan could be improved by providing details on how each performance and accountability challenge will be addressed. SSA asserted that the strategic plan addresses some of the challenges, but because the challenges are updated every year, they are more appropriately addressed in the annual performance plan and performance and accountability report. As noted in our discussion of the criteria used to analyze agencies' strategic plans, it is particularly important that agencies develop strategies that address management challenges that threaten their ability to meet long-term strategic goals, as one of the purposes of GPRA is to improve the management of federal agencies. 4. In our draft report, we observed that SSA's discussion of its interactions with other agencies, especially those that serve the same beneficiaries, was limited. SSA noted that such a discussion would be useful, but is not an OMB requirement. While we agree that this is not an OMB requirement, we have reported that given scarce resources and competing priorities, it would be useful to identify agency efforts to maximize its effect through cooperation and coordination across the federal government. Better strategic plans not only identify the need to coordinate with other agencies, but also discuss how agencies intend to coordinate common or complementary goals and strategies with other agencies. 5. With regard to SSA's performance and accountability report, we noted that SSA did not clearly state how program evaluations were used to answer questions about program performance and results and how those results can be improved. SSA disagreed with our observation, stating that many of its evaluations rely on surveys, and these surveys form the basis for its efforts to deliver high-quality service. SSA also noted that it listed other evaluations that are of great importance to its ongoing operations. We do not discount the usefulness of SSA's surveys in assessing its day-to-day management of programs. Rather, as we noted in the report, it would be helpful for SSA to clearly identify the range of evaluations conducted and how each of them contributed to improved program performance. For example, we recently recommended that SSA evaluate a new initiative to improve the integrity of Social Security number issuance to noncitizens; the description of such an evaluation would be helpful for SSA to determine how it can be best positioned to ensure the integrity of its enumeration process.[Footnote 144] Additionally, our September 2000 report on program evaluation states that GPRA recognizes the complementary nature of program evaluation and performance measurement. Strategic plans are to describe the program evaluations that were used in establishing and revising goals and to include a schedule for future program evaluations. Agencies are to summarize the findings of program evaluations in their annual performance reports. Additionally, SSA made technical comments that we incorporated into the report, as appropriate. [End of section] Appendix XII: GAO Contact and Staff Acknowledgments: GAO Contact: Patricia Dalton, (202) 512-6806: Acknowledgments: In addition to the persons mentioned above, Thomas Beall, Daniel Bertoni, Kay Brown, Joyce Corry, Elizabeth Curda, David Dornisch, William Fenzel, Kimberly Gianopoulos, Evan Gilman, Katie Harris, Susan Higgins, Benjamin Licht, William McKelligott, James Noel, Carol Petersen, Carolyn Samuels, Teresa Spisak, Daren Sweeney, Carolyn Taylor, Michael Volpe, Lynn Wasielewski, and Steven Westley made key contributions to this report. [End of section] Related GAO Products: GPRA/Managing for Results: Results-Oriented Government: Using GPRA to Address 21st Century Challenges, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-1166T]; (Washington, D.C. Sept. 18, 2003). Results-Oriented Management: Agency Crosscutting Actions and Plans in Border Control, Flood Mitigation and Insurance, Wetlands, and Wildland Fire Management, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-321]; (Washington, D.C. Dec. 20, 2002). Results-Oriented Management: Agency Crosscutting Actions and Plans in Drug Control, Family Poverty, Financial Institution Regulation, and Public Health Systems, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-320]; (Washington, D.C. Dec. 20, 2002). Performance and Accountability: Reported Agency Actions and Plans to Address 2001 Management Challenges and Program Risks, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-225]; (Washington, D.C. Oct. 31, 2002). Managing for Results: Next Steps to Improve the Federal Government's Management and Performance, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-439T]; (Washington, D.C. Feb. 15, 2002). Managing for Results: Federal Managers' Views on Key Management Issues Vary Widely Across Agencies, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-01-592]; (Washington, D.C. May 25, 2001). Managing for Results: Federal Managers' Views Show Need for Ensuring Top Leadership Skills, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-01-127]; (Washington, D.C. Oct. 20, 2000). Managing for Results: Barriers to Interagency Coordination, Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-106]; (Washington, D.C. Mar. 29, 2000). Management Reform: Elements of Successful Improvement Initiatives, Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-00-26]; (Washington, D.C. Oct. 15, 1999). Management Reform: Using the Results Act and Quality Management to Improve Federal Performance, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-99-151]; (Washington, D.C. July 29, 1999). Managing for Results: Opportunities for Continued Improvements in Agencies' Performance Plans, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-99-215]; (Washington, D.C. July 20, 1999). Agency Performance Plans: Examples of Practices That Can Improve Usefulness to Decisionmakers, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-99-69]; (Washington, D.C. Feb. 26, 1999). The Results Act: Assessment of the Governmentwide Performance Plan for Fiscal Year 1999, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD/GGD-98-159]; (Washington, D.C. Sept. 8, 1998). Managing for Results: An Agenda to Improve the Usefulness of Agencies' Annual Performance Plans, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-98-228]; (Washington, D.C. Sept. 8, 1998). The Results Act: An Evaluator's Guide to Assessing Agency Annual Performance Plans. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-10.1.20]; (Washington, D.C. April 1998). Agencies' Annual Performance Plans Under the Results Act: An Assessment Guide to Facilitate Congressional Decisionmaking, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-10.1.18]; (Washington, D.C. February 1998). Managing for Results: Agencies' Annual Performance Plans Can Help Address Strategic Planning Challenges, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-44]; (Washington, D.C. Jan. 30, 1998). Managing for Results: Building on Agencies' Strategic Plans to Improve Federal Management, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/AIMD-98-29]; (Washington, D.C. Oct. 30, 1997). Managing for Results: Critical Issues for Improving Federal Agencies' Strategic Plans, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-180]; (Washington, D.C. Sept. 16, 1997). Managing for Results: Using the Results Act to Address Mission Fragmentation and Program Overlap, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-97-146]; (Washington, D.C. Aug. 29, 1997). Managing for Results: The Statutory Framework for Improving Federal Management and Effectiveness, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/AIMD-97-144]; (Washington, D.C. June 24, 1997). The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-109]; (Washington, D.C. June 2, 1997). Agencies' Strategic Plans Under GPRA: Key Questions to Facilitate Congressional Review, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-10.1.16]; (Washington, D.C. May 1997). GPRA: Managerial Accountability and Flexibility Pilot Did Not Work as Intended, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-36]; (Washington, D.C. Apr. 10, 1997). Managing for Results: Enhancing the Usefulness of GPRA Consultations Between the Executive Branch and Congress, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-97-56]; (Washington, D.C. Mar. 10, 1997). Managing for Results: Key Steps and Challenges in Implementing GPRA in Science Agencies, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/RCED-96-214]; (Washington, D.C. July 10, 1996). Executive Guide: Effectively Implementing the Government Performance and Results Act, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-96-118]; (Washington, D.C. June 1996). Strategic Human Capital Management: Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government; (Exposure Draft), [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-893G]; (Washington, D.C. July 2003). Results-Oriented Cultures: Creating a Clear Linkage between Individual Performance and Organizational Success, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-488]; (Washington, D.C. Mar. 14, 2003). Results-Oriented Cultures: Insights for U.S. Agencies from Other Countries' Performance Management Initiatives, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-862]; (Washington, D.C. Aug. 2, 2002). Human Capital: Key Principles From Nine Private Sector Organizations, Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-28]; (Washington, D.C. Jan. 31, 2000). Linking Resources to Results: Performance Budgeting: Observations on the Use of OMB's Program Assessment Rating Tool for the Fiscal Year 2004 Budget, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-174]; (Washington, D.C. Jan. 30, 2004). Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Nuclear Regulatory Commission, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-258]; (Washington, D.C. Dec. 10, 2002). Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Administration for Children and Families, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-9]; (Washington, D.C. Dec. 10, 2002). Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Veterans Health Administration, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-10]; (Washington, D.C. Dec. 10, 2002). Managing for Results: Agency Progress in Linking Performance Plans With Budgets and Financial Statements, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-236]; (Washington, D.C. Jan. 4, 2002). Performance Budgeting: Fiscal Year 2000 Progress in Linking Plans With Budgets, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-99-239R]; (Washington, D.C. July 30, 1999). Performance Budgeting: Initial Experiences Under the Results Act in Linking Plans With Budgets, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD/GGD-99-67]; (Washington, D.C. Apr. 12, 1999). Performance Budgeting: Past Initiatives Offer Insights for GPRA Implementation, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-97-46]; (Washington, D.C. Mar. 27, 1997). Measuring Performance: Program Evaluation: An Evaluation Culture and Collaborative Partnerships Help Build Agency Capacity, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-454]; (Washington, D.C. May 2, 2003). Managing for Results: Next Steps to Improve the Federal Government's Management and Performance, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-439T]; (Washington, D.C. Feb. 15, 2002). Program Evaluation: Studies Helped Agencies Measure or Explain Program Performance, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-204]; (Washington, D.C. Sept. 29, 2000). Managing for Results: Strengthening Regulatory Agencies' Performance Management Practices, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-10]; (Washington, D.C. Oct. 28, 1999). Managing for Results: Measuring Program Results That Are Under Limited Federal Control, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-99-16]; (Washington, D.C. Dec. 11, 1998). Grant Programs: Design Features Shape Flexibility, Accountability, and Performance Information, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-137]; (Washington, D.C. June 22, 1998). Program Evaluation: Agencies Challenged by New Demand for Information on Program Results, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-53]; (Washington, D.C. Apr. 24, 1998). Managing for Results: Regulatory Agencies Identified Significant Barriers to Focusing on Results, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-83]; (Washington, D.C. June 24, 1997). Managing for Results: Analytic Challenges in Measuring Performance, Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS/GGD-97-138]; (Washington, D.C. May 30, 1997). Data Credibility: Performance Reporting: Few Agencies Reported on the Completeness and Reliability of Performance Data, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-372]; (Washington, D.C. Apr. 26, 2002). Managing for Results: Assessing the Quality of Program Performance Data, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-140R]; (Washington, D.C. May 25, 2000). Managing for Results: Challenges Agencies Face in Producing Credible Performance Information, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-52]; (Washington, D.C. Feb. 4, 2000). Performance Plans: Selected Approaches for Verification and Validation of Agency Performance Information, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-99-139]; (Washington, D.C. July 30, 1999). Using Performance Information: Managing for Results: Views on Ensuring the Usefulness of Agency Performance Information to Congress, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-35]; (Washington, D.C. Jan. 26, 2000). Managing for Results: Using GPRA to Assist Congressional and Executive Branch Decisionmaking, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-97-43]; (Washington, D.C. Feb. 12, 1997). Managing for Results: Achieving GPRA's Objectives Requires Strong Congressional Role, [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-96-79]; (Washington, D.C. Mar. 6, 1996). (450181): FOOTNOTES [1] Pub. L. No. 103-62. [2] U.S. General Accounting Office, Managing for Results: Opportunities for Continued Improvements in Agencies' Performance Plans, GAO/GGD/ AIMD-99-215 (Washington, D.C.: July 20, 1999). [3] PART is a diagnostic tool developed by OMB that it has been using to rate the effectiveness of federal programs with a particular focus on program results. OMB's goal is to review all federal programs over a 5-year period using the PART tool. OMB used the tool to review approximately 400 programs between the fiscal year 2004 budget cycle and the fiscal year 2005 budget cycle--234 programs were assessed last year and 173 were assessed this year. Some reassessed programs were combined for review for the 2005 budget, which is why the number of programs assessed over the 2 years does not add up to exactly 400 programs. [4] U.S. General Accounting Office, Program Performance Measures: Federal Agency Collection and Use of Performance Data, GAO/GGD-92-65 (Washington, D.C.: May 4, 1992). [5] Types of measures were defined in the questionnaire as follows: performance measures that tell us how many things we produce or services we provide (output measures); performance measures that tell us if we are operating efficiently (efficiency measures); performance measures that tell us whether or not we are satisfying our customers (customer service measures); performance measures that tell us about the quality of the products or services we provide (quality measures); and performance measures that would demonstrate to someone outside of our agency whether or not we are achieving our intended results (outcome measures). [6] U.S. General Accounting Office, Managing for Results: An Agenda to Improve the Usefulness of Agencies' Annual Performance Plans, GAO/GGD/ AIMD-98-228 (Washington, D.C.: Sept. 8, 1998). [7] U.S. General Accounting Office, Managing for Results: Critical Issues for Improving Agencies' Strategic Plans, GAO/GGD-97-180 (Washington, D.C.: Sept. 16, 1997). [8] GAO/GGD/AIMD-98-228. [9] Pub. L. No. 103-62. [10] The guidance on the preparation of strategic plans, annual performance plans, and program performance reports is contained in OMB Circular A-11, Part 6. [11] OMB Circular A-11, July 2003. [12] PART is a diagnostic tool developed by OMB that it has been using to rate the effectiveness of federal programs with a particular focus on program results. OMB's goal is to review all federal programs over a 5-year period using the PART tool. OMB used the tool to review approximately 400 programs between the fiscal year 2004 budget cycle and the fiscal year 2005 budget cycle--234 programs were assessed last year and 173 were assessed this year. Some reassessed programs were combined for review for the 2005 budget, which is why the number of programs assessed over the 2 years does not add up to exactly 400 programs. [13] U.S. General Accounting Office, GPRA: Managerial Accountability and Flexibility Pilot Did Not Work as Intended, GAO/GGD-97-36 (Washington, D.C.: Apr. 10, 1997). [14] U.S. General Accounting Office, Managing for Results: Agency Progress in Linking Performance Plans With Budgets and Financial Statements, GAO-02-236 (Washington, D.C.: Jan. 4, 2002). [15] GAO/GGD/AIMD-99-215. Based on how we had rated agencies' annual performance plans on their picture of performance, specificity of strategies and resources, and the degree of confidence that performance information will be credible, we assigned numeric values to each agencies' rating (e.g., clear=3, general=2, limited=1, unclear=0) and added them up to determine overall quality of high, medium, or low. An agency's plan was considered high quality if its score was between 7-9, a score of 5-6 was considered medium quality, and a score of 3-4 was low. No agencies received a score lower than 3. [16] GAO/GGD-92-65. [17] U.S. General Accounting Office, Performance Budgeting: Observations on the Use of OMB's Program Assessment Rating Tool for the Fiscal Year 2004 Budget, GAO-04-174 (Washington, D.C.: Jan. 30, 2004). [18] Types of measures were defined in the questionnaire as follows: performance measures that tell us how many things we produce or services we provide (output measures); performance measures that tell us if we are operating efficiently (efficiency measures); performance measures that tell us whether or not we are satisfying our customers (customer service measures); performance measures that tell us about the quality of the products or services we provide (quality measures); and performance measures that would demonstrate to someone outside of our agency whether or not we are achieving our intended results (outcome measures). [19] GAO/GGD/AIMD-98-228. [20] GAO-02-236; U.S. General Accounting Office, Performance Budgeting: Initial Experiences Under the Results Act in Linking Plans With Budgets, GAO/AIMD/GGD-99-67 (Washington, D.C.: Apr. 12, 1999); and Performance Budgeting: Fiscal Year 2000 Progress in Linking Plans With Budgets, GAO/AIMD-99-239R (Washington, D.C.: July 30, 1999). [21] U.S. General Accounting Office, Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Veterans Health Administration, GAO-03-10 (Washington, D.C.: Dec. 10, 2002). [22] U.S. General Accounting Office, Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Administration for Children and Families, GAO-03-09 (Washington, D.C.: Dec. 10, 2002). [23] U.S. General Accounting Office, Managing for Results: Efforts to Strengthen the Link Between Resources and Results at the Nuclear Regulatory Commission, GAO-03-258 (Washington, D.C.: Dec. 10, 2002). [24] GAO/GGD-97-180. [25] U.S. General Accounting Office, Managing for Results: Challenges Agencies Face in Producing Credible Performance Information, GAO/GGD- 00-52 (Washington, D.C.: Feb. 4, 2000). [26] Since 1990, GAO has periodically reported on government operations that it identifies as "high risk" because of the greater vulnerabilities to fraud, waste, abuse, and mismanagement. See U.S. General Accounting Office, High-Risk Series: An Update, GAO-03-119 (Washington, D.C.: January 2003). [27] U.S. General Accounting Office, Managing for Results: Using the Results Act to Address Mission Fragmentation and Program Overlap, GAO/ AIMD-97-146 (Washington, D.C.: Aug. 29, 1997). [28] GAO/GGD/AIMD-98-228. [29] Program activity refers to the list of projects and activities in the appendix portion of the Budget of the United States Government. Program activity structures are intended to provide a meaningful representation of the operations financed by a specific budget account. [30] GAO/GGD-97-36. [31] U.S. General Accounting Office, Major Management Challenges and Program Risks: Department of Energy, GAO-03-100 (Washington, D.C.: January 2003). [32] According to OMB's Statement of Federal Financial Accounting Standards No. 4--Managerial Cost Accounting Standards, July 31, 1995, a responsibility segment is a component of a reporting entity that is responsible for carrying out a mission, conducting a major line of activity, or producing one or a group of related products or services. In addition, responsibility segments usually possess the following characteristics: (1) their managers report to the entity's top management directly and (2) their resources and results of operations can be clearly distinguished from those of other segments of the entity. Managerial cost accounting should be performed to measure and report the costs of each segment's outputs. [33] SSA noted that its fifth strategic goal, "Valued Employees," supports the accomplishment of all its basic functions, so its resources are inherently included in the other four goals. [34] According to the IG, SSA began to implement an improved cost accounting system in fiscal year 2002, which was to be phased in over the next 3 to 4 years. [35] U.S. General Accounting Office, Performance Reporting: Few Agencies Reported on the Completeness and Reliability of Performance Data, GAO-02-372 (Washington, D.C.: Apr. 26, 2002). [36] U.S. General Accounting Office, Management Reform: Elements of Successful Improvement Initiatives, GAO/T-GGD-00-26 (Washington, D.C.: Oct. 15, 1999). [37] U.S. General Accounting Office, The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven, GAO/ GGD-97-109 (Washington, D.C.: June 2, 1997) and Managing for Results: Federal Managers' Views Show Need for Ensuring Top Leadership Skills, GAO-01-127 (Washington, D.C.: Oct. 20, 2000). [38] GAO/GGD-97-109 and GAO-01-127. [39] GAO-01-127. [40] U.S. General Accounting Office, Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government (Exposure Draft) GAO-03-893G (Washington, D.C.: July 1, 2003). [41] For a complete list and discussion of the practices, see U.S. General Accounting Office, Results oriented Cultures: Creating a Clear Linkage between Individual Performance and Organizational Success, GAO- 03-488 (Washington, D.C.: Mar. 14, 2003). [42] See for example, U.S. General Accounting Office, Managing for Results: Analytic Challenges in Measuring Performance, GAO/HEHS/GGD- 97-138 (Washington, D.C.: May 30, 1997); Program Evaluation: Agencies Challenged by New Demand for Information on Program Results, GAO/GGD- 98-53 (Washington, D.C.: Apr. 24, 1998); Managing for Results: Measuring Program Results That Are Under Limited Federal Control, GAO/ GGD-99-16 (Washington, D.C.: Dec. 11, 1998); and Managing for Results: Challenges Agencies Face in Producing Credible Performance Information, GAO/GGD-00-52 (Washington, D.C.: Feb. 4, 2000). [43] GAO/GGD-97-109, 6. [44] U.S. General Accounting Office, Managing for Results: Key Steps and Challenges in Implementing GPRA in Science Agencies, GAO/T-GGD/ RCED-96-214 (Washington, D.C.: July 10, 1996). [45] U.S. General Accounting Office, Managing for Results: Strengthening Regulatory Agencies' Performance Management Practices, GAO/GGD-00-10 (Washington, D.C.: Oct. 28, 1999). [46] U.S. General Accounting Office, Environmental Protection: Assessing the Impacts of EPA's Regulations Through Retrospective Studies, GAO/RCED-99-250 (Washington, D.C.: Sept. 14, 1999). [47] U.S. General Accounting Office, Grant Programs: Design Features Shape Flexibility, Accountability, and Performance Information, GAO/ GGD-98-137 (Washington, D.C.: June 22, 1998). [48] GAO/GGD-00-52. [49] U.S. General Accounting Office, Managing for Results: Barriers to Interagency Coordination, GAO/GGD-00-106 (Washington, D.C.: Mar. 29, 2000). [50] U.S. General Accounting Office, Wetlands Overview: Problems With Acreage Data Persist, GAO/RCED-98-150 (Washington, D.C.: July 1, 1998) and Results-Oriented Management: Agency Crosscutting Actions and Plans in Border Control, Flood Mitigation and Insurance, Wetlands, and Wildland Fire Management, GAO-03-321 (Washington, D.C.: Dec. 20, 2002). [51] GAO/GGD-00-106 and U.S. General Accounting Office, Results- Oriented Government: Using GPRA to Address 21st Century Challenges, GAO-03-1166T (Washington, D.C.: Sept. 18, 2003). [52] U.S. General Accounting Office, The Results Act: Assessment of the Governmentwide Performance Plan for Fiscal Year 1999, GAO/AIMD/GGD-98- 159 (Washington, D.C.: Sept. 8, 1998). [53] Congressional Research Service, Government Performance and Results Act: Overview of Associated Provisions in the 106th Congress, (Washington, D.C.: 2002). [54] This included reports that accompanied bills passed by both the House and Senate that were either enacted into law or vetoed by the President. [55] U.S. General Accounting Office, Managing for Results: Views on Ensuring the Usefulness of Agency Performance Information to Congress, GGD-00-35 (Washington, D.C.: Jan. 26, 2000). [56] U.S. General Accounting Office, Human Capital: Key Principles From Nine Private Sector Organizations, GAO/GGD-00-28 (Washington, D.C.: Jan. 31, 2000). [57] GAO/AIMD-97-146. [58] U.S. General Accounting Office, Executive Guide: Effectively Implementing the Government Performance and Results Act, GAO/GGD-96-118 (Washington, D.C.: June 1, 1996). [59] GAO/GGD-00-35. [60] U.S. General Accounting Office, Program Evaluation: Agencies Challenged by New Demand for Information on Program Results, GAO/GGD- 98-53 (Washington, D.C.: Apr. 24, 1998). [61] U.S. General Accounting Office, Program Evaluation: Studies Helped Agencies Measure or Explain Program Performance, GAO/GGD-00-204 (Washington, D.C.: Sept. 28, 2000). [62] GAO/GGD-00-52. [63] For information on the design and administration of the two earlier surveys, see GAO/GGD-97-109, GAO-01-127, and U.S. General Accounting Office, Managing For Results: Federal Managers' Views on Key Management Issues Vary Widely Across Agencies, GAO-01-592 (Washington, D.C.: May 25, 2001). [64] For example, candidates could be operations managers with hands-on experience managing a federal program or agency officials directly involved in carrying out the activities required under GPRA, such as developing a strategic or annual performance plan or annual performance report. [65] Due to last minute circumstances, a federal manager participated via teleconference from an agency's field office and another was unable to attend the focus group, but mailed his answers to questions we sent to all participants in advance of the focus groups. [66] We read the list of comments to the manager who participated via teleconference and voted on his behalf based on his preferences. [67] GAO/GGD/AIMD-99-215. Based on how we had rated agencies' annual performance plans on their picture of performance, specificity of strategies and resources, and the degree of confidence that performance information will be credible, we assigned numeric values to each agencies' rating (e.g., clear=3, general=2, limited=1, unclear=0) and added them up to determine overall quality of high, medium, or low. An agency's plan was considered high quality if its score was between 7-9, a score of 5-6 was considered medium quality, and a score of 3-4 was low. No agencies received a score lower than 3. [68] Government Performance and Results Act of 1993, Committee on Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong. 1st Sess. (1993). [69] Office of Management and Budget, Circular No. A-11, Part 6, Preparation and Submission of Strategic Plans, Annual Performance Plans, and Annual Program Performance Reports (Washington, D.C.: June 2002). [70] U.S. General Accounting Office, Agencies' Strategic Plans Under GPRA: Key Questions to Facilitate Congressional Review, GAO/GGD-10.1.16 (Washington, D.C.: May 1, 1997). [71] GAO/HEHS-97-176R; GAO/RCED-97-199R; GAO/RCED-97-208R; GAO/HEHS- 97-179R; GAO/RCED-97-205R; and GAO/RCED-97-224R. [72] GAO/HEHS-97-176R. [73] U.S. Department of Education, Office of the Deputy Secretary, Strategic Accountability Service, U.S. Department of Education FY 2004 Annual Plan (Washington, D.C.: March 2003). This document represents Education's performance plan for GPRA and will be referred to henceforth as the "annual plan." [74] U.S. General Accounting Office, Performance and Accountability Series: Major Management Challenges and Program Risks, Department of Education, GAO-01-245 (Washington, D.C.: January 2001). The January 2001 assessment was the last time we assessed the Department of Education under our Performance and Accountability Series before the release of the agency's 2002-2007 Strategic Plan. We further reported in October 2002 on how Education and other agencies reported responding to their management challenges and program risks. (U.S. General Accounting Office, Performance and Accountability: Reported Agency Actions and Plans to Address 2001 Management Challenges and Program Risks, GAO-03-225 (Washington, D.C.: Oct. 31, 2002). [75] Within one of its strategies, the plan states that "managers will be given the freedom to manage and will be held accountable for results." [76] Pub. L. No. 107-110, January 8, 2002. The No Child Left Behind Act of 2001 is a reauthorization of the Elementary and Secondary Education Act, one of the major pieces of authorizing legislation for the Department of Education. [77] GAO/RCED-97-199R. [78] GAO/RCED-97-224R. [79] U.S. General Accounting Office, Major Management Challenges and Program Risks: Department of Housing and Urban Development, GAO-03-103 (Washington, D.C.: January 2003). [80] GAO/RCED-97-205R. [81] At the time of our review, the most recent SBA strategic plan was for fiscal years 2001-2008. SBA released a new strategic plan for fiscal years 2003-2008 in October 2003. [82] U.S. General Accounting Office, Major Management Challenges and Program Risks: Small Business Administration, GAO-01-260 (Washington, D.C.: January 2001). [83] In the 2003-2008 plan, the word "advance" replaced the word "promote." [84] U.S. General Accounting Office, Major Management Challenges and Program Risks: Social Security Administration, GAO-03-117 (Washington, D.C.: January 2003). [85] The suspense file contains information on earnings that cannot be matched to an individual's record due to an invalid name/Social Security number combination. [86] GAO/RCED-97-208R. [87] At the time of our review, the Department of Transportation was in the process of revising its strategic plan. A draft copy of the updated strategic plan, dated July 1, 2003, was used for this review. [88] U.S. General Accounting Office, Major Management Challenges and Program Risks: Department of Transportation, GAO-03-108 (Washington, D.C.: January 2003). [89] Government Performance and Results Act of 1993, Committee on Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong. 1st Sess. (1993). [90] OMB Circular No. A-11: Part 6, Preparation and Submission of Strategic Plans, Annual Performance Plans, and Annual Program Performance Reports (Washington, D.C.: June 2002). [91] GAO/GGD/AIMD-99-215 and The Results Act: An Evaluator's Guide to Assessing Agency Annual Performance Plans, GAO/GGD-10.1.20 (Washington, D.C.: April 1998). [92] GAO/HEHS-98-172R, GAO/RCED-98-194R, GAO/RCED-98-159R, GAO/RCED- 98-180R, GAO/RCED-98-200R, and GAO/HEHS-98-178R. [93] The Homeland Security Act (Pub. L. No. 107-296), also requires that agencies provide a description of how the performance goals and objectives are to be achieved, including the operations, processes, training, skills and technology, and the human capital, information, and other resources and strategies required to meet those performance goals and objectives. [94] GAO/HEHS-98-172R, GAO/RCED-98-194R, GAO/RCED-98-159R, GAO/RCED- 98-180R, GAO/RCED-98-200R, and GAO/HEHS-98-178R. [95] Education's plan states that its fiscal year 2004 annual plan includes both department-level measures and program performance plans. These are organized into two volumes: the Annual Plan Fiscal Year 2004 U.S. Department of Education includes the department-level measures and the FY 2004 Program Performance Plan: U.S. Department of Education includes the program performance plans with their individual program measures. These volumes are presented in a slightly different electronic format for the public and other parties in general, which is available at Education's Web site. The two volumes will henceforth be referred to as Education's 2004 annual plan, or, where applicable, volume 1 and volume 2. [96] Education's 2004 annual plan represents its annual performance goals as "targets." According to GPRA, the definition for "performance goal" is "a target level of performance." [97] In this report, we refer to the multiyear, long-term objectives in Education's annual plan as "long-term goals." The strategic goals included in the plan represent overarching statements of aim or purpose that are used to group Education's long-term goals. [98] GAO/HEHS-97-176R. [99] U.S. General Accounting Office, Performance and Accountability Series--Major Management Challenges and Program Risks: A Governmentwide Perspective, GAO-01-241 (Washington, D.C.: January 2001) and GAO-01- 245. [100] GAO-03-225. [101] GAO-03-100. [102] GAO/RCED-98-159R. [103] GAO/RCED-98-200R. [104] Much of this improvement took place between the 1999 and 2000 plans. We reported that SSA's fiscal year 2000 performance plan showed significant improvement over its 1999 plan in U.S. General Accounting Office, Observations on the Social Security Administration's Fiscal Year 2000 Performance Plan, GAO/HEHS-99-162R (Washington, D.C.: July 20, 1999). [105] SSA reduced its strategic goals from five to four in its 2003- 2008 strategic plan. [106] GAO-03-225. [107] SSA refers to external factors as environmental factors. [108] GAO/RCED-98-180R. [109] Bureau of Transportation Statistics, Source & Accuracy Compendium, http://www.bts.gov/statpol/SAcompendium.html (Washington, D.C.: Aug. 15, 2003). [110] Office of Management and Budget, Memorandum: Program Assessment Rating Tool (PART) - Presentation in Congressional Justifications, M- 03-06 (Washington, D.C.: 2003). [111] Government Performance and Results Act of 1993, Committee on Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong. 1st Sess. (1993). [112] OMB Circular No. A-11, Part 6, Preparation and Submission of Strategic Plans, Annual Performance Plans, and Annual Program Performance Reports (Washington, D.C.: June 2002). [113] Chief Financial Officers Act of 1990 (Pub. L. No. 101-576). [114] GAO-02-372 and Executive Guide: Creating Value Through World- class Financial Management, GAO/AIMD-00-134 (Washington, D.C.: Apr. 1, 2000). [115] The Homeland Security Act (Pub. L. No. 107-296) requires agencies to include a review of the performance goals and evaluation of the performance plan relative to the agency's strategic human capital management. [116] Association of Government Accountants, Certificate of Excellence in Accountability Reporting: Reviewers Checklist, Fiscal Year 2001. (Washington, D.C.). [117] GAO-02-372. [118] GAO/HEHS-00-128R, GAO/RCED-00-209R, GAO/RCED-00-211R, GAO/RCED- 00-207R, GAO/HEHS-00-126R, and GAO/RCED-00-201R. [119] Education's annual performance goals are represented by its targets. [120] GAO/HEHS-00-128R. [121] Pub. L. No. 107-110, January 8, 2002. The No Child Left Behind Act of 2001 is a reauthorization of the Elementary and Secondary Education Act, one of the major pieces of authorizing legislation for Education. [122] GAO/HEHS-00-128R. [123] As amended by the Homeland Security Act of 2002. [124] GAO-03-225. [125] U.S. Department of Housing and Urban Development, Fiscal year 2002 Performance and Accountability Report (Washington, D.C.: 2003). [126] GAO/RCED-00-211R. [127] GAO-03-225. [128] Small Business Administration, SBA's Performance & Accountability Report for Fiscal Year 2002 (Washington, D.C.: 2003). [129] Our review of SBA's fiscal year 1999 report, GAO/RCED-00-207R, focused on our observations on only three of SBA's key outcomes, as well as the major management challenges addressed in the performance report. Since our review of SBA's 2002 performance report used somewhat different assessment criteria, we could not make valid comparisons on all aspects of the reports. [130] Social Security Administration, Performance and Accountability Report, Fiscal Year 2002 (Washington, D.C.: 2002). [131] Our review of SSA's fiscal year 1999 performance report (GAO/ HEHS-00-126R) focused on our observations on five of the agency's key outcomes, as well as the major management challenges addressed in the performance report. Since our review of SSA's 2002 performance report used different assessment criteria, we could not make valid comparisons on all aspects of the reports. [132] This program was an administration proposal to test allowing disabled beneficiaries to choose their own public or private vocational rehabilitation provider. [133] In our October 2002 report GAO-03-225, we noted that SSA reported progress on all six of its major management challenges in its fiscal year 2001 annual performance report. [134] We characterized program evaluations and their uses in GAO/GGD- 00-204. [135] SSA noted that its fifth strategic goal, "Valued Employees," supports the accomplishment of all its basic functions, so its resources are inherently included in the other four goals. [136] SSA began to implement an improved cost accounting system in fiscal year 2002, which will be phased in over the next 3 to 4 years. [137] U.S. Department of Transportation, Fiscal Year 2002 Performance and Accountability Report (Washington, D.C.: 2003). [138] Our review of DOT's fiscal year 1999 report, GAO/RCED-00-201R, focused on our observations regarding only four of the department's key outcomes, as well as the major management challenges addressed in the performance report. Since our review of DOT's 2002 performance report used different assessment criteria, we could not make valid comparisons on all aspects of the reports. [139] GAO-03-108. [140] Bureau of Transportation Statistics, Source & Accuracy Compendium, http://www.bts.gov/statpol/SAcompendium.html (Washington, D.C.: Aug. 15, 2003). [141] The Federal Emergency Management Agency became part of the Department of Homeland Security in March 2003. [142] Formerly part of the Department of Health and Human Services, the Social Security Administration once again became an independent agency on March 31, 1995. Congress established the position of Chief Financial Officer within SSA in the Social Security Independence and Program Improvements Act of 1994. [143] GAO/GGD-00-204. [144] U.S. General Accounting Office, Social Security Administration: Actions Taken to Strengthen Procedures for Issuing Social Security Numbers to Noncitizens, but Some Weaknesses Remain, GAO-04-12 (Washington, D.C.: Oct. 15, 2003). GAO's Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548: