This is the accessible text file for GAO report number GAO-09-543T entitled 'Defense Acquisitions: Measuring the Value of DOD's Weapon Programs Requires Starting with Realistic Baselines' which was released on April 1, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Testimony: Before the Panel on Defense Acquisition Reform, Committee on Armed Services, House of Representatives: United States Government Accountability Office: GAO: For Release on Delivery: Expected at 7:30 a.m. EDT: Wednesday April 1, 2009: Defense Acquisitions: Measuring the Value of DOD's Weapon Programs Requires Starting with Realistic Baselines: Statement of Michael J. Sullivan: Director Acquisition and Sourcing Management: GAO-09-543T: GAO Highlights: Highlights of GAO-09-543T, a testimony before the House Committee on Armed Services Panel on Defense Acquisition Reform. Why GAO Did This Study: In 2008, the cumulative cost growth in the Department of Defense’s portfolio of 96 major defense acquisition programs was $296 billion and the average delay in delivering promised capabilities to the warfighter was 22 months. These poor outcomes mean that other critical defense and national priorities may go unfunded and that warfighters may go without the equipment they need to counter the changing threats that they face. GAO has examined extensively the issues in DOD’s weapon system programs from the perspective of best practices in product development, and believes that significant improvements are possible. Because the ability to measure knowledge, processes, and outcomes is critical to achieving these improvements, GAO has used metrics to review the management and health of these programs from within the framework of best practices. This testimony discusses: 1) “knowledge metrics,” used to determine how well programs manage technology, design, and manufacturing risks; 2) outcome metrics--concerning cost, schedule, and capability--that serve as “health indicators” of how well programs are being executed in terms of predicted outcomes; and 3) the prerequisites that GAO believes must be met in order for a program’s plans and goals to be realistic. What GAO Found: GAO employs a set of knowledge metrics to determine whether programs have attained the right knowledge at critical points over the course of a weapon system acquisition, and facilitate the identification of potential problems that could lead to cost, schedule, or performance shortfalls. In essence, knowledge supplants risk over time. Key knowledge points and metrics include 1) achieving a high level of technology maturity at the start of program development, 2) reaching design stability at the system-level critical design review, and 3) demonstrating that critical manufacturing processes are in control before starting production. By applying these metrics to selected programs in DOD’s 2008 portfolio of major defense acquisitions, GAO found that most programs have started system development without mature technologies and moved into system demonstration with low levels of design stability. GAO has determined that programs with immature technologies and unstable designs have experienced significant cost and schedule growth. Program outcome metrics—quantitative measures of cost, schedule, and performance over time—provide useful indicators of the health of acquisition programs and whether they are meeting their intended goals. When assessed regularly for changes and the reasons that cause changes, these indicators can be valuable tools for improving insight into and oversight of individual programs as well as DOD’s total portfolio of major defense acquisitions. The collective performance of the programs in DOD’s portfolio is a key indicator of how well the acquisition system generates the return on investment that it promises to the warfighter, Congress and taxpayers. GAO recently reported that outcome metrics for DOD’s 2008 major defense acquisition portfolio show worsening performance when compared to the department’s 2003 portfolio. For example, total acquisition costs for programs in the 2008 portfolio increased 25 percent from first estimates compared to a 19-percent increase for programs in the 2003 portfolio. DOD is working with GAO and the Office of Management and Budget to develop a comprehensive set of outcome metrics to better assess its portfolio of programs. While knowledge and outcome metrics provide valuable information about the potential problems and health of programs, they are of limited value if DOD does not do a better job ensuring acquisitions begin with realistic plans and baselines prior to development start. GAO believes there is a clear set of prerequisites that must be met by each program’s acquisition strategy before a measurement of the program’s health will be of real value. These prerequisites include: 1) establishing an evolutionary, knowledge-based business case for each acquisition; 2) separating technology development from product development; 3) limiting time and requirements for product development to manageable levels; 4) employing systems engineering early on in the process to arrive at realistic cost and schedule estimates; 5) committing to fully funding a program once it is approved; and 6) setting priorities from the top to ensure that candidate programs are truly needed and have a solid plan for delivery. View [hyperlink, http://www.gao.gov/products/GAO-09-543T] or key components. For more information, contact Michael J. Sullivan at (202) 512-4841 or sullivanm@gao.gov. [End of section] Mr. Chairman and Members of the Committee: I am pleased to be here this morning to discuss how best to measure and determine whether DOD's acquisition system is providing value to the warfighter. Earlier this week, we reported that the cumulative cost growth in DOD's portfolio of 96 major defense acquisition programs was $296 billion and the average delay in delivering promised capabilities to the warfighter was 22 months. These outcomes mean that other critical defense and national priorities go unfunded and warfighters go without the equipment they need to counter ever changing threats that they face. This condition is unacceptable. We believe that significant improvement in the acquisition of weapon systems is possible and that the ability to measure knowledge, processes, and outcomes is critical to achieving that improvement. It is important to note that not one single metric or set of metrics is enough to monitor acquisitions and gain efficiencies. Today, we would like to break our discussion about how to measure the department's acquisitions into 3 basic sections: * First, we would like to present a set of metrics that we refer to as "knowledge metrics" and use to determine how well acquisition programs are managing and retiring predictable technology, design, and manufacturing risks and gaining knowledge. These metrics are valuable because they can predict problems and identify causes. * Second, we would like to discuss a set of outcome measures-- concerning cost, schedule, and capability--that serve as health indicators. These indicators measure how well programs are being executed and achieving predicted outcomes in terms of meeting original baselines for cost, schedule, and performance. These metrics have intrinsic value as simple measurements, just as a thermometer can warn a parent that a child has a fever. * Third, there are certain indicators that we look for--based on the work we have done examining best practices for product development-- that are, perhaps, more important than these knowledge and health metrics because they determine from the outset how realistic the acquisition plans and strategies of programs are. For the sake of today's discussion, we will refer to them as "prerequisite indicators." These prerequisites are most important because we question the value of ANY metric when measuring from an unrealistic baseline. We know that the knowledge and program health metrics we use to measure programs' progress and outcomes are valuable when used in realistic, market-driven product development environments. We also know that ALL of these metrics are important indicators for decision makers. Our extensive body of work examining world-class enterprises and the way they operate has validated their value for programs that must deliver a new product to market at a certain time and within a certain investment cost or suffer significant consequences. These metrics work because they are measuring realistic plans and goals that are supported by doable requirements, accurate cost and schedule estimates, and stable funding. The company developing the products suffers dire consequences, such as loss of market share, if these programs do not succeed. This statement draws from our extensive body of work on DOD's acquisition of weapon systems. A list of our key products is provided at the end of this statement. This work was conducted in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. "Knowledge" Metrics Identify Potential Problems and Their Likely Causes: We have conducted a body of work that examines weapon acquisition issues from a perspective that draws upon lessons learned from best practices in product development. Collectively, these practices comprise a process that is anchored in knowledge. Achieving the right knowledge at the right time enables leadership to make informed decisions about when and how best to move into various expensive acquisition phases. In essence, knowledge supplants risk over time. This building of knowledge consists of gathering information about technology, design, and manufacturing at three critical points over the course of a weapon system program (Figure 1). We have developed valuable "knowledge metrics" that measure this knowledge build and allow us to identify potential problems that could lead to cost, schedule, or performance shortfalls and their likely causes. The metrics can be described as: * Knowledge Point 1, evidenced by the balance between a product's required capabilities and the resources available to meet them. Focus should be on understanding technological and design implications and achieving a high level of technology maturity at the start of system development. This means that the critical technologies needed to meet essential product requirements must be demonstrated to work in their intended environment. The technology readiness level for each critical technology is the metric we use to measure technology maturity. [Footnote 1] * Knowledge point 2, evidenced by the development of engineering prototypes and the completion of engineering drawings for an integrated product at the system design review. This metric provides tangible evidence that the product's design is stable, meaning it has a high probability of meeting customer requirements, as well as cost, schedule, and reliability targets. A best practice is to achieve design stability at the system-level critical design review, usually held midway through development. Completion of at least 90 percent of engineering drawings is the metric we use to measure design stability. * Knowledge point 3, evidenced by the demonstration that critical manufacturing processes are in control--that is, they are repeatable, sustainable, and capable of consistently producing parts within the product's quality tolerances and standards--at the start of production. One hundred percent of critical manufacturing processes in control is the metric we use to evaluate manufacturing maturity. Figure 1: Knowledge-Based Acquisition Process: [Refer to PDF for image: illustration] Material development start: Technology Development (A); PDR: Knowledge point 1: Metric: technology readiness levels at 7 or higher (B); Product Development: - Integration; - CDR: Knowledge point 2: Metric: 90 percent of engineering drawings complete; - Demonstration; Production: (C): Knowledge point 3: Metric: 100 percent of critical manufacturing processes in control. Source: GAO analysis of commercial best practices. [End of figure] Each of these metrics gauges the point when the requisite level of knowledge has been attained for a product in relation to where that product is in its development. World-class firms we have visited work hard to establish metrics such as these and their decision makers are required to consider them before deciding to advance a program to the next level. Theses types of metrics also help decision makers gauge progress in meeting cost, schedule, and performance goals and ensure that managers will (1) conduct activities to capture relevant product development knowledge, (2) provide evidence that this knowledge has been captured, and (3) hold decision reviews to determine that appropriate knowledge has been captured before moving the product to the next phase. The result is a product development process that provides critical measurements of knowledge, holds decision makers accountable, and delivers the expected results in a predictable manner. Attachment 1 to this statement provides a detailed list of activities that would provide program managers with the requisite technology, design, and manufacturing knowledge at key points in time during development. We have used these metrics to identify problems on major weapon system acquisition programs and have found a strong correlation between each of them and cost and schedule outcomes. For example, for 47 weapon programs in DOD's 2008 portfolio of major defense acquisitions, we assessed the knowledge attained at key decision points in the acquisition process and found the following:[Footnote 2] * Most programs have started system development without mature technologies. Only 4 of the 36 programs that provided data on technical maturity at development start did so with fully mature critical technologies. Further, only 14 of 39 programs that provided data have or plan to have demonstrated all of their technologies in a realistic environment prior to system-level critical design review, at which point the system's design should be stable. The 5 newer programs--those initiated since 2003[Footnote 3]--have higher levels of technology maturity, with all 5 programs demonstrating their technologies in a relevant environment prior to development start, in accordance with DOD and statutory criteria. However only 1of these programs met the best practice standard of demonstrating critical technologies in an operational environment. Last year, we determined that programs with immature technologies at the start of system development experienced 44 percent higher cost growth than programs that began with mature technologies. * Programs that have held design reviews in recent years reported higher levels of design knowledge. However, designs, on average, are still far from stable. For the 24 programs in our assessment that held a critical design review since 2003, the average percentage of total expected design drawings releasable at this review was 65 percent, compared to a best practice standard of 90 percent. We have found that programs moving forward into system demonstration with low levels of design stability are more likely than other programs to encounter costly design changes and parts shortages that, in turn, cause labor inefficiencies, schedule delays, and quality problems. Attachment 3 represents our notional depiction of the problems and outcomes that can typically be expected when these knowledge metrics are followed versus when they are not. Generally speaking, programs that move forward without retiring technology, design, and manufacturing risk at appropriate junctures will encounter a cascade of problems beginning with design changes and continuing with parts shortages, changes to manufacturing processes, labor inefficiencies, and quality problems. All of these problems delay programs and add to their development costs. We have found, for example, that a significant portion--about 70 percent--of the total development cost growth in programs typically occurs after the design review. Outcome Metrics Provide Insight into the Health and Performance of Individual Weapon System Programs and DOD's Total Portfolio: Program outcome metrics--quantitative measures of cost, schedule, and performance, and changes in these factors over time--provide useful indicators of the health of acquisition programs and facilitate analyses of how well programs are meeting cost, schedule, and performance goals. When assessed regularly for changes and the reasons that cause changes, such indicators can be valuable tools for improving insight and oversight of individual programs as well as DOD's total portfolio of major defense acquisitions. Over the years we have reported cost, schedule and performance data--good and bad--on numerous weapon systems. Our work continues to identify systemic and program- specific causes for cost, schedule, and performance problems and has led us to designate, since 1990, DOD's management of major weapon system acquisitions as a high risk area. To improve acquisition performance and address the factors that keep weapon acquisitions on the high risk list, DOD is working with us and the Office of Management and Budget to develop a comprehensive set of outcome metrics to provide better, comprehensive, and consistent measures of program cost and schedule performance. Last year, this cooperative effort resulted in agreement to track trends and changes in programs from their original baselines, from 5 years ago, and from the previous year, for the following data points: * Development cost; * Procurement cost; * Total program cost; * Quantities to be procured; * Procurement unit costs; * Total program unit costs; * Cycle time from Milestone B to Initial Operational Capability: DOD initiated a pilot study of 7 major defense programs to assess the adequacy of the proposed metrics and their value in analyzing performance, and the results proved promising. DOD approved the outcome metrics and intends to collect and report such data on an annual basis. Efforts to develop similar metrics on schedule performance continue. We believe that the metrics DOD plans to use are valuable for providing insight into the performance of weapon system programs. We have used similar metrics for many years in assessing programs. For example, we recently reported that ten of DOD's largest acquisition programs, commanding about half the overall acquisition dollars in the department's 2008 portfolio of major programs, have experienced significant cost growth and have seen quantities reduced by almost a third (see table 1). The two largest programs--the Joint Strike Fighter and the Future Combat System--represent significant cost risk moving forward and will dominate the portfolio for years. Since these programs consume such a large portion of the funding that DOD spends on research and development and procurement, their performance also affects other major weapon acquisitions, smaller acquisition programs, and DOD's ability to fund and acquire other supplies and equipment as well. Table 1: Changes in Costs and Quantities for Ten of the Highest Cost Acquisition Programs: Fiscal year 2009 dollars in millions: Program: Joint Strike Fighter; Total cost: First full estimate: $206,410; Total cost: Current estimate: $244,772; Total quantity: First full estimate: 2,866; Total quantity: Current estimate: 2,456; Acquisition unit cost: Percentage change: 38. Program: Future Combat System; Total cost: First full estimate: $89,776; Total cost: Current estimate: $129,731; Total quantity: First full estimate: 15; Total quantity: Current estimate: 15; Acquisition unit cost: Percentage change: 45. Program: Virginia Class Submarine; Total cost: First full estimate: $58,378; Total cost: Current estimate: $81,556; Total quantity: First full estimate: 30; Total quantity: Current estimate: 30; Acquisition unit cost: Percentage change: 40. Program: F-22A Raptor; Total cost: First full estimate: $88,134; Total cost: Current estimate: $73,723; Total quantity: First full estimate: 648; Total quantity: Current estimate: 184; Acquisition unit cost: Percentage change: 195. Program: C-17 Globemaster III; Total cost: First full estimate: $51,733; Total cost: Current estimate: $73,571; Total quantity: First full estimate: 210; Total quantity: Current estimate: 190; Acquisition unit cost: Percentage change: 57. Program: V-22 Joint Services Advanced Vertical Lift Aircraft; Total cost: First full estimate: $38,726; Total cost: Current estimate: $55,544; Total quantity: First full estimate: 913; Total quantity: Current estimate: 458; Acquisition unit cost: Percentage change: 186. Program: F/A-18E/F Super Hornet; Total cost: First full estimate: $78,925; Total cost: Current estimate: $51,787; Total quantity: First full estimate: 1,000; Total quantity: Current estimate: 493; Acquisition unit cost: Percentage change: 33. Program: Trident II Missile; Total cost: First full estimate: $49,939; Total cost: Current estimate: $49,614; Total quantity: First full estimate: 845; Total quantity: Current estimate: 561; Acquisition unit cost: Percentage change: 50. Program: CVN 21 Nuclear Aircraft Class Carrier; Total cost: First full estimate: $34,360; Total cost: Current estimate: $29,914; Total quantity: First full estimate: 3; Total quantity: Current estimate: 3; Acquisition unit cost: Percentage change: -13. Program: P-8A Poseidon Multi-mission Maritime Aircraft; Total cost: First full estimate: $29,974; Total cost: Current estimate: $29,622; Total quantity: First full estimate: 115; Total quantity: Current estimate: 113; Acquisition unit cost: Percentage change: 1. Source: GAO analysis of DOD data. [End of table] While program outcome metrics are good measures of individual program performance, the collective performance of DOD's portfolio of major defense acquisition programs is a key indicator of how well the department's acquisition system generates the return on investment it promises to the warfighter, Congress, and the taxpayer. Portfolio metrics also provide senior leaders and Congress with a snapshot of the cumulative impact of current investment decisions and poor program performance on future budgets. In our annual assessment of selected weapon programs, we analyzed the performance of DOD programs at the portfolio level by comparing programs' initial cost, schedule, and quantity estimates to their current estimates, based on data obtained from the Selected Acquisition Reports. This year's cumulative results, reported earlier this week, [Footnote 4] are shown in table 2. Table 2: Analysis of DOD Major Defense Acquisition Program Portfolios[A]: Fiscal year 2009 dollars. Portfolio size: Number of programs; Fiscal Year 2003: 77; Fiscal Year 2007: 95; Fiscal Year 2008: 96. Portfolio size: Total planned commitments; Fiscal Year 2003: $1.2 trillion; Fiscal Year 2007: $1.6 trillion; Fiscal Year 2008: $1.6 trillion. Portfolio size: Commitments outstanding; Fiscal Year 2003: $724.2 billion; Fiscal Year 2007: $875.2 billion; Fiscal Year 2008: $786.3 billion. Portfolio indicators: Change to total RDT&E costs from first estimate; Fiscal Year 2003: 37 percent; Fiscal Year 2007: 40 percent; Fiscal Year 2008: 42 percent. Portfolio indicators: Change to total acquisition cost from first estimate; Fiscal Year 2003: 19 percent; Fiscal Year 2007: 26 percent; Fiscal Year 2008: 25 percent. Portfolio indicators: Total acquisition cost growth; Fiscal Year 2003: $183 billion; Fiscal Year 2007: $301.3 billion[B]; Fiscal Year 2008: $296.4 billion. Portfolio indicators: Share of programs with 25 percent increase in program acquisition unit cost growth; Fiscal Year 2003: 41 percent; Fiscal Year 2007: 44 percent; Fiscal Year 2008: 42 percent. Portfolio indicators: Average schedule delay in delivering initial capabilities; Fiscal Year 2003: 18 months; Fiscal Year 2007: 21 months; Fiscal Year 2008: 22 months. Source: GAO analysis of DOD data. [A] Data were obtained from DOD's Selected Acquisition Reports (dated December 2002, 2006, and 2007). In a few cases data were obtained directly from program offices. The number of programs reflects the programs with Selected Acquisition Reports; however, in our analysis we have broken a few Selected Acquisition Reports programs into smaller elements or programs. Not all programs had comparable cost and schedule data and these programs were excluded from the analysis where appropriate. Portfolio performance data do not include costs of developing Missile Defense Agency elements or the DIMHRS program. [B] The acquisition cost growth for the 2007 portfolio was $295 billion in 2008 constant dollars. [End of table] Analyzing the data and comparing metrics from different time periods provides unique insights into the relative health of the portfolio and trends. From 2003 to 2008, the number of programs in DOD's major defense acquisition portfolio has grown from 77 to 96. Total costs for these programs now total $1.6 trillion with almost one-half of this amount still to be spent. Outcome metrics for 2008 show worsening performance in all categories compared to the 2003 portfolio and mixed performance--some better, some worse--compared to the 2007 data. While DOD is committing substantially more investment dollars to developing and procuring new weapon systems, the total acquisition costs for the 2008 portfolio has grown by $296 billion over initial estimates and the average schedule delay in delivering capabilities to the warfighter averages 22 months. Implications for the future are obvious. Continued cost growth reduces DOD's buying power and results in less funding being available for other DOD priorities and programs. As program costs increase, DOD must request more funding to cover overruns, make trade- offs with existing programs, delay the start of new programs, take funds from other accounts, or reduce procurement quantities. Continued failure to deliver weapon systems on time delays providing critical capabilities to the warfighter and results in operating costly legacy systems longer than expected, finding alternatives to fill capability gaps, or going completely without the capability. Key Prerequisites Need to Be Met to Improve the Value of Metrics and Achieve Better Acquisition Program Outcomes: While the metrics discussed above can provide valuable knowledge about potential problems and additional information on the health of DOD's acquisition programs, metrics alone may not be sufficient if the department does not do a better job ensuring that acquisitions begin with realistic plans and baseline estimates for cost and schedules prior to development start. We believe there is a clear set of prerequisites that must be a part of any acquisition strategy before any measurement of the acquisition's health can be valuable. Otherwise, metrics measured against unrealistic estimates will do no good. These key prerequisites for obtaining realistic baselines include: * Establishing a clear, knowledge-based, evolutionary business case for the product. This business case must: validate that a need exists; determine that resources are available to develop a product that will meet the need; determine that the product developer has a knowledge- based plan and strategy to deliver the product; establish reasonable estimates for cost, delivery time and quantities; and ensure available funding for the product. All of these elements of the business case should also be agreed upon by major stakeholders across the requirements, funding, acquisition, and warfighting communities. * Separating technology development activities from product development activities. The process of developing technology culminates in discovery--the gathering of knowledge--and must, by its nature, allow room for unexpected results and delays. Leading firms do not ask their product managers to develop technology because they have learned the hard way that invention cannot be scheduled. When immature technologies are brought onto the critical path of product development programs too early, they often cause long delays in an environment where large workforces must be employed, complex tools, plants, and facilities must be operated, long and expensive supplier networks must be paid, and the product itself must sometimes be redesigned once the final form of the technologies is known. Successful programs give responsibility for maturing technologies to science and technology organizations, rather than the program or product development managers, because the science and technology environment is less costly. We have recommended in the past that DOD's risks should be taken in the science and technology arena and that more funding should be made available to this process to do so. * Limiting time and requirements for product development to manageable levels. Product developers should strive to deliver the best available capabilities within realistic timeframes and should expect to continue to develop new capabilities when they are technologically feasible. By limiting product development cycle times to 6 years or less, DOD could assimilate new technologies into weapon systems more frequently, accelerate delivery of new technology to the warfighter, hold program managers accountable, and make more frequent and predictable work in production, where contractors and the industrial base can profit by being efficient. Too many major acquisitions currently take the opposite approach by seeking to deliver a revolutionary "big bang" capability in one step. This means that programs are more risky, delivery takes as long as 15 years in some cases, and costs grow at exponential rates from the original baseline due to the risky nature of the acquisition strategy. We point to the private sector and some past defense acquisitions, such as the F-16 program, as models for this practice. * Employing early systems engineering discipline in order to develop realistic cost and schedule estimates prior to development start. Early systems engineering provides the knowledge a product developer needs to identify and resolve performance and resource gaps before product development begins either by reducing requirements, deferring them to the future, or increasing the estimated cost for the weapon system's development. Requirements that are too risky given the state of technology and design should not be allowed into this expensive environment. * Making a commitment to fully fund programs once they are approved. This would require the department to ensure that it does not have too many programs underway given the amount of available resources. * Setting priorities from the top to ensure that candidate programs are truly needed and have a solid plan for delivery. DOD will continue to experience poor acquisition outcomes until it begins making choices that reflect the most important needs of the joint warfighter and match requirements with available resources. The urge to accept all candidate programs and to go for the "big bang" capability without the knowledge to achieve it should be resisted. Only the best candidates--defined in terms of priorities, resource availability, and executability--should be approved. There is no doubt that the current state of the department's acquisition process is too expensive for the taxpayer and not timely enough for the warfighter. The following illustration reinforces this point. Figure 2: Cost Remaining Versus Annual Appropriations for Major Defense Acquisitions: [Refer to PDF for image: multiple line graph] Fiscal year: 1992; Annual RDTE and Procurement Appropriations: $100 billion; Costs Remaining for Major Defense Acquisitions: $324 billion. Fiscal year: 1993; Annual RDTE and Procurement Appropriations: $91 billion; Costs Remaining for Major Defense Acquisitions: $271 billion. Fiscal year: 1994; Annual RDTE and Procurement Appropriations: $79 billion; Costs Remaining for Major Defense Acquisitions: $253 billion. Fiscal year: 1995; Annual RDTE and Procurement Appropriations: $78 billion; Costs Remaining for Major Defense Acquisitions: $341 billion. Fiscal year: 1996; Annual RDTE and Procurement Appropriations: $78 billion; Costs Remaining for Major Defense Acquisitions: $345 billion. Fiscal year: 1997; Annual RDTE and Procurement Appropriations: $79 billion; Costs Remaining for Major Defense Acquisitions: $331 billion. Fiscal year: 1998; Annual RDTE and Procurement Appropriations: $82 billion; Costs Remaining for Major Defense Acquisitions: $280 billion. Fiscal year: 1999; Annual RDTE and Procurement Appropriations: $89 billion; Costs Remaining for Major Defense Acquisitions: $233 billion. Fiscal year: 2000; Annual RDTE and Procurement Appropriations: $94 billion; Costs Remaining for Major Defense Acquisitions: $309 billion. Fiscal year: 2001; Annual RDTE and Procurement Appropriations: $104 billion; Costs Remaining for Major Defense Acquisitions: $234 billion. Fiscal year: 2002; Annual RDTE and Procurement Appropriations: $111 billion; Costs Remaining for Major Defense Acquisitions: $586 billion. Fiscal year: 2003; Annual RDTE and Procurement Appropriations: $137 billion; Costs Remaining for Major Defense Acquisitions: $579 billion. Fiscal year: 2004; Annual RDTE and Procurement Appropriations: $148 billion; Costs Remaining for Major Defense Acquisitions: $722 billion. Fiscal year: 2005; Annual RDTE and Procurement Appropriations: $165 billion; Costs Remaining for Major Defense Acquisitions: $745 billion. Fiscal year: 2006; Annual RDTE and Procurement Appropriations: $157 billion; Costs Remaining for Major Defense Acquisitions: $841 billion. Fiscal year: 2007; Annual RDTE and Procurement Appropriations: $160 billion; Costs Remaining for Major Defense Acquisitions: $767 billion. Source: DOD (data); GAO (analysis and presentation). [End of figure] This figure depicts an investment strategy for major weapon systems that continues to increase the costs to develop our existing weapons well into the future while the funding available to retire those costs appears capped at a very low level. While costs continue to rise as the result of more and more risky programs being added to the portfolio, our ability to allocate funds for these costs appears to be, at best, capped at very low percentages of the total cost. We could measure the risk of these acquisitions much better than we have in the past if we set the appropriate prerequisites for their initiation, measure the knowledge that must be in place at various points, and continue to monitor their health in terms of cost and schedule. Concluding Remarks: Measuring the performance of weapon system programs both individually and collectively is critical for determining whether the warfighter and the taxpayer are receiving the promised return on investment. No single metric, however, can capture the whole picture of how well programs are performing. It is important to look at knowledge and outcome metrics. Knowledge metrics provide key information for determining whether programs have the requisite knowledge to move from one phase of development to the next and are at risk of cost and schedule overruns. Outcomes metrics are also needed to provide temperature checks on the health and status of individual programs and the portfolio of programs as a whole. These metrics are vital for informing program decision making and helping to manage programs. Metrics by themselves do not solve problematic acquisitions. Ultimately, DOD still needs to do a better job planning and executing programs to achieve better outcomes. Critical to achieving successful outcomes is establishing knowledge-based, realistic program baselines. Without realistic baselines, there is no foundation for accurately measuring the knowledge and health of programs. Over the past several years, our work has highlighted a number of underlying causes for why DOD does not effectively manage the acquisition of weapon system programs. DOD recently revised its acquisition policy to provide a better foundation for developing weapon systems, however, reform will not be achieved without fundamental changes to the overall acquisition culture and environment that exists in DOD. I would be pleased to discuss these causes and issues with the Committee at a future time. Mr. Chairman, this concludes my prepared statement. I would be happy to answer any questions you may have at this time. Contacts and Acknowledgements: For further information about this statement, please contact Michael J. Sullivan (202) 512-4841 or sullivanm@gao.gov. Contact points for our Office of Congressional Relations and Public Affairs may be found on the last page of this statement. Individuals who made key contributions to this statement include Cheryl Andrew, Ridge Bowman, Bruce Fairbairn, Susan Neill, John Oppenheim, and Ron Schwenn. [End of section] Attachment 1: Knowledge-Based Activities: Knowledge Point 1: Start of product development activities: Best practice metric: Technology readiness level 7 (indicating technologies work in an operational environment): * Demonstrate technologies to high readiness levels. * Ensure that requirements for the product increment are informed by preliminary design using systems engineering process (such as prototyping of preliminary design). * Establish cost and schedule estimates for product on the basis of knowledge from preliminary design using system engineering tools (such as prototyping of preliminary design). * Constrain development phase (5 to 6 years or less) for incremental development. * Ensure development phase fully funded (programmed in anticipation of milestone). * Align program manager tenure to complete development phase. * Contract strategy that separates system integration and system demonstration activities. * Conduct independent cost estimate. * Conduct independent program assessment. * Conduct major milestone decision review for development start. Knowledge Point 2: Transition from system integration to system demonstration activities; Best practice metric: 90 percent of design drawings are complete by the critical design review: * Complete system critical design review. * Complete 90 percent of engineering design drawing packages. * Complete subsystem and system design reviews. * Demonstrate with system integration prototype that design meets requirements. * Complete the failure modes and effects analysis. * Identify key system characteristics. * Identify critical manufacturing processes. * Establish reliability targets and growth plan on the basis of demonstrated reliability rates of components and subsystems. * Conduct independent cost estimate. * Conduct independent program assessment. * Conduct major milestone decision review to enter system demonstration. Knowledge Point 3: Initiation of producing a product to be delivered to customer; Best practice metric: 100 percent of critical manufacturing processes are in control: * Demonstrate manufacturing processes. * Build and test production-representative prototypes to demonstrate product in operational environment. * Test production-representative prototypes to achieve reliability goal. * Collect statistical process control data. * Demonstrate that critical processes are capable and in statistical control. * Independent cost estimate. * Independent program assessment. * Conduct major milestone decision review to begin production. Source: GAO analysis of commercial best practices. [End of table] [End of section] Attachment 2: Technology Readiness Levels: Technology readiness level: 1. Basic principles observed and reported; Description: Lowest level of technology readiness. Scientific research begins to be translated into applied research and development. Examples might include paper studies of a technology's basic properties; Hardware: Software: None (paper studies and analysis); Demonstration Environment: None. Technology readiness level: 2. Technology concept and/or application formulated; Description: Invention begins. Once basic principles are observed, practical applications can be invented. The application is speculative and there is no proof or detailed analysis to support the assumption. Examples are still limited to paper studies; Hardware: Software: None (paper studies and analysis); Demonstration Environment: None. Technology readiness level: 3. Analytical and experimental critical function and/or characteristic proof of concept; Description: Active research and development is initiated. This includes analytical studies and laboratory studies to physically validate analytical predictions of separate elements of the technology. Examples include components that are not yet integrated or representative; Hardware: Software: Analytical studies and demonstration of nonscale individual components (pieces of subsystem); Demonstration Environment: Lab. Technology readiness level: 4. Component and/or breadboard. Validation in laboratory environment; Description: Basic technological components are integrated to establish that the pieces will work together. This is relatively "low fidelity" compared to the eventual system. Examples include integration of "ad hoc" hardware in a laboratory; Hardware: Software: Low fidelity breadboard; Integration of nonscale components to show pieces will work together. Not fully functional or form or fit but representative of technically feasible approach suitable for flight articles; Demonstration Environment: Lab. Technology readiness level: 5. Component and/or breadboard validation in relevant environment; Description: Fidelity of breadboard technology increases significantly. The basic technological components are integrated with reasonably realistic supporting elements so that the technology can be tested in a simulated environment. Examples include "high fidelity" laboratory integration of components; Hardware: Software: High fidelity breadboard; Functionally equivalent but not necessarily form and/or fit (size weight, materials, etc). Should be approaching appropriate scale. May include integration of several components with reasonably realistic support elements/ subsystems to demonstrate functionality; Demonstration Environment: Lab demonstrating functionality but not form and fit. May include flight demonstrating breadboard in surrogate aircraft. Technology ready for detailed design studies. Technology readiness level: 6. System/subsystem model or prototype demonstration in a relevant environment; Description: Representative model or prototype system, which is well beyond the breadboard tested for TRL 5, is tested in a relevant environment. Represents a major step up in a technology's demonstrated readiness. Examples include testing a prototype in a high fidelity laboratory environment or in simulated realistic environment; Hardware: Software: Prototype; Should be very close to form, fit and function. Probably includes the integration of many new components and realistic supporting elements/subsystems if needed to demonstrate full functionality of the subsystem; Demonstration Environment: High-fidelity lab demonstration or limited/ restricted flight demonstration for a relevant environment. Integration of technology is well defined. Technology readiness level: 7. System prototype demonstration in a realistic environment; Description: Prototype near or at planned operational system. Represents a major step up from TRL 6, requiring the demonstration of an actual system prototype in a realistic environment, such as in an aircraft, vehicle or space. Examples include testing the prototype in a test bed aircraft; Hardware: Software: Prototype. Should be form, fit and function integrated with other key supporting elements/subsystems to demonstrate full functionality of subsystem; Demonstration Environment: Flight demonstration in representative realistic environment such as flying test bed or demonstrator aircraft; Technology is well substantiated with test data. Technology readiness level: 8. Actual system completed and "flight qualified" through test and demonstration; Description: Technology has been proven to work in its final form and under expected conditions. In almost all cases, this TRL represents the end of true system development. Examples include developmental test and evaluation of the system in its intended weapon system to determine if it meets design specifications; Hardware: Software: Flight qualified hardware; Demonstration Environment: Developmental Test and Evaluation (DT&E) in the actual system application. Technology readiness level: 9. Actual system "flight proven" through successful mission operations; Description: Actual application of the technology in its final form and under mission conditions, such as those encountered in operational test and evaluation. In almost all cases, this is the end of the last "bug fixing" aspects of true system development. Examples include using the system under operational mission conditions; Hardware: Software: Actual system in final form; Demonstration Environment: Operational Test and Evaluation (OT&E) in operational mission conditions. Source: GAO and its analysis of National Aeronautics and Space Administration data. [End of table] [End of section] Attachment 3: Notional Illustration Showing the Different Paths That a Product's Development Can Take: [Refer to PDF for image: illustration] This illustration is a graph plotting schedule/time against cost/investment. Two paths are shown: 1) Mature product and; 2) Continual changes needed to reach maturity. Mature product: takes less time and has a lower cost due to: * stable design; * Demonstrated; * Built and tested; * Manufacturing processes in control. Continual changes needed to reach maturity: takes more time and has a greater cost due to: * Unstable design; * Late drawings and design changes; * Tooling and design changes (Fully integrated aircraft not built or tested); * Production manufacturing process not in control; * Labor inefficiencies and quality issues. Source: GAO. [End of figure] [End of section] Related GAO Products: Defense Acquisitions: Assessment of Major Weapon Programs. [hyperlink, http://www.gao.gov/products/GAO-09-326SP]. Washington, D.C.: March 30, 2009. Defense Acquisitions: DOD Must Prioritize Its Weapon System Acquisitions and Balance Them with Available Resources. [hyperlink, http://www.gao.gov/products/GAO-09-501T]. Washington, D.C.: March 18, 2009. Defense Acquisitions: DOD Must Balance Its Needs with Available Resources and Follow an Incremental Approach to Acquiring Weapon Systems. [hyperlink, http://www.gao.gov/products/GAO-09-431T]. Washington, D.C.: March 3, 2009. Defense Acquisitions: Perspectives on Potential Changes to DOD's Acquisition Management Framework. [hyperlink, http://www.gao.gov/products/GAO-09-295R]. Washington, D.C.: February 27, 2009. Defense Management: Actions Needed to Overcome Long-standing Challenges with Weapon Systems Acquisition and Service Contract Management. [hyperlink, http://www.gao.gov/products/GAO-09-362T]. Washington, D.C.: February 11, 2009. Defense Acquisitions: Fundamental Changes Are Needed to Improve Weapon Program Outcomes. [hyperlink, http://www.gao.gov/products/GAO-08-1159T]. Washington, D.C.: September 25, 2008. Defense Acquisitions: DOD's Requirements Determination Process Has Not Been Effective in Prioritizing Joint Capabilities. [hyperlink, http://www.gao.gov/products/GAO-08-1060]. Washington, D.C.: September 25, 2008. Defense Acquisitions: A Knowledge-Based Funding Approach Could Improve Major Weapon System Program Outcomes. [hyperlink, http://www.gao.gov/products/GAO-08-619]. Washington, D.C.: July 2, 2008. Defense Acquisitions: Better Weapon Program Outcomes Require Discipline, Accountability, and Fundamental Changes in the Acquisition Environment. [hyperlink, http://www.gao.gov/products/GAO-08-782T]. Washington, D.C.: June 3, 2008. Defense Acquisitions: Assessments of Selected Weapon Programs. [hyperlink, http://www.gao.gov/products/GAO-08-467SP]. Washington, D.C.: March 31, 2008. Best Practices: Increased Focus on Requirements and Oversight Needed to Improve DOD's Acquisition Environment and Weapon System Quality. [hyperlink, http://www.gao.gov/products/GAO-08-294]. Washington, D.C.: February 1, 2008. Cost Assessment Guide: Best Practices for Estimating and Managing Program Costs. [hyperlink, http://www.gao.gov/products/GAO-07-1134SP]. Washington, D.C.: July 2007. Defense Acquisitions: Assessments of Selected Weapon Programs. [hyperlink, http://www.gao.gov/products/GAO-07-406SP.Washington, D.C.: March 30, 2007. Best Practices: An Integrated Portfolio Management Approach to Weapon System Investments Could Improve DOD's Acquisition Outcomes. [hyperlink, http://www.gao.gov/products/GAO-07-388]. Washington, D.C.: March 30, 2007. Defense Acquisitions: Major Weapon Systems Continue to Experience Cost and Schedule Problems under DOD's Revised Policy. [hyperlink, http://www.gao.gov/products/GAO-06-368]. Washington, D.C.: April 13, 2006. DOD Acquisition Outcomes: A Case for Change. [hyperlink, http://www.gao.gov/products/GAO-06-257T]. Washington, D.C.: November 15, 2005. Best Practices: Better Support of Weapon System Program Managers Needed to Improve Outcomes. [hyperlink, http://www.gao.gov/products/GAO-06-110]. Washington, D.C.: November 1, 2005. Defense Acquisitions: Stronger Management Practices Are Needed to Improve DOD's Software-Intensive Weapon Acquisitions. [hyperlink, http://www.gao.gov/products/GAO-04-393]. Washington, D.C.: March 1, 2004. Defense Acquisitions: DOD's Revised Policy Emphasizes Best Practices, but More Controls Are Needed. [hyperlink, http://www.gao.gov/products/GAO-04-53]. Washington, D.C.: November 10, 2003. Best Practices: Setting Requirements Differently Could Reduce Weapon Systems' Total Ownership Costs. [hyperlink, http://www.gao.gov/products/GAO-03-57. Washington, D.C.: February 11, 2003. Defense Acquisitions: Factors Affecting Outcomes of Advanced Concept Technology Demonstration. [hyperlink, http://www.gao.gov/products/GAO-03-52]. Washington, D.C.: December 2, 2002. Best Practices: Capturing Design and Manufacturing Knowledge Early Improves Acquisition Outcomes. [hyperlink, http://www.gao.gov/products/GAO-02-701]. Washington, D.C.: July 15, 2002. Defense Acquisitions: DOD Faces Challenges in Implementing Best Practices. [hyperlink, http://www.gao.gov/products/GAO-02-469T]. Washington, D.C.: February 27, 2002. [End of section] Footnotes: [1] Technology readiness levels, originally developed by the National Aeronautics and Space Administration, are measured on a scale of 1 to 9, beginning with paper studies of a technology's feasibility and culminating with a technology fully integrated into a completed product. See Attachment 2 for the definitions of technology readiness levels. [2] We did this by collecting data directly from program offices using a questionnaire. These programs are primarily in development and, therefore, most relevant to current decisions about which programs should receive substantial investments of research and development funding now and large amounts of procurement funding in the future. Defense Acquisitions: Assessment of Selected Weapons Programs. [hyperlink, http://www.gao.gov/products/GAO-09-326SP]. Washington, D.C.: March 30, 2009. [3] In 2003, DOD revised its primary acquisition policy to state that technologies should be demonstrated in a relevant environment prior to starting an acquisition program. In 2006, this standard became a statutory requirement for all major defense acquisition programs in the National Defense Authorization Act for Fiscal Year 2006, Pub. L. No. 109-163, § 801, codified at 10 U.S.C. § 2366b. [4] GAO, Defense Acquisitions: Assessments of Selected Weapon Programs, [hyperlink, http://www.gao.gov/products/GAO-09-326SP] (Washington, D.C.: Mar. 30, 2009). [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO’s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO’s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548: