This is the accessible text file for GAO report number GAO-11-53 entitled 'DOD Business Transformation: Improved Management Oversight of Business System Modernization Efforts Needed' which was released on October 8, 2010. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: United States Government Accountability Office: GAO: October 2010: DOD Business Transformation: Improved Management Oversight of Business System Modernization Efforts Needed: GAO-11-53: GAO Highlights: Highlights of GAO-11-53, a report to congressional requesters. Why GAO Did This Study: The Department of Defense (DOD) invests billions of dollars annually to modernize its business systems, which have been on GAO’s high-risk list since 1995. DOD is in the process of implementing nine enterprise resource planning (ERP) efforts which perform business-related tasks such as general ledger accounting and supply chain management. These efforts are essential to transforming DOD’s business operations. GAO was asked to (1) provide the status of the ERPs as of December 31, 2009; (2) determine whether selected ERPs followed schedule and cost best practices; and (3) determine if DOD has defined the performance measures to assess whether the ERPs will meet their intended business capabilities. To accomplish these objectives, GAO reviewed data on the status of each ERP from the program management officers and interviewed the DOD and military departments’ chief management officers. What GAO Found: Based upon the data provided by DOD, six of the nine ERPs have experienced schedule delays ranging from 2 to 12 years and five have incurred cost increases ranging from $530 million to $2.4 billion. DOD has stated that the ERPs will replace over 500 legacy systems that cost hundreds of millions of dollars to operate annually. However, delays in implementing the ERPs require DOD to fund the legacy systems longer than anticipated, thereby reducing the funds available for other DOD priorities. In 2007, 2008, and 2009, GAO made 19 recommendations to improve the management of DOD’s ERP efforts. While DOD agreed with the recommendations, 14 have not yet been fully implemented. GAO analyzed four of the nine ERPs to determine whether scheduling and cost estimating best practices were being followed. Regarding scheduling practices, GAO found that none of the programs had developed a fully integrated master schedule as an effective tool to help in the management of the programs. A reliable schedule is crucial to estimating the overall schedule and cost of a program. Without a reliable schedule, DOD is unable to predict, with any degree of confidence, if the estimated completion dates are realistic. Regarding the cost estimates, GAO found that although the four ERPs’ cost estimates generally met the criteria for three of the four best practices—well-documented, accurate, and comprehensive—three ERPs did not fully meet the credibility criteria because potential limitations were not discussed. More specifically, the three ERPs lacked a sensitivity analysis or a risk and uncertainty analysis as stipulated in GAO, Office of Management and Budget, and DOD guidance, thus diminishing the credibility of the estimates. While the ERPs are critical to transforming DOD’s business operations, DOD lacks a comprehensive set of performance measures to assess these systems and their contribution to transforming business operations. Management needs to define what constitutes a successful implementation in terms that can be used to assess whether the system is (1) being used as expected and (2) providing the intended benefits. Accordingly, the actual measures used to accomplish these objectives will differ depending on the system. For example, measures for a logistical system may focus on reducing inventory levels, while those for a financial system may focus on reducing prompt payment penalties. Without performance measures to evaluate how well the ERPs are accomplishing their intended goals, DOD decision makers do not have all the information they need to determine whether DOD investments are accomplishing their desired goals, and program managers do not have the information they need to ensure that their individual program is helping DOD to achieve business transformation and thereby improve upon its primary mission of supporting the warfighter. What GAO Recommends: In addition to reiterating its existing recommendations, GAO is making eight recommendations to the Secretary of Defense aimed at improving schedule and cost practices and the development of performance measures to evaluate whether the ERPs’ intended goals are being accomplished. DOD concurred with our recommendations and plans to take action to implement them. View [hyperlink, http://www.gao.gov/products/GAO-11-53] or key components. For more information, contact Asif A. Khan at (202) 512- 9095 or khana@gao.gov. [End of section] Contents: Letter: Background: Status of DOD's ERP Implementation Efforts: DOD Did Not Follow Key Best Practices for Estimating ERP Schedules and Cost, Resulting in Unreliable Estimates: ERP Success in Transforming Business Operations Has Not Been Defined or Measured: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Objective, Scope, and Methodology: Appendix II: Comments from the Department of Defense: Appendix III: Status of DOD's Actions on Previous GAO Recommendations Related to Business Systems Modernization: Appendix IV: Assessments of Four DOD ERP Programs' Integrated Master Schedules: Appendix V: Assessments of Four DOD ERP Program Cost Estimates: Appendix VI: GAO Contacts and Staff Acknowledgments: Tables: Table 1: Reported Full Deployment Schedule Slippage for Each ERP as of December 31, 2009: Table 2: Reported Original and Current Life-Cycle Cost Estimate for Each ERP as of December 31, 2009: Table 3: Defense Agencies' Scheduled Implementation of DAI: Table 4: Extent to Which Program Schedules Met Best Practices: Table 5: Extent Cost Estimates Met Best Practices: Table 6: Status of DOD's Actions to Address GAO Recommendations in GAO- 07-860: Table 7: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-822: Table 8: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-866: Table 9: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-896: Table 10: Status of DOD's Actions to Address GAO Recommendations in GAO-09-841: Table 11: Analysis of the Air Force's DEAMS Program Schedule: Table 12: Analysis of the Air Force's ECSS Solutions Development Project Schedule: Table 13: Analysis of the Air Force's ECSS Reports, Interfaces, Conversions, and Extensions (RICE) Program Schedule: Table 14: Analysis of the Army's GFEBS Program Schedule: Table 15: Analysis of the Army's GCSS-Army Program Schedule: Table 16: The 12 Steps of High-Quality Cost Estimating, Mapped to the Steps of a High-Quality Cost Estimate: Table 17: Analysis of the Air Force's DEAMS Cost Estimate: Table 18: Analysis of the Air Force's ECSS Cost Estimate: Table 19: Analysis of the Army's GFEBS Cost Estimate: Table 20: Analysis of the Army's GCSS-Army Cost Estimate: Figure: Figure 1: DOD's Fiscal Year 2011 Business Systems Budget Request by DOD Components (Dollars in Thousands): Abbreviations: ATEC: Army Test and Evaluation Command: BPR: business process reengineering: BSM: Business System Modernization: BTA: Business Transformation Agency: CARD: Cost Analysis Requirements Document: COTS: commercial off-the-shelf: DAI: Defense Agencies Initiative: DBSMC: Defense Business Systems Management Committee: DCMO: deputy chief management officer: DEAMS: Defense Enterprise Accounting and Management System: DID: Data Item Description: DIMHRS: Defense Integrated Military Human Resources System: DLA: Defense Logistics Agency: DOD: Department of Defense: EBS: Enterprise Business System: ECSS: Expeditionary Combat Support System: ERAM: Enterprise Risk Assessment Methodology: ERP: enterprise resource planning: EVM: earned value management: FFP: firm-fixed price: GCSS-Army: Global Combat Support System-Army: GCSS-MC: Global Combat Support System-Marine Corps: GFEBS: General Fund Enterprise Business System: IMS: integrated master schedule: IOT&E: initial operational test and evaluation: IPPS-A: Integrated Personnel and Pay System-Army: IRB: investment review board: IT: information technology: IUID: item-unique identification: IV&V: independent verification and validation: LMP: Logistics Modernization Program: MAIS: major automated information system: MDA: Milestone Decision Authority: MDAP: major defense acquisition program: MSO: Must Start On: NAVAIR: Naval Air Systems Command: Navy ERP: Navy Enterprise Resource Planning: NTC: National Training Center: OMB: Office of Management and Budget: PCA: pre-certification authority: PMO: program management office: RICE: reports, interfaces, conversions, and extensions: SFIS: Standard Financial Information Structure: TAV: total asset visibility: [End of section] United States Government Accountability Office: Washington, DC 20548: October 7, 2010: Congressional Requesters: The Department of Defense's (DOD) business systems[Footnote 1] modernization program has been on our high-risk list[Footnote 2] since 1995 because of the size, complexity, and significance of the related efforts. DOD's business systems modernization entails investments in and the implementation of comprehensive, integrated business systems for managing an organization's resources, commonly referred to as enterprise resource planning (ERP)[Footnote 3] systems and the elimination of hundreds of legacy systems. DOD officials have said that successful implementation of ERPs is key to resolving the long- standing weaknesses in the department's business operations in areas such as business transformation, financial management, and supply chain management,[Footnote 4] and improving the department's capability to provide DOD management and the Congress with accurate and reliable information on the results of its operations. DOD has identified 10 ERPs,[Footnote 5] 1 of which has been fully implemented, as essential to its efforts to transform its business operations. According to DOD, as of December 2009, it had invested approximately $5.8 billion to develop and implement these ERPs and will invest additional billions before the remaining 9 ERPs are fully implemented. Our prior reviews of several ERPs have found that the department has not effectively employed acquisition management controls or delivered the promised capabilities on time and within budget.[Footnote 6] This report provides information to support your continuing oversight of DOD's progress in modernizing its business systems to address long- standing weaknesses and ultimately to transform its business operations. As agreed with your office, our objectives were to (1) provide the status as of December 31, 2009 of the nine ERPs DOD identified as essential to transforming its business operations, (2) assess the scheduling and cost estimating practices of selected ERPs to determine the extent to which the program management offices (PMO) were applying best practices, and (3) ascertain whether DOD and the military departments have defined the performance measures to determine whether the systems will meet their intended business capabilities. To address the first objective, we reviewed status information obtained from each PMO, such as the reported amount of funds expended on the implementation of the nine ERPs, the estimated number of legacy systems to be replaced by each ERP, and the reported annual cost of maintaining these legacy systems. We also reviewed past GAO reports [Footnote 7] that were specific to the department's efforts to implement the nine ERPs to identify prior recommendations and assess DOD's progress in addressing the 19 recommendations discussed in these reports. For the purposes of this report, we did not include information on the Defense Logistics Agency (DLA) Business System Modernization (BSM)/ Enterprise Business System (EBS). According to DLA, the BSM effort was fully implemented in July 2007, and transformed how the agency conducts its operations in five core business processes: order fulfillment, demand and supply planning, procurement, technical/quality assurance, and financial management. Subsequently, in September 2007, the name of the program was changed to the EBS, which is a continuation of the ERP's capabilities to support internal agency operations. To address the second objective, we assessed the scheduling and cost estimating practices for four of the nine ERPs[Footnote 8] to determine the extent to which the PMOs were applying best practices for scheduling and cost estimating. For the four ERPs, we obtained and analyzed the most current schedule and cost estimate for each program and compared them against the criteria set forth in GAO's cost guide. [Footnote 9] In using the guide, we determined the extent to which the schedule was prepared in accordance with the best practices[Footnote 10] that are fundamental to having a reliable schedule. In assessing each program's cost estimates, we used the GAO cost guide to evaluate the PMOs' estimating methodologies, assumptions, and results to determine whether the cost estimates were comprehensive, accurate, well-documented, and credible. We did not conduct detailed schedule and cost assessments for the remaining five programs because (1) the implementation strategy has not been fully defined for two of the ERPs, (2) one of the ERPs is near full deployment, and (3) we have previously reported[Footnote 11] on two ERPs' schedule and cost estimating practices. To address the third objective, we reviewed the extent to which DOD and the military departments included performance measures in their congressional reports on business transformation. In addition, we met with the military departments' deputy chief management officers (DCMO) to obtain an understanding of how they define success in terms of deploying their respective ERPs. We also met with the DOD DCMO and the Director of the Business Transformation Agency (BTA) to obtain an understanding of their respective roles and responsibilities in the oversight of DOD's ERP implementation efforts. Additional details on our scope and methodology are presented in appendix I. We conducted this performance audit from June 2009 through October 2010 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. We requested comments on a draft of this report from the Secretary of Defense or his designee. We received written comments from the Deputy Chief Management Officer, which are reprinted in appendix II. Background: DOD is one of the largest and most complex organizations in the world. In fiscal year 2009, DOD reported that its operations consisted of $1.8 trillion in assets, $2.2 trillion in liabilities, approximately 3.2 million military and civilian personnel--including active and reserve components--and disbursements of over $947 billion.[Footnote 12] Execution of these operations spans a wide range of defense organizations, including the military departments and their respective major commands and functional activities, large defense agencies and field activities, and various combatant and joint operational commands that are responsible for military operations for specific geographic regions or theaters of operation. To execute military operations, the department performs interrelated and interdependent business functions, including financial management, logistics management, health care management, and procurement. To support its business functions, DOD has reported that it relies on about 2,080 business systems,[Footnote 13] including accounting, acquisition, logistics, and personnel systems. Funding of DOD's Business Systems: To fund its existing business systems environment, DOD requested for fiscal year 2011 nearly $17.4 billion to operate, maintain, and modernize its reported 2,080 business systems (see fig. 1). Of this amount, about $12.2 billion is for operations and maintenance and the remaining $5.2 billion is for planned or ongoing DOD business systems development modernization efforts. Figure 1: DOD's Fiscal Year 2011 Business Systems Budget Request by DOD Components (Dollars in Thousands): [Refer to PDF for image: illustrated table] Component: Army; Current services: $3,031,957; Development/modernization: $1,768,150; Total: $4,800,107; Percent: 27.7%. Component: Air Force; Current services: $2,323,175; Development/modernization: $1,666,103; Total: $3,989,278; Percent: 23.0%. Component: Navy; Current services: $2,310,296; Development/modernization: $536,492; Total: $2,846,788; Percent: 16.4%. Component: TRICARE Management Activity; Current services: $1,403,434; Development/modernization: $403,857; Total: $1,807,291; Percent: 10.4%. Component: Defense Logistics Agency; Current services: $763,438; Development/modernization: $160,478; Total: $923,916; Percent: 5.3%. Component: Defense Information Systems Agency; Current services: $689,584; Development/modernization: $25,190; Total: $714,774; Percent: 4.1%. Component: Defense Finance and Accounting Service; Current services: $397,239; Development/modernization: $29,812; Total: $427,051; Percent: 2.5%. Component: Defense Human Resources Activity; Current services: $253,215; Development/modernization: $67,950; Total: $321,165; Percent: 1.9%. Component: Transportation Command; Current services: $186,370; Development/modernization: $101,973; Total: $288,343; Percent: 1.7%. Component: Business Transformation Agency; Current services: $53,332; Development/modernization: $145,190; Total: $198,522; Percent: 1.2%. Component: Washington Headquarters Service Current services: $153,579; Development/modernization: $27,119; Total: $180,698; Percent: 1.0%. Component: Missile Defense Agency; Current services: $0; Development/modernization: $152,208; Total: $152,208; Percent: 0.8%. Component: Defense Commissary Agency; Current services: $117,861; Development/modernization: $3,616; Total: $121,477; Percent: 0.7%. Component: Defense Contract Management Agency; Current services: $103,391; Development/modernization: $13,933; Total: $117,324; Percent: 0.7%. Component: Department of Defense Dependents Education; Current services: $94,590; Development/modernization: $0; Total: $94,590; Percent: 0.6%. Component: Office of the Secretary of Defense; Current services: $29,755; Development/modernization: $49,098; Total: $78,853; Percent: 0.5%. Component: Joint Chiefs of Staff; Current services: $60,151; Development/modernization: $10,963; Total: $71,114; Percent: 0.4%. Component: Other DOD components; Current services: $180,479; Development/modernization: $23,104; Total: $203,58; Percent: 31.2%. Component: Total; Current services: $12,151,846; Development/modernization: $5,185,236; Total: $17,337,082; Percent: 100%. Source: GAO based upon fiscal year 2011 budget request data provided by DOD. This data has not been validated. [End of figure] The Office of Management and Budget (OMB) requires that funds requested for information technology (IT) projects be classified as either "steady state" (or "current services" in DOD) or as "development/modernization." Current services represents funds for operating and maintaining systems at current levels (i.e., without major enhancements). The development modernization budget category represents funds for developing new IT systems or making major enhancements to existing systems. Some systems have both current services and development modernization funding. While current services are to be used for operating the system at various locations, development modernization funds are to be used for activities such as developing and expanding system functionality at existing locations and deploying the system to new locations. Generally, current services are financed through Operation and Maintenance appropriations, whereas development modernization funding can come from several or a combination of several appropriations, such as Research, Development, Test, and Evaluation; Procurement; or the Defense Working Capital Fund. DOD's Acquisition System Framework: ERPs are developed within the defense acquisition system framework, which is intended to translate mission needs and requirements into stable, affordable, and well-managed acquisition programs.[Footnote 14] The defense acquisition system framework was updated in December 2008 and consists of five program life-cycle phases and three related milestone decision points which are described below. Materiel solution analysis (previously concept refinement). The purpose of this phase is to refine the initial system solution (concept) and create a strategy for acquiring the solution. A decision is made at the end of this phase (Milestone A) regarding whether to move to the next phase. - Milestone A authorizes acquisition of the program and permission to begin planning and development of the system technology. Technology development. The purpose of this phase is to determine the appropriate set of technologies to be integrated into the investment solution by iteratively assessing the viability of the various technologies while simultaneously refining user requirements. Once the technology has been demonstrated, a decision is made (Milestone B) whether to move to the next phase. - Milestone B authorizes product development of the program based on well-defined technology and a reasonable system design plan. * Engineering and manufacturing development (previously system development and demonstration). The purpose of this phase is to develop a system and demonstrate through developer testing that the system can function in its target environment. A decision is made at the end of this phase (Milestone C) whether to move to the next phase. - Milestone C authorizes entry of the system into the production and deployment phase or into limited deployment in support of operational testing. * Production and deployment. The purpose of this phase is to achieve an operational capability that satisfies the mission needs, as verified through independent operational test and evaluation, and to implement the system at all applicable locations. * Operations and support. The purpose of this phase is to operationally sustain the system in the most cost-effective manner over its life cycle. Overview of DOD Business Systems Investment Review Process: In 2005, DOD adopted a "tiered accountability" approach to improve control and accountability over the billions of dollars it invests annually in DOD business systems. Under this approach, executive leadership for the direction, oversight, and execution of DOD investments is the responsibility of several entities within DOD and its components. As indicated below, the investment control process begins at the component level and works its way up through a hierarchy of review and approval authorities, depending on the size and significance of the investment.[Footnote 15] * Defense Business Systems Management Committee (DBSMC) serves as the highest-ranking governance body for business systems modernization activities and approves funding request for investments costing more than $1 million within the department. * Investment review boards (IRB)[Footnote 16] are responsible for the review, approval, and oversight of the planning, design, acquisition, deployment, operation, maintenance, and modernization of defense business systems. The IRBs are also responsible for recommending business systems to the DBSMC for certification,[Footnote 17] which equates to recommending funding, for all business system investments costing more than $1 million. * The Milestone Decision Authority (MDA) is the senior DOD official who has overall authority to approve entry of an acquisition program into the next phase of the acquisition process and is accountable for cost, schedule, and performance reporting, including congressional reporting. * DOD Component Acquisition Executive is responsible for providing a written memorandum to the MDA through the cognizant IRB that (1) states that the program complies with applicable DOD statutory and regulatory requirements, (2) describes any conditions or issues applicable to the requested acquisition decision, and (3) recommends approval of the acquisition decision request. * A DOD component pre-certification authority (PCA) acts as the component's principle point of contact with the IRBs. The PCA is responsible for identifying the component's systems that require IRB certifications and prepares, reviews, approves, and validates investment documentation as required. The PCA also submits to the appropriate IRB the component's precertification memorandum that asserts the status and validity of the business system's investment information during the certification and annual review processes. The MDA, IRBs, the DBSMC or a combination of these can place conditions or issues needing resolution upon the individual programs during the defense business system's funding certification and acquisition decision review processes. These conditions are generally noted in a memorandum. Further, DOD's business investment management system includes two types of reviews for business systems: certification and annual reviews. Certification reviews apply to modernization projects with total costs over $1 million. These reviews focus on program alignment with the business enterprise architecture and must be completed before components obligate funds for programs. As noted above, the IRBs recommend certification to the DBSMC, which approves the expenditure of funds. The annual reviews apply to all business programs and are undertaken to determine whether the system development effort is meeting its milestones and addressing its certification conditions. Additionally, the Duncan Hunter National Defense Authorization Act for Fiscal Year 2009 directs that the executive-level oversight of DOD- wide business systems modernization and overall business transformation--including defining and measuring success in enterprise resource planning--is the responsibility of a military department- level chief management officer and the DCMO.[Footnote 18] DOD's ERP Efforts: The department stated that the following nine ERPs are critical to transforming the department's business operations and addressing some of its long-standing weaknesses. A brief description of each ERP is presented below. * The General Fund Enterprise Business System (GFEBS) is intended to support the Army's standardized financial management and accounting practices for the Army's general fund,[Footnote 19] with the exception of that related to the Army Corps of Engineers, which will continue to use its existing financial system, the Corps of Engineers Financial Management System.[Footnote 20] GFEBS will allow the Army to share financial, asset and accounting data across the active Army, the Army National Guard, and the Army Reserve. The Army estimates that when fully implemented, GFEBS will be used to control and account for about $140 billion in spending. * The Global Combat Support System-Army (GCSS-Army) is expected to integrate multiple logistics functions by replacing numerous legacy systems and interfaces. The system will provide tactical units with a common authoritative source for financial and related non-financial data, such as information related to maintenance and transportation of equipment. The system is also intended to provide asset visibility for accountable items. GCSS-Army will manage over $49 billion in annual spending by the active Army, National Guard, and the Army Reserve. * The Logistics Modernization Program (LMP) is intended to provide order fulfillment, demand and supply planning, procurement, asset management, material maintenance, and financial management capabilities for the Army's working capital fund. The Army has estimated that LMP will be populated with 6 million Army-managed inventory items valued at about $40 billion when it is fully implemented. * The Navy Enterprise Resource Planning System (Navy ERP) is intended to standardize the acquisition, financial, program management, maintenance, plant and wholesale supply, and workforce management capabilities at six Navy commands.[Footnote 21] Once it is fully deployed, the Navy estimates that the system will control and account for approximately $71 billion, or 50 percent, of the Navy's estimated appropriated funds--after excluding the appropriated funds for the Marine Corps and military personnel and pay. * The Global Combat Support System-Marine Corps (GCSS-MC) is intended to provide the deployed warfighter enhanced capabilities in the areas of warehousing, distribution, logistical planning, depot maintenance, and improved asset visibility. According to the PMO, once the system is fully implemented, it will control and account for approximately $1.2 billion of inventory. * The Defense Enterprise Accounting and Management System (DEAMS) is intended to provide the Air Force the entire spectrum of financial management capabilities, including collections, commitments and obligations, cost accounting, general ledger, funds control, receipts and acceptance, accounts payable and disbursement, billing, and financial reporting for the general fund. According to Air Force officials, when DEAMS is fully operational, it is expected to maintain control and accountability for about $160 billion. * The Expeditionary Combat Support System (ECSS) is intended to provide the Air Force a single, integrated logistics system--including transportation, supply, maintenance and repair, engineering and acquisition--for both the Air Force's general and working capital funds. Additionally, ECSS is intended to provide the financial management and accounting functions for the Air Force's working capital fund operations. When fully implemented, ECSS is expected to control and account for about $36 billion of inventory. * The Service Specific Integrated Personnel and Pay Systems are intended to provide the military departments an integrated personnel and pay system.[Footnote 22] * Defense Agencies Initiative (DAI) is intended to modernize the defense agencies' financial management processes by streamlining financial management capabilities and transforming the budget, finance, and accounting operations. When DAI is fully implemented, it is expected to have the capability to control and account for all appropriated, working capital and revolving funds at the defense agencies implementing the system. Status of DOD's ERP Implementation Efforts: Based upon the information provided by the PMOs, six of the ERPs have experienced schedule slippages (see table 1) based on comparing the estimated date that each program was originally scheduled to achieve full deployment[Footnote 23] to the full deployment date as of December 2009. For the remaining three ERPs, the full deployment date has either remained unchanged or has not been established. The GFEBS PMO noted that the acquisition program baseline approved in November 2008, established a full deployment date in fiscal year 2011 and that date remains unchanged. Additionally, according to the GCSS-Army PMO a full deployment date has not been established for this effort. The PMO noted that a full deployment date will not be established for the program until a full deployment decision has been approved by the department. A specific timeframe has not been established for when the decision will be made. Further, in the case of DAI, the original full deployment date was scheduled for fiscal year 2012, but the PMO is in the process of reevaluating the date and a new date has not yet been established. Table 1: Reported Full Deployment Schedule Slippage for Each ERP as of December 31, 2009: Army: Component/system name: GFEBS; Originally scheduled fiscal year for full deployment: 2011; Actual or latest estimated fiscal year for full deployment: 2011; Schedule slippage: None. Component/system name: GCSS-Army; Originally scheduled fiscal year for full deployment: [A]; Actual or latest estimated fiscal year for full deployment: [A]; Schedule slippage: Not applicable. Component/system name: LMP; Originally scheduled fiscal year for full deployment: 2005; Actual or latest estimated fiscal year for full deployment: 2011; Schedule slippage: 6 years. Navy: Component/system name: Navy ERP; Originally scheduled fiscal year for full deployment: 2011; Actual or latest estimated fiscal year for full deployment: 2013; Schedule slippage: 2 years. Component/system name: GCSS-MC; Originally scheduled fiscal year for full deployment: 2010; Actual or latest estimated fiscal year for full deployment: 2013; Schedule slippage: 3 years[B]. Air Force: Component/system name: DEAMS; Originally scheduled fiscal year for full deployment: 2014; Actual or latest estimated fiscal year for full deployment: 2017; Schedule slippage: 3 years. Component/system name: ECSS; Originally scheduled fiscal year for full deployment: 2012; Actual or latest estimated fiscal year for full deployment: 2016; Schedule slippage: 4 years. DOD: Component/system name: Service Specific Integrated Personnel and Pay Systems; Originally scheduled fiscal year for full deployment: 2006; Actual or latest estimated fiscal year for full deployment: Army--2014; Navy--2017; Air Force--2018; Schedule slippage: 12 years[C]. Component/system name: DAI; Originally scheduled fiscal year for full deployment: 2012; Actual or latest estimated fiscal year for full deployment: [D]; Schedule slippage: Not applicable. Source: DOD program management offices. [A] The PMO has not yet determined the full deployment date. [B] The PMO stated that the estimated full deployment date is only for phase 1. The full deployment date for the entire program has not yet been determined. [C] Originally, this ERP was referred to as the Defense Integrated Military Human Resources System (DIMHRS) and was intended to provide a joint, integrated, standardized personnel/pay system for all military personnel departmentwide. The original full deployment date represents the estimated date for DIMHRS. Each military service is now responsible for developing its own integrated personnel and pay system. [D] As of December 2009, the DAI PMO had not determined the revised full deployment date. [End of table] Besides schedule slippages, five of the ERP efforts have reported a cost increase and one program--GFEBS--reported a cost decrease of $17 million (see table 2). The reported life-cycle[Footnote 24] cost estimate for GCSS-MC only represents the estimated cost for phase[Footnote 25] 1 of the program. The cost of the remaining phases has not yet been determined and therefore, a total life-cycle cost estimate for the entire program has not been determined. Additionally, a current life-cycle cost estimate has not been determined for the Service Specific Integrated Personnel and Pay Systems and DAI. Table 2: Reported Original and Current Life-Cycle Cost Estimate for Each ERP as of December 31, 2009: Dollars in millions. Army: Component/system name: GFEBS; Original life-cycle cost estimate: $1,354; Current life-cycle cost estimate: $1,337; Reported cost increase: $(17). Component/system name: GCSS-Army; Original life-cycle cost estimate: $3,900; Current life-cycle cost estimate: $3,900; Reported cost increase: 0. Component/system name: LMP; Original life-cycle cost estimate: $2,630; Current life-cycle cost estimate: $2,630[A]; Reported cost increase: 0. Navy: Component/system name: Navy ERP; Original life-cycle cost estimate: $1,870; Current life-cycle cost estimate: $2,400; Reported cost increase: $530. Component/system name: GCSS-MC; Original life-cycle cost estimate: $126; Current life-cycle cost estimate: $934; Reported cost increase: $808[B]. Air Force: Component/system name: DEAMS; Original life-cycle cost estimate: $1,100; Current life-cycle cost estimate: $2,048; Reported cost increase: $948. Component/system name: ECSS; Original life-cycle cost estimate: $3,000; Current life-cycle cost estimate: $5,200; Reported cost increase: $2,200[C]. DOD: Component/system name: Service Specific Integrated Personnel and Pay Systems; Original life-cycle cost estimate: $577[D]; Current life-cycle cost estimate: Army[D]; Navy-$1,300; Air Force- $1,700; Reported cost increase: At least $2,423. Component/system name: DAI; Original life-cycle cost estimate: $209; Current life-cycle cost estimate: [E]; Reported cost increase: Not applicable. Source: DOD Program Management Offices. [A] At the time LMP was designated as a major automated information system (MAIS) program in December 2007, it was required to comply with the DOD guidance for MAIS programs. This guidance requires, among other things, that a MAIS program have a completed and approved acquisition program baseline--the baseline description of the program, including the life-cycle cost estimate--prior to Milestone B approval. The $2.6 billion is the only life-cycle cost estimate that has been developed for the program. [B] The current life-cycle cost estimate for GCSS-MC is for phase one. The remaining two phases will have separate baselines. [C] Originally, ECSS was to be implemented in three phases, but now, it will be implemented in four phases. [D] The original life-cycle cost estimate represents the estimate for DIMHRS. While the Navy and Air Force have estimated their respective life-cycle cost estimate, the Army is in the process of completing its life-cycle cost estimate. [E] As of December 2009, the life-cycle cost estimate for DAI had not been finalized. According to the PMO, the life-cycle cost estimate is expected to be approved at Milestone B in fiscal year 2011. [End of table] According to the PMOs, while there have been schedule slippages and cost increases, for several of the nine ERP efforts, the functionality that was envisioned and planned when each program was initiated remains the same today. While the original intent of each program remains the same, the anticipated savings that were to accrue to the department may not be fully realized. Delays in implementing the ERPs result in DOD having to fund the operation and maintenance of the legacy systems longer than anticipated, thereby reducing funds that could be used for other DOD priorities. Furthermore, we have previously reported on the department's effort in implementing some of the ERPs and made 19 recommendations to improve DOD's management and oversight of these efforts. As of October 2010, the department has taken sufficient action to implement 5 of the recommendations. Appendix III provides details on the specific recommendations and the department's efforts to address them. The following information describes in more detail the status of each ERP. General Fund Enterprise Business System: Figure: DOD Program Data for GFEBS, as of December 31, 2009: [Refer to PDF for image: text box] Date of initiation: October 2004. Program owner: Assistant Secretary of the Army for Financial Management and Comptroller. Reported life-cycle cost estimate: $1.336.7 billion: * Development and Modernization: $642.4 million; * Operations and Maintenance: $694.3 million. Reported amount expended: $416.8 million. Reported legacy systems to be replaced: 87. Reported annual cost of maintaining legacy systems: $57.8 million. Number of system interfaces: 56. Date of last certification of funding: September 2, 2009, by the DBSMC. Number of system users: 79,000. Number of locations: 200. Source: DOD’s GFEBS Program Management Office. These data have not been validated. [End of figure] Program Status: According to the GFEBS PMO, the system will be implemented in four phases. Phases 1 and 2 were completed in October 2008 and provided full functionality to 250 users at the Management Command, Fort Jackson, South Carolina. The implementation of phase 2 set the stage for GFEBS to be deployed to the rest of the Army. The PMO currently estimates that phases 3 and 4 will be deployed Army-wide with full functionality by December 2011. PMO officials told us that the establishment of the December 2011 milestone resulted from conditions placed on the GFEBS program at Milestone B, directing the Army to develop an integrated strategy for the implementation of GFEBS and GCSS-Army--meaning that both systems were to be implemented using a standard configuration and set of common master data.[Footnote 26] The PMO also stated that the original life-cycle cost estimate of approximately $1.3 billion covering fiscal years 2005 through 2022 remained unchanged as of December 31, 2009. On May 30, 2009, GFEBS was authorized by the MDA to proceed with a limited deployment to initial operational test and evaluation (IOT&E) [Footnote 27] sites. In January 2010 and again in March 2010, the GFEBS program was authorized to continue its deployment to a limited number of sites. According to the MDA, this limited deployment process allows the program to gain additional operational experience with the GFEBS application and conduct additional user testing. MDA approval is required for deployment to additional sites and full deployment. Before GFEBS will be granted approval for full deployment of phase 4, the PMO must address several conditions[Footnote 28] that were placed on the program by the MDA. According to the PMO, all of the conditions were addressed in December 2009 and presented to the IRB for approval. However, the decision on the deployment of the system to additional locations is pending and scheduled to occur during fiscal year 2010. In December 2009, the U.S. Army Test and Evaluation Command (ATEC) reported on concerns with GFEBS's data accuracy, reliability, and timeliness.[Footnote 29] More specifically, the report noted that Army "installations certifying year-end data with caveats and notes related to inaccurate, incomplete, and missing data." Furthermore, the report noted that "because of incomplete or not implemented business processes, users at times, executed their mission using the "workarounds" of the legacy systems that the GFEBS is intended to replace or subsume." The report recommended that the deployment of GFEBS be limited until the problems are resolved and the corrective actions have been validated by ATEC. According to the PMO, in conjunction with ATEC, a plan of action and milestones has been developed to address the issues. The PMO noted that GFEBS is undergoing an additional operational test and evaluation limited user test; and at the conclusion of the testing, a determination will be made whether the ATEC issues have been addressed. Global Combat Support System-Army: Figure: DOD Program Data for GCSS-Army, as of December 31, 2009: [Refer to PDF for image: text box] Date of initiation: December 2003[A]. Program owner: Army Deputy Chief of Staff for Logistics. Reported life-cycle cost estimate: $3.9 billion: * Development and Modernization: $1.8 billion; * Operations and Maintenance: $2.1 billion. Reported amount expended: $581 million. Reported legacy systems to be replaced: 7. Reported annual cost of maintaining legacy systems: $63 million. Number of system interfaces: 106. Date of last certification of funding: September 2, 2009, by the DBSMC. Number of system users: 169,880. Number of locations: 379. Source: DOD's GCSS-Army Program Management Office. These data have not been validated. [A] Prior to the initiation of the current ERP effort, the Army had been developing custom software since May 1997. [End of figure] Program Status: GCSS-Army is being implemented in three phases with phases 1 and 2 being proof-of-concept demonstrations that have been ongoing since December 2007 at the National Training Center (NTC) in Fort Irwin, California and testing and evaluation are scheduled to be completed in January 2012. The GCSS-Army team is conducting critical activities, such as data cleansing and training users at the NTC site. Phase 3 is intended to provide full functionality and is scheduled to begin implementation in October 2013, but a full deployment date has not yet been determined. According to the PMO, the exact locations for the implementation of phase 3 have not been determined because the deployment schedule by specific location has not yet been finalized. In July 2008, the MDA, in approving Milestone B, directed GCSS-Army to develop and implement a strategy to better facilitate interactions with GFEBS and LMP. Under the federated strategy, GCSS-Army will use GFEBS' financial template to allow the Army to integrate data on logistics, financial, maintenance, property, and accountability of assets. This strategy is intended to standardize transactional input and business processes across the Army ERPs to enable common cost management activities; provide accurate, reliable, and real-time data; and tie budgets to execution. According to the PMO, this change in implementation strategy resulted in: * the Cost Analysis Improvement Group's direction that an additional year of support be added to the cost estimate because of the additional time needed to deploy the system and: * a revised strategy that resulted in an increase in the number of required reports, interfaces, conversions, and extensions that need to be developed or tested for GCSS-Army's integration with GFEBS. Logistics Modernization Program: Figure: DOD Program Data for LMP, as of December 31, 2009] [Refer to PDF for image: text box] Date of initiation: December 1999. Program owner: Army Materiel Command. Reported life-cycle cost estimate: $2.630 billion: * Development and Modernization: $637 million; * Operations and Maintenance: $1.993 billion. Reported amount expended: $1.1 billion. Reported legacy systems to be replaced: 2. Reported annual cost of maintaining legacy systems: $25 million. Number of system interfaces: 27. Date of last certification of funding: September 2, 2009, by the DBSMC. Number of system users: 21,000. Number of locations: 104. Source: DOD's LMP Program Management Office. These data have not been validated. [End of figure] Program Status: LMP was deployed at the Army Communications-Electronics Command and Tobyhanna Army Depot in July 2003. In May 2009, the second deployment of LMP became operational at the Army Aviation and Missile Command and Corpus Christi and Letterkenny Army Depots. The final deployment of LMP is scheduled to occur in October 2010 at the Army Sustainment Command, the Joint Munitions and Lethality Command, the Tank- automotive and Armaments Command, and the Anniston and Red River Army Depots. LMP has experienced schedule slippages primarily because requirements management[Footnote 30] and system testing were ineffective which we reported on in May 2004[Footnote 31] and June 2005.[Footnote 32] For example, at the Tobyhanna Army Depot deployment in fiscal year 2003, customers were not being properly billed for work performed which affected the accurate recording of revenue, and account balances could not be reconciled when transferred from the legacy systems to LMP. As a result, the full deployment date of the system has slipped by 6 years. Furthermore, in April 2010,[Footnote 33] we reported that the Army's management processes for ensuring data reliability that were established prior to the second deployment of LMP were not effective. Specifically, the Army was unable to ensure that the data used by LMP were of sufficient quality to enable the depots to perform their day- to-day missions after LMP became operational. As a result of these data quality issues, depot personnel had to develop and use manual work-around processes until they could correct the data in LMP, which prevented the Army from achieving the expected benefits from LMP. Data quality issues occurred despite improvements made by the Army to address similar issues experienced during the first deployment of LMP because the Army's testing strategy did not provide reasonable assurance that the data being used by LMP were accurate and reliable. We made recommendations to help improve the third deployment of LMP. We are following up on the Army's efforts to implement our recommendations and will report on those actions separately. The PMO further noted that the original life-cycle cost estimate of approximately $2.6 billion[Footnote 34] covering fiscal years 2000 through 2021 remained unchanged as of December 2009. PMO officials told us that there were no issues or conditions that had been placed upon LMP by the MDA, IRBs or the DBSMC that needed to be resolved as of December 2009. Navy Enterprise Resource Planning System: Figure: DOD Program Data Provided for Navy ERP, as of December 31, 2009: [Refer to PDF for image: text box] Date of Initiation: July 2003. Program owner: Assistant Secretary of the Navy, Research, Development, and Acquisition. Reported life-cycle cost estimate: $2.4 billion: * Development and Modernization: $1.0 billion; * Operations and Maintenance: $1.4 billion. Reported amount expended: $691.3 million. Reported legacy systems to be replaced: 98. Reported annual cost of maintaining legacy systems: $102 million. Number of system interfaces: 51. Date of last certification of funding: September 2, 2009, by the DBSMC. Number of system users: 66,000. Number of locations: 53. Source: DOD's Navy ERP Program Management Office. These data have not been validated. [End of figure] Program Status: Navy ERP is to be implemented in two phases. As part of phase 1, the financial and acquisition functionalities of Navy ERP were deployed to the Naval Air Systems Command, Naval Supply Systems Command, and the Space and Naval Warfare Systems Command. Those functionalities are scheduled for deployment for the general fund at the Naval Sea Systems Command in October 2010 and the Navy Working Capital Fund in October 2011 and the Office of Naval Research and Strategic Systems Planning in October 2012. Phase 2 is currently in progress with the deployment of the wholesale and retail supply functionalities to the Navy. According to the PMO, Navy ERP is currently being used by 38,000 users and is executing approximately $37 billion of the Navy's total obligational authority. Further, the PMO noted that in fiscal year 2010, 19 legacy systems have already been retired. The Navy ERP implementation has experienced slippages of 2 years. Originally, the Navy ERP was to achieve full deployment in fiscal year 2011, but now full deployment is planned for fiscal year 2013. According to program documentation, these slippages occurred, in part, because of problems experienced in data conversion and adopting new business procedures associated with implementing the ERP. The delay occurred at the Naval Air Systems Command and affected the deployment schedule for the other locations. In addition to slippages in schedule, there have also been increases in the life-cycle cost estimate. The 2003 original life-cycle cost estimate for the Navy ERP was about $1.87 billion. This estimate was later revised in August 2004, December 2006, and again in September 2007 to $2.4 billion. According to the September 2007 acquisition program baseline, the estimated $2.4 billion is for acquisition, operations, and support for fiscal years 2004 through 2023. Moreover, in September 2008,[Footnote 35] we reported that not effectively implementing key IT management controls, such as earned value management, has contributed to the more than 2-year schedule delay and almost $600 million cost overrun on the program since it began, and will likely contribute to future delays and overruns if not corrected. The IRB has identified two issues or conditions that the Navy ERP PMO has to address: (1) provide a description of how the Navy plans to use the item-unique identification (IUID)[Footnote 36] and (2) provide an updated checklist to BTA showing compliance with the Standard Financial Information Structure (SFIS).[Footnote 37] The PMO stated that it presented a plan to the Navy Comptroller in February 2010 describing how it will use IUID and provided BTA the SFIS checklist in April 2010. Global Combat Support System-Marine Corps: Figure: DOD Program Data for GCSS-MC, as of December 31, 2009: [Refer to PDF for image: text box] Date of Initiation: September 2003. Program owner: Assistant Secretary of the Navy, Research, Development, and Acquisition. Reported life-cycle cost estimate: $934 million: * Development and Modernization: $489 million; * Operations and Maintenance: $445 million. Reported amount expended: $245 million. Reported legacy systems to be replaced: 4. Reported annual cost of maintaining legacy systems: $4.5 million. Number of system interfaces: 42. Date of last certification of funding: June 1, 2010 by the DBSMC. Number of system users: 33,000. Number of locations: 6. Source: DOD's GCSS-MC Program Management Office. These data have not been validated. [End of figure] Program Status: GCSS-MC was authorized to "Go Live" for field user evaluation in March 2010 and Milestone C was granted in May 2010. GCSS-MC is to be implemented in three phases. Phase 1 is intended to provide a wide range of asset management capabilities such as planning inventory requirements to support current and future demands; requesting and tracking the status of products (e.g., supplies and personnel) and services (e.g., maintenance and engineering); allocating resources (e.g., inventory, warehouse capacity, and personnel) to support unit demands for specific products; and scheduling maintenance resources (e.g., manpower, equipment, and supplies) for specific assets, such as vehicles. Phases 2 and 3 are intended to provide additional functionally such as transportation and wholesale inventory management. To date, there have been program slippages and cost increases. The PMO told us that full deployment for phase 1 was originally scheduled to be achieved in November 2009. However, the current estimated full deployment date for phase 1 is January 2013.[Footnote 38] GCSS-MC program officials informed us that the schedule slippage for phase 1 occurred incrementally over time during the design, build, and test phases of the program. The slippages occurred because of issues associated with system interfaces and the conversion of data from the legacy systems. Moreover, in July 2008,[Footnote 39] we reported that not effectively implementing key IT management controls, such as economically justifying investment in the system, has in part contributed to a 3-year schedule slippage and about $193 million cost overrun on the first phase of the program and will likely contribute to future delays and overruns if not corrected. These schedule slippages caused the program to exceed the MAIS critical-breach criteria for time-certain development, which is the failure to achieve initial operating capability within 5 years of Milestone A approval. PMO officials also told us that initially, GCSS- MC had an estimated cost of approximately $126 million over a 7-year life cycle.[Footnote 40] This cost estimate was later revised in 2005 to approximately $249 million over a 13-year life cycle.[Footnote 41] Currently, the PMO estimates the total life-cycle cost estimate for phase 1 to be approximately $934 million. The total life-cycle cost estimate for the additional phases has not been determined. According to the PMO, phase 2 is in the preliminary planning stage and all additional phases will have separate acquisition program baselines. [Footnote 42] As a result, a total life-cycle cost estimate for the entire system may not be available for several years. The IRB directed that the GCSS-MC PMO (1) provide a component-wide plan that addresses how GCSS-MC will include the capability to use IUID and (2) brief the Navy DCMO on the extent to which business process reengineering (BPR) has been performed to address the statutory requirement regarding BPR in Section 1072 of the Fiscal Year 2010 National Defense Authorization Act. In this regard, the act directs the Chief Management Officer to determine whether or not appropriate business process re-engineering efforts have been undertaken to ensure that (1) the business process to be supported by the business system will be as streamlined and efficient as practicable and (2) the need to tailor the ERP to meet unique requirements or incorporate unique interfaces has been eliminated. The PMO stated that it provided the BPR information to the Navy DCMO and the DCMO indicated that the PMO had addressed the requirements contained in the act. Defense Enterprise Accounting and Management System: Figure: DOD Program Data for DEAMS, as of December 31, 2009: [Refer to PDF for image: text box] Date of initiation: August 2003. Program owner: Assistant Secretary of the Air Force for Financial Management and Comptroller. Reported life-cycle cost estimate: $2.048 billion: * Development and Modernization: $1.030 billion; * Operations and Maintenance: $1.018 billion. Reported amount expended: $139.1 million. Reported legacy systems to be replaced: 10. Reported annual cost of maintaining legacy systems: $55.9 million. Number of system interfaces: 100. Date of last certification of funding: December 14, 2009, by the DBSMC. Number of system users: 30,000. Number of locations: 179. Source: DOD's DEAMS Program Management Office. These data have not been validated. [End of figure] Program Status: DEAMS will be deployed in three phases. Phase 1 deployed limited functionality--recording commitments--to about 650 system users at Scott Air Force Base in July 2007. According to the PMO, as part of phase 1, additional functionality was deployed to an additional 870 users in May 2010. Further, the PMO noted that DEAMS is currently scheduled to achieve initial operating capability for phase 2 for the U.S. Transportation Command and most of the Air Force's major commands in fiscal year 2014. According to the PMO, the final phase of DEAMS will be deployed to the remaining Air Force's major commands by fiscal year 2017, thereby providing the entire spectrum of general fund capabilities to the entire Air Force. The Air Force expects DEAMS to reach full deployment in fiscal year 2017--which is a 3-year slippage from the full deployment date reported at program initiation. According to the PMO, DEAMS has experienced a 3-year schedule slippage because of problems caused by software code defects, integration test delays and to accommodate schedule risk. DEAMS program management officials acknowledged that the standardization of computer desktops across the Air Force contributed to schedule slippages. Our August 2008 report discussed this specific problem.[Footnote 43] In addition to schedule slippages, DEAMS also had an increase in its life-cycle cost estimate. In August 2008, we reported that the Air Force's life-cycle cost estimate for DEAMS was about $1.1 billion through fiscal year 2021.[Footnote 44] According to the PMO, as of December 2009, the life-cycle cost estimate for the DEAMS is approximately $2 billion through fiscal year 2027. The PMO stated that the increase in the life-cycle cost estimate can be attributed to changes in the program implementation strategy from two phases to three phases and program development and testing issues. The IRB directed the DEAMS's PMO to (1) create an IUID compliance plan indicating when the system will include the capability to use IUID, (2) identify the date that one of the legacy systems will be subsumed, (3) provide a plan on how DEAMS will meet Environmental Liabilities Recognition Valuation and Reporting requirements, and (4) comply with Section 1072 of the National Defense Authorization Act for Fiscal Year 2010 related to business process reengineering. According to the PMO, the Air Force has addressed these issues. Expeditionary Combat Support System: Figure: DOD Program Data for ECSS, as of December 31, 2009: [Refer to PDF for image: text box] Date of Initiation: January 2004. Program owner: Deputy Chief of Staff for Logistics, Installations, and Mission Support, Headquarters, U.S. Air Force. Reported life-cycle cost estimate: $5.2 billion: * Development and Modernization: $3.4 billion; * Operations and Maintenance: $1.8 billion. Reported amount expended: $518.9 million. Reported legacy systems to be replaced: 240. Reported annual cost of maintaining legacy systems: $325 million. Number of system interfaces: 157 (phase 1) and 673 (phases 2, 3, and 4). Date of last certification of funding: September 2, 2009, by the DBSMC. Number of system users: 250,000. Number of locations: 186. Source: DOD's ECSS Program Management Office. These data have not been validated. [End of figure] Program Status: ECSS will be deployed in four phases. The Air Force anticipates that phase 1 will begin deployment in June 2012, with phase 2 scheduled for deployment in April 2014, phase 3 in January 2015, and phase 4 in November 2015. According to the PMO, each phase will provide additional functionality to the system users. Phase 1 will focus on base materiel and equipment management, phase 2 will concentrate on global materiel and equipment management and enterprise planning, phase 3 will involve depot maintenance repair and overhaul, and phase 4 will involve flight line maintenance and ammunition management. The PMO estimated that full deployment will be achieved in July 2016--a slippage of at least 4 years. According to the PMO, the slippage can be attributed to (1) two contract award protests--both denied by GAO-- and (2) the change in the implementation strategy, which had originally called for the system to be implemented in three phases. Also, in our August 2008 report,[Footnote 45] we noted that the life- cycle cost estimate was approximately $3 billion for the entire ECSS program when it was scheduled for three phases. According to the ECSS PMO, the current life-cycle cost estimate is approximately $5.2 billion. Funding has not yet been approved for phases 2 through 4. The PMO noted that ECSS will seek approval at each phase's critical milestone in order to go forward to the next phase. The Air Force DCMO told us that Air Force leadership (including the Secretary of the Air Force, Air Force Chief of Staff, and Senior Acquisition Executive) reviewed the program to determine whether it should be restructured or canceled. The leadership was specifically concerned about the size, scope, and pace of the program. The program was restructured, and in June 2009, the decision was made to pursue only the revised phase 1 pending a demonstration of the program's ability to deliver to the revised schedule. The DCMO told us that the Air Force will make a decision on (1) whether to implement phase 1 and (2) whether to budget for the other phases in June 2010. According to the PMO, it anticipates the Air Force fully funding phase 1 and the long-lead requirements for phase 2 in the fiscal year 2012 program objective memorandum.[Footnote 46] Because of changes in the implementation strategy, in September 2009, the DOD MDA approved a revised Milestone A for ECSS. The revised milestone provides for additional funding, and it grants the Air Force authority to continue with ECSS technology development and prepare for Milestone B for phase 1. In preparing for Milestone B, the Air Force was directed to: * present quarterly reports regarding the progress of the program, including internal and external challenges and risks, to the IRBs for weapons system, material, service, and financial management; * complete an enterprise risk assessment methodology review of the program 120 days prior to Milestone B; and: * provide a cost analysis requirement document to the Air Force Analysis Agency to support the development of an independent cost estimate. According to the PMO, each of these actions was completed by May 20, 2010. Service Specific Integrated Personnel and Pay Systems[Footnote 47] Figure: DOD Program Data for Service Specific Integrated Personnel and Pay Systems, as of December 31, 2009: [Refer to PDF for image: text box] Date of Initiation: February 1998. Program owner: Army-—Army’s Program Executive Office, Enterprise Information Systems; Navy-—Chief of Naval Operations; Air Force-—Air Force Program Executive Office and Service Acquisition Executive. Reported life-cycle cost estimate: Army-—Has not yet been determined; Navy-—$1.3 billion; Air Force-—$1.7 billion. Reported amount expended: $841.1 million. Legacy systems to be replaced: Army-—65; Navy-—7; Air Force-—Has not yet been determined. Reported annual cost of maintaining legacy systems: Army—-$39 million; Navy-—$69 million; Air Force—-Has not yet been determined. Number of system interfaces: Has not yet been determined. Date of last certification of funding: Not applicable for the military services as of December 2009. Number of system users: Has not yet been determined. Number of locations: Has not yet been determined. Sources: The Army, Navy, and Air Force program management offices. These data have not been validated. [End of figure] Program Status: In a January 2009 memorandum, the Deputy Secretary of Defense changed the department's strategy for implementing an integrated personnel and pay system. The memorandum directed the BTA to develop the pay module and provide it to the military departments. Each military department would be responsible for implementing an integrated personnel and pay system for its respective service. In revising the department's strategy, a subsequent memorandum issued September 2009 by the Under Secretary of Defense (Acquisition, Technology and Logistics) noted that the capabilities needed by DOD to develop integrated personnel and pay systems are best met through the military departments because of several risks, including governance, technical complexities, and past failed attempts of developing DIMHRS as a one-fits-all solution. The memorandum further noted that military departments were to use, to the maximum extent practical, the DIMHRS requirements related to the pay module developed by BTA. Highlighted below is the status of each of the military department's efforts to implement an integrated personnel and pay system. Integrated Personnel and Pay System-Army (IPPS-A): Army PMO officials told us that in accordance with the September 2009 memorandum, the Army intends to use the BTA-developed pay module, develop the personnel module and implement an integrated system. Once IPPS-A is developed it will be implemented in several phases. The first deployment is planned for the Army National Guard, followed by the Army Reserves, and then the active Army. The PMO stated that the personnel and pay portion will be deployed to all Army components by August 2014. The Army anticipates that full deployment will occur late in fiscal year 2014. The PMO informed us that the Army is in the process of developing the life-cycle cost estimate. Navy Future Pay and Personnel Solution: According to PMO officials, the Navy is in the process of evaluating the extent to which the BTA-developed pay module can be used to meet its needs for an integrated system. Navy anticipates that this evaluation will be completed by the second quarter of fiscal year 2011. PMO officials told us that if the pay module can be used, the system will be implemented in two phases. Phase 1 will consolidate the existing legacy personnel systems and establish a single personnel record. Phase 2 will be the implementation of the pay module. Navy would begin deployment in fiscal year 2014 for phase 1 and fiscal year 2015 for phase 2, with full deployment being achieved in fiscal year 2017. PMO officials told us that the Navy estimates that the life- cycle cost estimate for its integrated personnel and pay system will be about $1.3 billion. The PMO further stated that if an alternative to using the BTA-developed pay module is selected, the implementation dates and estimated cost may change. In the September 2009 memorandum, it was noted that the Marine Corps will continue to use the Marine Corps Total Force System because it is already an integrated personnel and pay system. Air Force Integrated Personnel and Pay System: At the time of our review, Air Force was evaluating the BTA-developed pay module to assess whether it could be used. According to the PMO, the system will be implemented in three phases, provided the existing BTA-developed pay module can be used. Phase 1 will consist of transferring data from the legacy systems to the new integrated personnel and pay system and will include implementation of leave/ benefits for all. Phase 2 will provide an integrated personnel and pay solution for active Air Force officers, and phase 3 will deploy the system to the rest of the Air Force personnel including guard and reserve personnel. The quantity and content of the phases may change as the Air Force evolves the acquisition and deployment strategies. According to the PMO, it is anticipated that full deployment will be achieved in April 2018. The Air Force PMO currently estimates the life- cycle cost estimate to be about $1.7 billion covering fiscal year 2010 through fiscal year 2027. The PMO told us that as the Air Force better defines its implementation strategy, the implementation dates and life- cycle cost estimate could change. The PMO also said that the Air Force is in the process of ascertaining how many legacy systems can be eliminated through its implementation of an integrated personnel and pay system. Defense Agencies Initiative: Figure: DOD Program Data for DAI, as of December 31, 2009: [Refer to PDF for image: text box] Date of Initiation: January 2007. Program owner: The Business Transformation Agency was the first entity to implement DAI. Each defense agency will be responsible for the management and oversight of its respective implementation. Reported life-cycle cost estimate: Has not yet been determined; * Development and Modernization: Has not yet been determined; * Operations and Maintenance: Has not yet been determined. Reported amount expended: $40.2 million. Reported legacy systems to be replaced: 17. Reported annual cost of maintaining legacy systems: $35 million. Number of system interfaces: 24. Date of last certification of funding: September 30, 2009 by the DBSMC. Number of system users: 15,000 (estimated). Number of locations: 11 (estimated). Source: DOD's DAI Program Management Office. These data have not been validated. [End of figure] Program Status: DAI became operational at BTA in October 2008 and at the Defense Technical Information Center in October 2009. Table 3 lists the defense agencies that are scheduled to implement DAI in fiscal years 2011 through 2013. Table 3: Defense Agencies' Scheduled Implementation of DAI: Defense agency: Uniform Services University of the Health Services; Scheduled Implementation of DAI: Fiscal year 2011. Defense agency: Missile Defense Agency; Scheduled Implementation of DAI: Fiscal year 2011. Defense agency: Defense Threat Reduction Agency; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: Defense Information Systems Agency; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: Defense Technology Security Administration; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: Chemical Biological Defense Program; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: TRICARE Management Agency--Headquarters; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: Defense Media Agency; Scheduled Implementation of DAI: Fiscal year 2012. Defense agency: Defense Information System Agency--General Fund; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Defense Acquisition University; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Defense POW/Missing Personnel Office; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Defense Advanced Research Projects Agency; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Defense Security Service; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Office of Economic Adjustment; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: Center for Countermeasures; Scheduled Implementation of DAI: Fiscal year 2013. Defense agency: National Defense University; Scheduled Implementation of DAI: Fiscal year 2013. Source: Business Transformation Agency. [End of table] There has been some slippage in the implementation schedule. However, at the time of our review, a revised full deployment date for all of the agencies scheduled to use DAI had not been established. According to the department's fiscal year 2011 IT budget request, additional defense agencies have expressed an interest in using DAI. However, the Financial Management IRB and the DBSMC must grant approval to any entity that wants to use DAI. DOD's budget request notes that the total cost of the program is affected by the number of agencies participating. The budget request further notes that a more accurate implementation-plus-sustainment cost can be determined once all of the signed memorandums of intent from agencies wanting to use DAI have been received. DOD Did Not Follow Key Best Practices for Estimating ERP Schedules and Cost, Resulting in Unreliable Estimates: Our analysis of the schedules and cost estimates for four ERP programs--DEAMS, ECSS, GFEBS, and GCSS-Army--found that none of the programs are fully following best practices for developing reliable schedules and cost estimates. More specifically, none of the programs had developed a fully integrated master schedule (IMS) that reflects all activities, including both government and contractor activities. In addition, none of the programs established a valid critical path or conducted a schedule risk analysis.[Footnote 48] We have previously reported that the schedules for GCSS-MC and Navy ERP were developed using some of these best practices, but several key practices were not fully employed that are fundamental to having a schedule that provides a sufficiently reliable basis for estimating costs, measuring progress, and forecasting slippages.[Footnote 49] We recommended that each program follow best practices to update its respective schedule. DOD generally agreed with the recommendations. Additional details on the status of the recommendations are discussed in appendix III. The success of any program depends on having a reliable schedule of the program's work activities that will occur, how long they will take, and how the activities are related to one another. As such, the schedule not only provides a road map for systematic execution of a program, but also provides the means by which to gauge progress, identify and address potential problems, and promote accountability. Our analysis of the four programs' cost estimates found that ECSS, GFEBS, and GCSS-Army did not include a sensitivity analysis, while cost estimates for GFEBS did not include a risk and uncertainty analysis. GAO, OMB, and DOD guidance[Footnote 50] stipulate that risk and uncertainty analysis should be performed to determine the level of risk associated with the dollar estimate. Furthermore, a sensitivity analysis would assist decision makers in determining how changes to assumptions or key cost drivers (such as labor or equipment) could affect the cost estimate. We have previously reported that the cost estimates for Navy ERP and GCSS-MC are comprehensive and well- documented, but only partially accurate and credible.[Footnote 51] We recommended that each program update its respective cost estimate following best practices. The department generally agreed with the recommendations. Additional details on the status of the recommendations are discussed in appendix III. For DOD management to make good decisions, the program estimate must reflect the degree of uncertainty so that a level of confidence can be given about the estimate. A reliable cost estimate provides the basis for informed investment decision making, realistic budget formulation and program resourcing, meaningful progress measurement, proactive course correction, and accountability for results. Program Schedules Not Developed in Accordance with Key Scheduling Practices: Our cost guide best practices and related federal guidance call for a program schedule to be programwide, meaning that it should include an integrated breakdown of the work to be performed by both the government and its contractors over the expected life of the program. [Footnote 52] Our guidance identifies nine scheduling best practices that are integral to a reliable and effective master schedule: (1) capturing all activities, (2) sequencing all activities, (3) assigning resources to all activities, (4) establishing the duration of all activities, (5) integrating schedule activities horizontally and vertically, (6) establishing the critical path for all activities, (7) identifying float between activities, (8) conducting a schedule risk analysis, and (9) updating the schedule using logic and durations to determine the dates. The scheduling best practices are interrelated so that deficiencies in one best practice will cause deficiencies in other best practices. For example, if the schedule does not capture all activities, then there will be uncertainty about whether activities are sequenced in the correct order and whether the schedule properly reflects the resources needed to accomplish the work. The schedule should use logic and durations in order to reflect realistic start and completion dates for program activities. Maintaining the integrity of the schedule logic is not only necessary to reflect true status, but is also required before conducting follow-on schedule risk analyses. If the schedule is not properly updated, positive and negative float will not change properly. Positive float indicates the amount of time the schedule can fluctuate before affecting the end date. Negative float indicates critical path effort that may require management action such as overtime, second or third shifts, or resequencing of work. Moreover, if activities are not properly sequenced with logical links, it is not certain whether the critical path--which represents the chain of dependent activities with the longest total duration--is valid. Table 4 summarizes the results of our review of the four programs. Table 4: Extent to Which Program Schedules Met Best Practices: Best practice: 1. Capturing all activities; Extent best practice met: DEAMS: Partially; Extent best practice met: ECSS[A]: Substantially; Extent best practice met: GFEBS: Substantially; Extent best practice met: GCSS-Army: Partially. Best practice: 2. Sequencing all activities; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Partially; Extent best practice met: GFEBS: Partially; Extent best practice met: GCSS-Army: Partially. Best practice: 3. Assigning resources to all activities; Extent best practice met: DEAMS: Fully met; Extent best practice met: ECSS[A]: Minimally; Extent best practice met: GFEBS: Not Met; Extent best practice met: GCSS-Army: Substantially. Best practice: 4. Establishing the duration of all activities; Extent best practice met: DEAMS: Substantially; Extent best practice met: ECSS[A]: Substantially; Extent best practice met: GFEBS: Fully Met; Extent best practice met: GCSS-Army: Fully Met. Best practice: 5. Integrating schedule activities horizontally and vertically; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Partially; Extent best practice met: GFEBS: Minimally; Extent best practice met: GCSS-Army: Partially. Best practice: 6. Establishing the critical path for all activities; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Partially; Extent best practice met: GFEBS: Partially; Extent best practice met: GCSS-Army: Partially. Best practice: 7. Identifying reasonable float between activities; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Partially; Extent best practice met: GFEBS: Minimally; Extent best practice met: GCSS-Army: Substantially. Best practice: 8. Conducting a schedule risk analysis; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Not met; Extent best practice met: GFEBS: Not met; Extent best practice met: GCSS-Army: Minimally. Best practice: 9. Updating schedule using logic and durations to determine dates; Extent best practice met: DEAMS: Minimally; Extent best practice met: ECSS[A]: Partially; Extent best practice met: GFEBS: Partially; Extent best practice met: GCSS-Army: Substantially. Sources: GAO analysis based on data provided by the PMOs. Note: "Not met" means the program provided no evidence that satisfies any of the criterion. "Minimally" means the program provided evidence that satisfies a small portion of the criterion. "Partially" means the program provided evidence that satisfies about half of the criterion. "Substantially" means the program provided evidence that satisfies a large portion of the criterion. "Fully met" means the program provided evidence that completely satisfies the criterion. [A] In reviewing ECSS we analyzed two project schedules: (1) solutions development and (2) reports, interfaces, conversions, and extensions (RICE). We analyzed two schedules because the ECSS IMS is made up of 46 individual project schedules. The ratings were exactly the same for the nine practices. [End of table] Highlighted below are examples of the specific weaknesses we found in each of the nine best practices.[Footnote 53] Appendix IV contains a detailed discussion of the extent to which the four ERPs we analyzed met the nine best practice criteria. * Capturing all activities. A schedule should reflect all activities as defined in the program's work breakdown structure to include activities to be performed by the government and the contractor. Our analysis found that the ERP program schedules differed in the extent to which they capture all activities, as well as in the integration of government and contractor activities. The DEAMS PMO does not have a single schedule that integrates government and contractor activities. While the PMO maintains internal schedules that reflect government- only activities, these activities are not linked to contractor activities. In addition, many contractor activities within the DEAMS schedule are not mapped to the work breakdown structure, hampering management's ability to ensure all effort is included in the schedule. While the GCSS-Army schedule identifies contractor activities, it contains only key government milestones for the program. Other government activities, such as testing events and milestones beyond December 2010, are not captured in the schedule. The ECSS program schedule contains detailed activities associated with government effort and contractor effort. However, the government activities are not fully linked to contractor activities, so that updates to government activities do not have a direct impact on scheduled contractor activities. While the GFEBS's schedule captures government and contractor activities, dependencies between key milestones in deployment, software release, and maintenance are not linked, thereby precluding a comprehensive view of the entire program. Without fully integrating government activities with contractor activities, the schedule will not be able to reliably estimate the date the program is to be finished if a significant amount of key activities are not adequately captured. * Sequencing all activities. The schedule should be planned so that it can meet program critical dates. To meet this objective, activities need to be logically sequenced in the order that they are to be carried out and no artificial date constraints should be included in the schedule. In particular, activities that must finish prior to the start of follow-on activities (i.e., predecessor activities), as well as activities that cannot begin until other activities are completed (i.e., successor activities), should be identified. None of the contractor schedules we assessed fully met the criteria for sequencing all activities. For example, the DEAMS schedule has over 60 percent of the remaining activities missing logic links to predecessor or successor activities. Missing predecessors or successors reduce the credibility of the calculated dates. The DEAMS schedule also has date constraints[Footnote 54] that keep the schedule from responding correctly to changes. The ECSS schedule has 78 instances of unusual logic that cause activities to finish at the same time that their predecessor activities start.[Footnote 55] The GCSS-Army schedule has constraints on 1,503 of the remaining activities that keep the schedule from responding to changes. Moreover, the GFEBS schedule has date constraints and linked summary activities that interfere with the critical path.[Footnote 56] Missing or incorrect logic reduces the credibility of the calculated dates in the schedule because the schedule will not reflect the effects of slipping activities on the critical path, scheduled resources, or scheduled start dates of future activities. * Assigning resources to all activities. The schedule should realistically reflect what resources (i.e., labor, material, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist. Because of the fixed price contractual arrangements with ERP contractors, resources are not reflected in the periodic updates of the schedules submitted to the PMOs. While the GCSS-Army IMS does not include resources, scheduled activities can be traced to control account plans which have resources laid out by month by labor category. On the other hand, the DEAMS PMO provided evidence that resources were assigned to activities in the schedule. In the case of ECSS, PMO officials stated the contractor assigned resources to scheduled activities, but we were not able to verify whether resources were assigned. The GFEBS contractor's schedules had resources assigned to activities in earlier releases of the system, but according to the PMO, resources are no longer assigned to activities. Without resource information, DOD management has insufficient insight into current or projected over-allocation of contractor resources, thus increasing the risk of slippage in the estimated completion date. * Establishing the duration of all activities. The schedule should reflect how long each activity will take to execute and activity durations should be as short as possible with specific start and end dates. The four programs properly reflected how long each activity should take to execute. In addition, activities were generally shorter than 44 working days--or 2 working months--which represents best practices for activity durations. * Integrating schedule activities horizontally and vertically. The schedule should be integrated horizontally and vertically. Horizontal integration means that the schedule links the products and outcomes associated with already-sequenced activities. Horizontal integration also demonstrates that the overall schedule is rational, planned in a logical sequence or to reflect interdependencies between work and planning packages and provides a way to evaluate current status. When schedules are vertically integrated, lower-level schedules are clearly traced to upper-tiered milestones, allowing for total schedule integration and enabling different teams to work to the same schedule expectations. The program schedules we assessed partially met the criteria for horizontal and vertical integration. In general, as discussed earlier, issues with missing or convoluted logic and artificially constrained dates prevent the program schedules from being horizontally integrated. Schedules that are not horizontally integrated may not depict relationships between different program elements and product handoffs. While ECSS and GCSS-Army program schedules are vertically integrated, the inability to clearly trace lower-level schedules to upper-tiered milestones prevent the DEAMS and GFEBS program schedules from being fully vertically integrated. * Establishing the critical path. The establishment of a critical path--the longest duration path through the sequenced list of activities--is necessary for examining the effects of any activity slipping along this path. The calculation of a critical path is directly related to the logical sequencing of events. Missing or convoluted logic and artificially constrained dates prevent the calculation of a valid critical path, and can mark activities as critical that are not truly critical. The program schedules either partially or minimally met the criteria for establishing a critical path. While the ECSS PMO has insight into detailed contractor activities, officials acknowledged that it is difficult to establish a critical path using the program's current schedule. Instead, the program tracks high-level milestones in a separate schedule and officials stated that they are in the process of revamping the program's work breakdown structure in order to establish a clearer relationship between work products and hence a more accurate critical path. While the GFEBS PMO stated that it receives weekly updates from the contractor and manages to a critical path, our analysis of GFEBS concluded that the critical path was not reliable because of artificial date constraints and unrealistic float.[Footnote 57] Conversely, GCSS-Army and DEAMS officials stated that regardless of insight into detailed contractor activities, a critical path is not possible because it would be too complex. However, as program complexity increases, so must the schedule's sophistication. Further, our analysis of the DEAMS schedule found that it would be impossible to develop a critical path because 60 percent of the remaining activities are missing logic links. Likewise, our analysis found that a critical path within the GCSS-Army schedule is not possible because of artificial date constraints placed on key activities. While the level of complexity in an ERP is daunting, a critical path through at least a higher-level version of the detailed schedule would assist management in identifying which slipped tasks will have detrimental effects on the completion date. By managing to pre-defined, constrained dates instead of a critical path, management does not have a clear picture of the tasks that must be performed to achieve the target completion date. * Identifying reasonable float between activities. The schedule should identify float--the time that a predecessor activity can slip before the delay affects successor activities--so that schedule flexibility can be determined. As a general rule, activities along the critical path typically have the least amount of float. The DEAMS, ECSS, and GFEBS schedules did not meet the criteria for identifying reasonable float. The missing or convoluted logic and artificially constrained dates identified above prevent the proper calculation of float, which in turn affects the identification of a valid critical path. Without proper insight into float, management cannot determine the flexibility of tasks and therefore cannot properly reallocate resources from tasks that can safely slip to tasks that cannot slip without adversely affecting the estimated program completion date. * Conducting a schedule risk analysis. A schedule risk analysis uses statistical techniques to predict a level of confidence in meeting a completion date. The purpose of the analysis is to develop a probability distribution of possible completion dates that reflect the project and its quantified risks. This analysis can help management to understand the most important risks and to focus on mitigating these risks. We found that none of the PMOs have explicitly linked program risks to their schedule in the form of a schedule risk analysis. The ECSS PMO stated that it actively monitors schedule risk, but it has not performed a schedule risk analysis. The GFEBS PMO stated that while schedule risks have been discussed in team meetings, it has not performed a formal schedule risk analysis. However, the GFEBS PMO stated that it is open to improving in the area of schedule risk analysis. The DEAMS PMO stated that while it has tied risks to activities the program considers to be on the critical path, a formal schedule risk analysis was not performed because the schedule provided by the contractor lacks a sufficient level of detail to do such an analysis. The GCSS-Army contractor recently conducted a high-level schedule risk analysis on two major milestones. In addition, GCSS-Army PMO officials acknowledged the importance of a detailed schedule risk analysis and stated that the PMO intends to include this requirement in the contract within the next few months. A schedule risk analysis is important because it allows high-priority risks to be identified and mitigated, and the level of confidence in meeting projected completion dates can be predicted. Without a schedule risk analysis, the PMO cannot reliably determine the level of confidence in meeting the completion date. However, if the schedule risk analysis is to be credible, the program must have a quality schedule that reflects reliable logic and clearly identifies the critical path--conditions that none of the ERP schedules met. * Updating the schedule using logic and durations to determine dates. The schedule should use logic and durations in order to reflect realistic start and completion dates. The schedule should be continually monitored to determine when forecasted completion dates differ from the planned dates, which can be used to determine whether schedule variances will affect future work. There are differences in the four programs' ability to update the schedule using logic and durations to determine schedule dates. For example, the GCSS-Army schedule substantially met the criteria for updating the schedule, but the GFEBS schedule has over 100 instances of activities that should have occurred, yet have no actual start dates or finish dates. In addition, the ECSS schedules had a status date of January 1, 2010--a federal holiday--while schedules provided to the DEAMS program office by the contractor did not have a status date. A status date denotes the date of the latest update to the schedule and therefore defines the point in time at which completed work and remaining work are calculated. Without a valid status date, management is not able to determine what work is completed and what work is remaining. An invalid or missing status date is also an indication that management is not using the schedule to effectively oversee and monitor the effort. Furthermore, maintaining the integrity of the schedule logic is not only necessary to reflect true status, but is also required before conducting a schedule risk analysis. Each of the PMOs acknowledged the importance of many of the scheduling best practices, but stated that its ability to meet the prescribed best practices is limited because of the complexity of the ERP development process and the use of the firm-fixed price contract. Under the terms of these firm-fixed price contracts, the contractors are not required to provide detailed and timely scheduling data, which are essential for preparing order and an accurate and reliable schedule for the implementation of the system. While some of the necessary information is not being provided by the contractor, this does not relieve the PMOs of the responsibility for developing an IMS that fully meets prescribed best practices. Without the development of an IMS that meets scheduling best practices, the PMOs and the department are not positioned to adequately monitor and oversee the progress of the billions of dollars being invested in the modernization of DOD's business systems. Lacking a credible IMS, management is unable to predict, with any degree of confidence, whether the estimated completion date is realistic. An integrated schedule is key in managing program performance and is necessary for determining what work remains and the expected cost to complete it. A schedule delay can also lead to an increase in the cost of the project because, for example, labor, supervision, facilities, and escalation cost more if the program takes longer. A schedule and cost risk assessment recognizes the interrelationship between schedule and cost and captures the risk that schedule durations and cost estimates may vary. But without a fully integrated master schedule, the full extent of schedule uncertainty is not known, and therefore cannot be incorporated into the cost uncertainty analysis as schedule risk. Subsequent to the completion of our field work, ECSS and GFEBS PMOs provided updated schedules for assessment, but we were unable to perform a detailed evaluation of the updated schedules. Program officials for ECSS and GFEBS indicated that the updated schedules addressed some areas in which their previous schedules were deficient according to GAO's assessment of the nine scheduling best practices. In response to limitations that we identified and shared with the GFEBS PMO, the program office enacted several formal changes to its existing schedule. The ECSS PMO provided us with an updated IMS that contains details on future activities beyond the scheduled activities we originally assessed. Although we did not assess the new schedule, according to the ECSS PMO, the updated schedule is an improvement over past versions of the ECSS schedule and addresses many of the deficiencies GAO identified in the earlier version. Although Cost Estimates Meet Most Best Practices, the Lack of Sensitivity and Uncertainty Analyses Results in Estimates That May Not Be Credible: We have identified[Footnote 58] four characteristics of a reliable cost estimate (1) well-documented, (2) comprehensive, (3) accurate, and (4) credible. The four characteristics encompass 12 best practices for effective program cost estimates that are identified in appendix V. The results of our review of the DEAMS, ECSS, GFEBS, and GCSS-Army cost estimates are summarized in table 5. Table 5: Extent Cost Estimates Met Best Practices: Best practice: Well-documented; DEAMS: Substantially; ECSS: Substantially; GFEBS: Fully met; GCSS-Army: Substantially. Best practice: Comprehensive; DEAMS: Fully met; ECSS: Fully met; GFEBS: Fully met; GCSS-Army: Substantially. Best practice: Accurate; DEAMS: Fully met; ECSS: Substantially; GFEBS: Substantially; GCSS-Army: Partially. Best practice: Credible; DEAMS: Fully met; ECSS: Partially; GFEBS: Minimally; GCSS-Army: Partially. Sources: GAO analysis based on information provided by the PMOs. Note: "Not met" means the program provided no evidence that satisfies any of the criterion. "Minimally" means the program provided evidence that satisfies a small portion of the criterion. "Partially" means the program provided evidence that satisfies about half of the criterion. "Substantially" means the program provided evidence that satisfies a large portion of the criterion. "Fully met" means the program provided evidence that completely satisfies the criterion. [End of table] Highlighted below are examples of the specific strengths and weaknesses we found in each of the four best practices. Appendix V contains a detailed discussion of the extent to which the four ERPs we analyzed met the four best practices criteria. * Well-documented. The cost estimates should be supported by detailed documentation that describes the purpose of the estimate, the program background and system description, the scope of the estimate, the ground rules and assumptions, all data sources, estimating methodology and rationale, and the results of the risk analysis. Moreover, this information should be captured in such a way that the data used to derive the estimate can be traced back to, and verified against, their sources. The cost estimates for DEAMS, ECSS, GFEBS, and GCSS-Army are well-documented. The cost estimates have clearly defined purposes and are supported by documented descriptions of key program or system characteristics (e.g., relationships with other systems, performance parameters). Additionally, they capture in writing such things as the source data used and their significance, the calculations performed and their results, and the rationale for choosing a particular estimating method or reference. This information is captured in such a way that the data used to derive the estimate can be traced back to, and verified against, the sources. The final cost estimates are reviewed and accepted by management on the basis of confidence in the estimating process and the estimate produced by the process. * Comprehensive. The cost estimates should include costs of the program over its full life cycle, provide a level of detail appropriate to ensure that cost elements are neither omitted nor double-counted, and document all cost-influencing ground rules and assumptions. We found that cost estimates for DEAMS, ECSS, GFEBS, and GCSS-Army are comprehensive. The cost estimates include both government and contractor costs over the program's life cycle, from the inception of the program through design, development, deployment, and operation and maintenance to retirement. They also provide an appropriate level of detail to ensure that cost elements are neither omitted nor duplicated and include documentation of all cost- influencing ground rules and assumptions. * Accurate. The cost estimates should be based on an assessment of most likely costs (adjusted for inflation), documented assumptions, and historical cost estimates and actual experiences on other comparable programs. Estimates should be cross-checked against an independent cost estimate for accuracy, double counting, and omissions. In addition, the estimates should be updated to reflect any changes. Our analysis also found the cost estimates for DEAMS, ECSS, and GFEBS to be accurate. The cost estimates provide for results that are unbiased and are not overly conservative or optimistic. In addition, the cost estimates are updated regularly to reflect material changes in the program, and steps are taken to minimize mathematical mistakes and their significance. Among other things, the cost estimates are grounded in historical record of cost estimating and actual experiences on comparable programs. Our analysis found the cost estimate for GCSS-Army to be partially accurate because we could not verify how actual incurred costs were used to update the cost estimate. * Credible. The cost estimates should discuss any limitations of the analysis because of uncertainty, or biases surrounding data or assumptions. Risk and uncertainty analysis should be performed to determine the level of risk associated with the estimate. Further, the estimate's results should be cross-checked against an independent cost estimate.[Footnote 59] While we found that the ERP programs were generally following the cost estimating best practices, our analysis also found that the cost estimates for ECSS, GFEBS, and GCSS-Army are not fully credible. As stipulated in OMB and DOD guidance, ECSS and GCSS-Army did not include a sensitivity analysis, and GFEBS did not include a sensitivity analysis or a cost risk and uncertainty analysis. In the case of GFEBS, our July 2007 report[Footnote 60] noted that a sensitivity analysis had not been developed in calculating the life-cycle cost estimate. Cost estimates should discuss any limitations of the analysis because of uncertainty or biases surrounding data and assumptions. Major assumptions should be varied, and other outcomes recomputed to determine how sensitive they are to changes in the assumptions. Having a range of costs around a point estimate is more useful to decision makers because it conveys the level of confidence in achieving the most likely cost and also informs them of cost, schedule, and technical risks. In addition, as discussed earlier, because each of the four programs we assessed did not meet best practices for schedule estimating, none of the cost estimates could be considered credible because they did not assess the cost effects of schedule slippage. While individual phases of a multi- phased project may be completed on time, the project as a whole can be delayed, and phases that are not part of an IMS may not be completed efficiently which could result in future cost overruns. A reliable cost estimate is a key variable in calculating return on investment, and it provides the basis for informed investment decision making, realistic budget formulation and program resourcing, meaningful progress measurement, proactive course correction, and accountability for results. According to OMB[Footnote 61] programs must maintain current and well-documented cost estimates, and these estimates must encompass the full life cycle of the program. OMB states that generating reliable cost estimates is a critical function necessary to support OMB's capital programming process. Without reliable estimates, programs are at increased risk of experiencing cost increases, missed deadlines, and performance shortfalls. ERP Success in Transforming Business Operations Has Not Been Defined or Measured: DOD has not yet defined success for ERP implementation in the context of business operations and in a way that is measurable. Accepted practices in system development include testing the system in terms of the organization's mission and operations--whether the system performs as envisioned at expected levels of cost and risk when implemented within the organization's business operations. The Clinger-Cohen Act of 1996 recognizes the importance of performance measurement in requiring agencies to (1) establish goals for improving the efficiency and effectiveness of agency operations and (2) ensure that performance measurements determine how well the information technology supports programs of the executive agency.[Footnote 62] DOD also has recognized the importance of performance measures, which the department directs should be (1) written in terms of desired outcomes, (2) quantifiable, (3) able to measure the degree to which the desired outcome is achieved, (4) independent of the particular automated system tested and not focused on system-performance criteria, and (5) designed to include benefits to the DOD component and the enterprise. In regard to the ERPs, measures determining whether the system is being used as expected and is providing the desired benefits from a business perspective will vary depending on the specific type of business functions the system is performing. For example, in a logistical system, the new system and its processes may be expected to help accomplish such items as (1) reducing inventory levels; (2) increasing the inventory turnover rate which shows that the items actually needed are the items being procured; and (3) increasing the accuracy of the projected completion dates for repair projects which allows for better equipment utilization. On the other hand, a financial system may measure benefits in such areas as (1) reducing prompt payment penalties; (2) improving the financial statement preparation process by having the system automatically generate the statements which reduces the potential for manual error; and (3) improving management oversight of an entity's operations and providing the detailed data necessary to evaluate abnormalities that may be detected. Developing and using specific performance measures to evaluate a system effort should help management understand whether the expected benefits are being realized. While the definition of success and performance measures for DOD's ERPs will differ between organizational levels, components, and subcomponents, our previous work has shown that performance measures should be aligned toward a shared direction. In this regard, all members of the organization need to understand the ultimate result to be achieved and all parties should work toward the same goal and desired results. This alignment should extend throughout the organization and cover the activities that an entity is expected to perform to support the intent of the program.[Footnote 63] DOD has not taken actions to align the definitions and related performance measures used by its components to measure progress and determine success. The DCMOs told us that they had not yet developed a DOD-wide definition of success or related performance measures for ERPs. While acknowledging the importance of these practices, the officials told us that they are still in the early stages of implementing processes for managing and overseeing their business systems modernization efforts, in accordance with the fiscal year 2010 Defense Authorization Act. Successful implementation of the ERPs is critical to transforming business operations. Without defining ERP success in terms of support for mission and business operations and establishing the related performance measures, the military services and the department cannot ensure that the performance of deployed ERPs has been realistically and accurately measured. Our April 2010 report,[Footnote 64] which focused on the second deployment of LMP at the Corpus Christi and Letterkenny Army Depots, illustrates the importance of establishing performance measures. Based on our observations at the Corpus Christi and Letterkenny Army Depots, we found that the Army's measures for assessing LMP implementation at the two deployment sites did not accurately reflect whether the locations were able to perform their day-to-day operations using LMP as envisioned. Rather, the measures used by the Army assessed the success at the two locations from a system-software perspective. While this is important, performance measures from a business perspective were not considered to determine whether the depots were able to use LMP to perform their mission to repair items. Without performance measures to evaluate how well these systems are accomplishing their desired goals, DOD decision makers including program managers do not have all the information they need to evaluate their investments to determine whether the individual programs are helping DOD achieve business transformation and thereby improve upon its primary mission of supporting the warfighter. Conclusions: Modernizing the department's business systems is a critical part of transforming DOD's business operations, addressing some of its high- risk areas, and providing more accurate and reliable financial information to the Congress on the results of DOD's operations. However, DOD continues to experience difficulties that hinder its ability to implement these efforts on time and within budget. The department has not followed best practices and developed a reliable IMS for several of these modernization efforts. As a result, it lacks the assurance that these ERPs will be completed by the projected date. Furthermore, while DOD generally followed best practices in developing the programs' cost estimates for these efforts, with the exception of DEAMS, none of the programs has prepared a sensitivity analysis. The lack of a sensitivity analysis increases the chances that decisions will be made without a clear understanding of the possible impact on the estimates of costs and benefits of each program. In addition, because each of the four programs we assessed did not meet best practices for schedule estimating, none of the cost estimates could be considered credible because they did not assess the cost effects of schedule slippage. It is critical to correct the underlying issues to help ensure that the billions of dollars spent annually are being used in the most efficient and effective manner. While modernizing its business systems is not a risk-free endeavor, additional funds spent because of schedule slippages are funds that could have been available for other departmental priorities. Furthermore, the longer it takes to implement these critical business systems, the longer the department will continue to use its existing duplicative, stovepiped systems environment and further erode the estimated savings that were to accrue to DOD as a result of modernizing its business systems. Additionally, the department has not defined the measures to ascertain if the systems are providing the desired functionality to achieve DOD's business transformation goals. DOD has stated that the successful implementation of the ERPs is critical to transforming its business operations and addressing some of its high-risk areas. However, we found that the department has not yet developed performance measures to ascertain whether the systems, once implemented, are providing the intended functionality. If the systems cannot be used for their intended purpose, transformation will be difficult if not impossible to achieve and the billions of dollars being invested in these systems may not generate the benefits and efficiencies as intended. Further, we reaffirm our prior recommendations related to the actions needed to improve the department's management and oversight of the ERPs. Recommendations for Executive Action: To strengthen DOD's management oversight and accountability over business system investments and help provide for the successful implementation of the ERPs, we recommend that the Secretary of Defense take the following eight actions: * Direct the Secretary of the Army to ensure that the Chief Management Officer of the Army directs the PMO for the GFEBS to develop an IMS that fully incorporates best practices. The schedule should: - sequence all activities, - assign resources to all activities, - integrate schedule activities horizontally and vertically, - establish the critical path for all activities, - identify float between activities, - conduct a schedule risk analysis, and: - update schedule using logic and durations to determine dates. * Direct the Secretary of the Army to ensure that the Chief Management Officer of the Army direct the PMO for GCSS-Army to develop an IMS that fully incorporates best practices. The schedule should: - capture all activities, - sequence all activities, - integrate schedule activities horizontally and vertically, - establish the critical path for all activities, and: - conduct a schedule risk analysis. *Direct the Secretary of the Air Force to ensure that the Chief Management Officer of the Air Force directs the PMO for DEAMS to develop an IMS that fully incorporates best practices. The schedule should: - capture all activities, - sequence all activities, - integrate schedule activities horizontally and vertically, - establish the critical path for all activities, - identify float between activities, - conduct a schedule risk analysis, and: - update schedule using logic and durations to determine dates. * Direct the Secretary of the Air Force to ensure that the Chief Management Officer of the Air Force directs the PMO for ECSS to develop an IMS that fully incorporates best practices. The schedule should: - sequence all activities, - assign resources to all activities, - integrate schedule activities horizontally and vertically, - establish the critical path for all activities, - identify float between activities, - conduct a schedule risk analysis, and: - update schedule using logic and durations to determine dates. * Direct the Secretary of the Army to ensure that the Chief Management Officer of the Army directs the PMO for GFEBS to update the cost estimates by preparing sensitivity and risk and uncertainty analyses using best practices. * Direct the Secretary of the Army to ensure that the Chief Management Officer of the Army directs the PMO for GCSS-Army to update the cost estimates by using actual cost and preparing a sensitivity analysis using best practices. * Direct the Secretary of the Air Force to ensure that the Chief Management Officer of the Air Force directs the PMO for ECSS to update the cost estimates by preparing a sensitivity analysis using best practices. * Direct the department's Chief Management Officer and the chief management officers of the military departments to establish performance measures based on quantitative data that will enable the department to assess whether each respective military service's ERP efforts are providing the intended business capabilities to the system users. Agency Comments and Our Evaluation: DOD provided written comments on a draft of this report. In its comments, DOD concurred with the eight recommendations and cited actions planned to address them. For example, the department recognized the importance of an integrated master schedule as a key program management tool fundamental to having a reliable program schedule. The department stated that the appropriate military department Chief Management Officer will direct program managers to implement the recommendations and further noted that the Chief Management Officer will oversee the implementation of the recommendations. Further, DOD stated that guidance will be issued requiring DOD business systems investments to include performance measures that can be used to assess the expected benefits of the investments. Additionally, the department noted that the performance measures will be incorporated into DOD's Business Enterprise Architecture. We are sending copies of this report to the Secretary of Defense; the Secretary of the Army; the Secretary of the Navy; the Secretary of the Air Force; the Deputy Secretary of Defense; the Under Secretary of Defense (Comptroller); the Chief Management Officer of the Army, the Navy, and the Air Force; the program management office for each business system that was included in the audit; and other interested congressional committees and members. This report also is available at no charge on the GAO Web site at [hyperlink, http://www.gao.gov]. Please contact Asif A. Khan at (202) 512-9095 or khana@gao.gov or Nabajyoti Barkakati at (202) 512-4499 or barkakatin@gao.gov if you or your staff have questions on matters discussed in this report. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. Key contributors to this report are listed in appendix VI. Signed by: Asif A. Khan: Director: Financial Management and Assurance: Signed by: Nabajyoti Barkakati: Chief Technologist: Applied Research and Methods Center for Science, Technology, and Engineering: List of Requesters: The Honorable Evan Bayh: Chairman: The Honorable Richard Burr: Ranking Member: Subcommittee on Readiness and Management Support: Committee on Armed Services: United States Senate: The Honorable Thomas R. Carper: Chairman: The Honorable John McCain: Ranking Member: Subcommittee on Federal Financial Management, Government Information, Federal Services, and International Security: Committee on Homeland Security and Governmental Affairs: United States Senate: The Honorable George Voinovich: Ranking Member: Subcommittee on Oversight of Government Management, the Federal: Workforce, and the District of Columbia: Committee on Homeland Security and Governmental Affairs: United States Senate: The Honorable Tom Coburn: Ranking Member: Permanent Subcommittee on Investigations: Committee on Homeland Security and Governmental Affairs: United States Senate: [End of section] Appendix I: Objective, Scope, and Methodology: Our objectives were to (1) provide the status as of December 31, 2009 of the nine enterprise resource planning (ERP) systems that the Department of Defense (DOD) identified as essential to transforming its business operations; (2) assess the scheduling and cost estimating practices of selected ERPs to determine the extent to which the program management offices (PMO) were applying best practices; and (3) ascertain whether DOD and the military departments have defined the performance measures to determine whether the systems will meet their intended business capabilities. To address the first objective, we obtained and reviewed information provided by the PMO responsible for the nine ERP efforts.[Footnote 65] More specifically, we obtained data related to the following for each program: (1) when the program was initiated, (2) the program's accountable official, (3) the purpose of the program, (4) the cost of the program, (5) the implementation schedule, (6) the number of legacy systems intended to be replaced, (7) the cost of the legacy systems, (8) the date the program was last certified, and (9) the conditions placed on the program by the various review boards. For the purposes of this report, we did not include information on the Defense Logistics Agency Business System Modernization/Enterprise Business System. According to DOD the Business System Modernization effort was fully implemented in July 2007 and transformed how the agency conducts its operations in five core business processes: order fulfillment, demand and supply planning, procurement, technical/quality assurance, and financial management. Subsequently, in September 2007, the name of the program was changed to the Enterprise Business System, which is a continuation of the ERP's capabilities to support internal agency operations. We also reviewed various DOD documents such as the Enterprise Transition Plans issued in September 2008 and December 2009, the Defense Business Systems Management Committee meeting minutes and briefings, the Selected Capital Investment Reports, which are prepared in support of the funding requests for the ERPs, the Congressional Report on Defense Business Operations for fiscal years 2009 and 2010 and the Major Automated Information System Reports for fiscal years 2008 and 2009--to corroborate the information obtained from the PMOs. In instances where we identified discrepancies, we followed up with the PMOs to obtain an explanation. Most of the financial information in this report was obtained through interviews with or responses to GAO questions from knowledgeable PMO officials for the nine ERP systems. As part of the first objective, we also reviewed past GAO reports[Footnote 66] that were specific to the department's efforts to implement the nine ERPs to identify prior recommendations and assess DOD's progress in addressing the 19 recommendations discussed in these reports. To assess the scheduling and cost estimating practices of selected ERPs, we selected the General Fund Enterprise Business System (GFEBS), the Global Combat Support System-Army (GCSS-Army), the Defense Enterprise Accounting and Management Systems (DEAMS), and the Expeditionary Combat Support System (ECSS). The other programs were excluded because (1) the Logistics Modernization Program is expected to be fully deployed soon, (2) it is too soon to assess the department's integrated personnel and pay efforts because of the recent change in the Defense Integrated Military Human Resources System implementation strategy and the Defense Agencies Initiative has yet to develop its implementation schedule for the various defense agencies, and (3) we reported[Footnote 67] on concerns with the Marine Corps and Navy schedule and cost estimating practices in July 2008 and September 2008, respectively. In performing our analysis for the four ERPs, we reviewed the schedules and cost estimates available at the time of our review and evaluated them using the criteria set forth in GAO's cost guide.[Footnote 68] In using the guide, we determined the extent to which each schedule was prepared in accordance with the best practices[Footnote 69] that are fundamental to having a reliable schedule. In assessing each program's cost estimates, we used the GAO cost guide to evaluate the PMOs' estimating methodologies, assumptions, and results to determine whether the cost estimates were comprehensive, accurate, well-documented, and credible. We discussed the results of our assessments with the PMOs, lead schedulers, and cost estimators. To address the third objective, we obtained and reviewed the 2009 and 2010 reports[Footnote 70] on business transformation submitted to congressional defense committees by each military service to determine the extent to which these reports included performance measures. In addition, we met with the military departments' deputy chief management officers to obtain an understanding of how they define success in terms of their respective ERPs. We also met with the personnel within the department's DCMO office and the Director, Business Transformation Agency, to obtain an understanding of their respective roles and responsibilities for the implementation of the ERPs within the department. We conducted this performance audit from June 2009 through October 2010 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix II: Comments from the Department of Defense: Deputy Chief Management Officer: 9010 Defense Pentagon: Washington, DC 20301-9010: September 23, 2010: Mr. Asif A. Khan: Director,Financial Management and Assurance: U.S. Government Accountability Office: 441 G Street, NW Washington, DC 20548: Dear Mr. Khan: The Department of Defense (DoD) response to the U.S. Government Accountability Office (GAO) draft report 10-951, "DOD Business Transformation: Improved Management Oversight of Business System Modernization Needed," dated August 25, 2010 (GAO Code 197086), is contained in this letter. The Department concurs with all GAO recommendations contained in the draft report. [The report number is now GAO-11-53] Your audit highlighted the need to incorporate best practices into development of an integrated master schedule (IMS). The Department recognizes IMS as a key program management tool fundamental to program schedule reliability. The appropriate Military Department Chief Management Officers will direct Program Managers to implement these recommendations and have oversight over implementation. Additionally, the Department will issue guidance requiring DoD business systems investments to include performance measures which can be used to assess expected business benefits. For strategic alignment, these measures will also be incorporated into the Business Enterprise Architecture. Sincerely, Signed by: Elizabeth A. McGrath: [End of section] Appendix III: Status of DOD's Actions on Previous GAO Recommendations Related to Business Systems Modernization: [End of section] Tables 6 through 10 provide information on the status of DOD's actions to address our recommendations in previous reports. Table 6: Status of DOD's Actions to Address GAO Recommendations in GAO- 07-860: GAO recommendation: 1. The Secretary of Defense should direct the Secretary of the Army and the Director, Business Transformation Agency (BTA) to jointly develop a concept of operations that (1) clearly defines the ERP vision for accomplishing total asset visibility (TAV) within the Army; (2) addresses how its business systems and processes, individually and collectively, will provide the desired functionality to achieve total asset visibility; and (3) determines the desired functionality among the selected systems; DOD action taken to address the recommendation: The Army's March 2010 report to Congress[A] stated that the Army lacks a concept of operations that describes at a high level how the GFEBS, GCSS-Army, and LMP systems relate to each other and how information flows between and through the systems. Furthermore, the Army found that representatives from the three systems were not able to articulate (1) what specific data would be exchanged between the three systems and (2) which system would be considered the official system of record for master data that needed to be consistent between the three systems. The Army did not provide a timeframe for completing the concept of operations; Status of GAO recommendation: Open. GAO recommendation: 2. The Secretary of Defense should direct the Secretary of the Army and the Director, BTA to jointly develop policies, procedures, and processes to support the oversight and management of selected groupings of business systems that are intended to provide a specific capability or functionality, such as TAV from a portfolio perspective, utilizing indicators such as costs, schedule, performance, and risks; DOD action taken to address the recommendation: In June 2010, the Under Secretary of the Army established the Business Systems Information Technologies Executive Steering Group. The purpose of the group is to advise the Army Chief Management Officer on Army-wide requirements for the synchronization, integration, prioritization, and resourcing of Army business systems. The Army's efforts to establish an enterprisewide focus on systems investments should improve the Army's ability to oversee the billions of dollars it is investing in its business systems. The group meets the intent of the recommendation; Status of GAO recommendation: Closed. GAO recommendation: 3. The Secretary of Defense should direct the Secretary of the Army and the Director, BTA to jointly establish an independent verification and validation (IV&V) function for GFEBS, GCSS-Army, and LMP. Additionally, direct that all IV&V reports for each system be provided to Army management, the appropriate investment review board (IRB), and BTA; DOD action taken to address the recommendation: In August 2009, the Army awarded a contract to carry out the IV&V function for these systems. Under the contract, the contractor is to provide reports on each of the systems to the Program Executive Office Enterprise Information Systems, which reports to the Army's Deputy Chief Management Officer (DCMO). The Army's action to establish an IV&V function under the direction of the Army's DCMO, if fully and effectively implemented, should enable the Army to improve its management and oversight of its business systems investments. Given the responsibility of the Chief Management Officer for overseeing and monitoring the implementation of Army's business systems, the Army's action meets the intent of the recommendation; Status of GAO recommendation: Closed. GAO recommendation: 4. The Secretary of Defense should direct the Secretary of the Army and the Director, BTA to jointly require that any future GFEBS economic analysis identify costs and benefits in accordance with the criteria specified by DOD and Office of Management and Budget (OMB) guidance, to include a sensitivity analysis; DOD action taken to address the recommendation: While the Army has developed an updated economic analysis, it was not prepared in accordance with DOD and OMB guidance.[B] For example, the economic analysis did not include a sensitivity analysis or a cost uncertainty analysis. Cost estimates should discuss any limitations of the analysis because of uncertainty or biases surrounding data and assumptions. Major assumptions should be varied, and other outcomes recomputed to determine how sensitive they are to changes in the assumptions. Having a range of costs around a point estimate--the best guess at the cost estimate, given the underlying data--is more useful to decision makers because it conveys the level of confidence in achieving the most likely cost and also informs management on cost, schedule, and risks; Status of GAO recommendation: Open. GAO recommendation: 5. The Secretary of Defense should direct the Secretary of the Army and the Director, BTA to jointly direct that LMP utilize system testers that are independent of the LMP system developers to help ensure that the system is providing the users of the system the intended capabilities; DOD action taken to address the recommendation: The Army has stated that LMP system testers are now independent of the system developers. We are in the process of evaluating the Army's actions as part of our ongoing work on the third deployment of LMP; Status of GAO recommendation: Open. Source: GAO analysis of data provided by DOD. [A] U.S. Army, Report to Congress on Business Transformation (Mar. 1, 2010). [B] Department of Defense Instruction 7041.3 and OMB Circular No. A-94. [End of table] Table 7: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-822: GAO recommendation: 1. The Secretary of Defense direct the Secretary of the Navy to ensure that investment in the next acquisition phase of the program's first increment is conditional upon fully disclosing to program oversight and approval entities the steps under way or planned to address each of the risks discussed in the report, including the risk of not being architecturally compliant and being duplicative of related programs, not producing expected mission benefits commensurate with reliably estimated costs, not effectively implementing earned valued management (EVM), not mitigating known program risks, and not knowing whether the system is becoming more or less mature and stable. We further recommend that investment in all future Global Combat Support System-Marine Corps (GCSS-MC) increments be limited if the management control weaknesses that are the source of these risks, and which are discussed in the report, have not been fully addressed; DOD action taken to address the recommendation: An Enterprise Risk Assessment Methodology (ERAM)-based review was conducted on GCSS-MC, and the results were presented at a May 2009 IRB meeting. According to DOD, the assessment included a review of the program's risk management database and policies. The ERAM process identified seven risk areas, some of which relate to risks discussed in our report. DOD reported that the governance-related risks identified in our report require longer-term actions; while the program had nevertheless demonstrated compliance with its business enterprise architecture and that the IRB reviewed and certified compliance with the architecture in October 2009. DOD also reported that the program implemented a new risk management process in March 2009 and developed metrics related to system maturity and stability, such as metrics to track defects during developmental test and evaluation, and is tracking change requests and generating monthly trend analyses of each. In addition, DOD reported that the program is working closely with the Milestone Decision Authority, via the IRB, to correct management control weaknesses. As of October 2010, DOD had yet to provide the supporting documentation for the above actions taken by the department; Status of GAO recommendation: Open. GAO recommendation: 2. The Secretary of Defense direct the appropriate organization within DOD to collaborate with relevant organizations to standardize the cost element structure for the department's ERP programs and to use this standard structure to maintain cost data for its ERP programs, including GCSS-MC, and to use this cost data in developing future cost estimates; DOD action taken to address the recommendation: In April 2010, DOD reported that planning is underway within the BTA and the Office of Acquisition Resources and Analysis for development of a common set of high-level work elements, such as testing, design, and training, to augment detailed work breakdown structures developed by program managers for their respective ERP programs. DOD also stated that it plans to use the common set of high-level work elements, along with a common set of cost elements--buckets of cost types such as program management, technical labor, hardware, and software--to capture historical costs across ERP programs. DOD also stated that it still plans to track and maintain ERP cost data through the Business Capability Lifecycle Integrated Management Information Environment, and use the data to develop future cost estimates and an economic analysis. As of October 2010, DOD did not provide timeframes for completion of these actions; Status of GAO recommendation: Open. GAO recommendation: 3. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that the program's current economic analysis is adjusted to reflect the risks associated with it not reflecting cost data for comparable ERP programs, and otherwise not having been derived according to other key cost estimating practices, and that future updates to the GCSS-MC economic analysis similarly do so; DOD action taken to address the recommendation: In April 2010, DOD reported that the GCSS-MC program developed its Cost Analysis Requirements Document and Economic Analysis Development Plan in partnership with the Office of the Secretary of Defense (Cost Analysis and Program Evaluation) to ensure that the GCSS-MC economic analysis addresses DOD-wide assumptions and risks. DOD stated that the independent cost estimate prepared by the Naval Center for Cost Analysis and approved in January 2010 was risk adjusted and included cross-checks from similar ERP systems and models. As of October 2010, DOD had yet to provide the supporting documentation for the above actions; Status of GAO recommendation: Open. GAO recommendation: 4. To enhance GCSS-MC's use of EVM, we recommend that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that the program office (1) monitors the actual start and completion dates of work activities performed so that the impact of deviations on downstream scheduled work can be proactively addressed; (2) allocates resources, such as labor hours and material, to all key activities on the schedule; (3) integrates key activities and supporting tasks and subtasks; (4) identifies and allocates the amount of float time needed for key activities to account for potential problems that might occur along or near the schedule's critical path; (5) performs a schedule risk analysis to determine the level of confidence in meeting the program's activities and completion date; (6) allocates schedule reserve for high-risk activities on the critical path; and (7) discloses the inherent risks and limitations associated with any future use of the program's EVM reports until the schedule has been risk-adjusted; DOD action taken to address the recommendation: In April 2010, DOD reported that the schedule was rebaselined and is now used to monitor and report actual versus planned start and completion dates of work activities; allocate resources to activities; integrate key activities and supporting tasks and sub-tasks; identify and allocate the amount of float time needed for key activities, and allocate schedule reserve for high-risk activities on the critical path. DOD also reported that the program conducted a schedule risk analysis which resulted in more detailed task definitions, the ability to provide detailed weekly status reports to the program manager, and more effective analysis, monitoring and risk assessment of the program's scheduled activities and completion dates. As of October 2010, DOD had yet to provide the supporting documentation for the above actions; Status of GAO recommendation: Open. GAO recommendation: 5. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that the program office (1) adds each of the risks discussed in this report to its active inventory of risks, (2) tracks and evaluates the implementation of mitigation plans for all risks, (3) discloses to appropriate program oversight and approval authorities whether mitigation plans have been fully executed and have produced the intended outcome(s), and (4) only closes a risk if its mitigation plan has been fully executed and produced the intended outcome(s); DOD action taken to address the recommendation: In April 2010, DOD reported that the program office took a number of actions to strengthen risk management. First, it included all risks reported by GAO as well as risks identified through DOD's ERAM in its risk database. Second, program risks are now reviewed, tracked and managed on a continuous basis, and the program office conducts weekly risk meetings to track and evaluate mitigation plan implementation for all risks. Third, monthly risk boards are convened to discuss risks and mitigation plans with GCSS-MC senior leadership, and risks are not closed without risk board approval. Further, the program office meets monthly with the Program Executive Office for Enterprise Information Systems and quarterly with the Assistant Secretary of the Navy (Research Development and Acquisition) to discuss program risks. Also, the program office reports risks to other program oversight bodies, such as the Weapons Systems Lifecycle Management and Materiel Supply and Services Management IRB, and the Defense Business Systems Management Committee. Fourth, the program office revised its Risk Management Plan, in March 2009, to reflect these new processes and policies. As of October 2010, DOD had yet to provide the supporting documentation for the above actions; Status of GAO recommendation: Open. GAO recommendation: 6. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that the program office (1) collects the data needed to develop trends in unresolved system defects and change requests according to their priority and severity and (2) discloses these trends to appropriate program oversight and approval authorities; DOD action taken to address the recommendation: In April 2010, DOD reported that during system developmental test and evaluation, completed in October 2009, the program office developed metrics to track defects and correction of defects throughout the test period, and that the metrics were made available to the BTA and the Cost Analysis and Program Evaluation Office. DOD reported that the program office is also (1) collecting defect data over time across severity levels and using diagrams to show trends and (2) managing change requests according to its configuration management plan and generating trend analysis reports to track them. As of October 2010, DOD had yet to provide the supporting documentation for the above actions; Status of GAO recommendation: Open. Source: GAO analysis of data provided by DOD. [End of table] Table 8: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-866: GAO recommendation: 1. The Secretary of Defense direct the Secretary of the Air Force to direct Air Force program management officials for ECSS and DEAMS to ensure that risk management activities at all levels of the program are identified and communicated to program management to facilitate oversight and monitoring. Key risks described at the appropriate level of detail should include and not be limited to risks associated with interfaces, data conversion, change management, and contractor oversight; DOD action taken to address the recommendation: In July 2009, the DEAMS Program Charter noted that risk management activities at all levels of the program would be identified and communicated to the program manager to facilitate oversight and monitoring. The charter noted that program risk will include, but not be limited to, interfaces, data conversion, change management, and contractor oversight. The charter also notes that the risk management process will include risk identified by various reviews, including GAO audits. As of July 2010, ECSS was still in the process of revising its risk management plan to address our recommendation; Status of GAO recommendation: Open. GAO recommendation: 2. The Secretary of Defense direct the Secretary of the Air Force to direct the Air Force program management offices to test ECSS and DEAMS on relevant computer desktop configurations prior to deployment at a given location; DOD action taken to address the recommendation: The intent of the recommendation was to reduce program risk and ensure that when DEAMS and ECSS were deployed to a given location they would operate as intended. According to the DEAMS PMO, the PMO has performed appropriate testing prior to the system being operational and if necessary, changes are made prior to the implementation of the system. The Defense Finance and Accounting Service is also participating in the testing, thereby helping to ensure that the accounting information will process correctly. As of July 2010, ECSS had not yet become operational at a given location; Status of GAO recommendation: Open. Source: GAO analysis of data provided by DOD. [End of table] Table 9: Status of DOD's Actions to Address GAO Recommendations in GAO- 08-896: GAO recommendation: 1. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that future Navy ERP estimates include uncertainty analyses of estimated benefits, reflect the risks associated with not having cost data for comparable ERP programs, and are otherwise derived in full accordance with the other key estimating practices, and economic analysis practices discussed in this report; DOD action taken to address the recommendation: In July 2010, DOD reported that uncertainty analysis will be applied to the Navy ERP's benefit estimate in support of the next milestone review, Full Deployment Decision Review, planned for the first quarter of fiscal year 2011. The benefit estimation model is being updated to include variations among key cost drivers, such as labor category efficiency and legacy system sustainment difficulty factors, through the use of Monte Carlo simulation. In addition, DOD reported that the Navy ERP program is working with the Space and Naval Warfare Systems Command and Naval Center for Cost Analysis, as they conduct an independent assessment of the program's life-cycle cost estimate. According to DOD, the assessment will include a review of the risk/uncertainty approach and methodologies used to develop the cost estimate; Status of GAO recommendation: Open. GAO recommendation: 2. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that (1) an integrated baseline review on the last two releases of the first increment is conducted, (2) compliance against the 32 accepted industry earned value management (EVM) practices is verified, and (3) a plan to have an independent organization perform surveillance of the program's EVM system is developed and implemented; DOD action taken to address the recommendation: In July 2010, DOD reported that the Navy ERP program office conducted an integrated baseline review of its second release, which resulted in recommendations to mature and implement EVM processes. Because the third release is no longer a part of Navy ERP's program of record, this recommendation is not applicable to this release. In addition, DOD reported that the Navy Center for Earned Value Management planned to conduct surveillance of the Navy ERP's EVM system in September 2010, and that it would review compliance against the 32 accepted industry EVM practices; Status of GAO recommendation: Open. GAO recommendation: 3. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that the schedule (1) includes the logical sequencing of all activities, (2) reflects whether all required resources will be available when needed, (3) defines a critical path that integrates all three releases, (4) allocates reserve for the high-risk activities on the entire program's critical path, and (5) incorporates the results of a schedule risk analysis for all three releases and recalculates program cost and schedule variances to more accurately determine a most likely cost and schedule overrun; DOD action taken to address the recommendation: As of July 2010, Navy ERP continues to make progress in addressing this recommendation. For example, it is using metrics to track and logically link activities and account for resources and their availability, and it plans to conduct a schedule risk assessment in September 2010 so that reserves can be established for high-risk activities. Further, in July 2010, DOD reported that it was not feasible to define a critical path integrating all three releases because (1) key functionality deliverables for the first release were completed prior to the second release's development and (2) the third release was removed from Navy ERP's program of record. However, the March 2010 metrics report shows that not all activities are logically sequenced, which can affect the calculation of the critical path and finish date. Further, because the schedules are not integrated and personnel are assigned to activities across multiple releases, if deployment activities in one schedule were to be delayed, the other schedule that requires the same resources would likely also be delayed; Status of GAO recommendation: Open. GAO recommendation: 4. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to ensure that (1) the plans for mitigating the risks associated with converting data from legacy systems to Navy ERP and positioning the commands for adopting the new business processes embedded in the Navy ERP are re- evaluated in light of the recent experience with the Naval Air Systems Command (NAVAIR) and adjusted accordingly, (2) the status and results of these and other mitigation plans' implementation are periodically reported to program oversight and approval authorities, (3) these authorities ensure that those entities responsible for implementing these strategies are held accountable for doing so, and (4) each of the risks discussed in this report are included in the program's inventory of active risks and managed accordingly; DOD action taken to address the recommendation: The department has taken actions to address the intent of this recommendation. First, the Navy ERP program office reevaluated its plans for mitigating risks associated with data conversion and adopting new business processes. Second, the program manager and System Command officials report monthly to the Navy ERP Senior Integration Board (NESIB) on performance, and periodically brief oversight and approval authorities on the implementation of risk mitigation plans. Third, the NESIB requires actionable reporting on performance by the program manager and System Command officials, and the program manager is to report to the Milestone Decision Authority on implementation of risk mitigation strategies. Fourth, the program's risk inventory has been updated to include risks related to adopting new business processes and data conversion; Status of GAO recommendation: Closed. Source: GAO analysis of data provided by DOD. [End of table] Table 10: Status of DOD's Actions to Address GAO Recommendations in GAO-09-841: GAO recommendation: 1. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to (1) revise the Navy ERP procedures for controlling system changes to explicitly require that a proposed change's life-cycle cost impact be estimated and considered in making change request decisions and (2) capture the cost and schedule impact of each proposed change in the Navy ERP automated control tracking tool; DOD action taken to address the recommendation: The Navy ERP program updated its Enterprise Change Request Process and Procedures to explicitly require that a change's life-cycle cost impact be estimated as part of the change control process. In addition, the change control tracking tool now captures cost and schedule impact information. As a result, management of the Navy ERP's change control process has been strengthened. As a result, approval authorities should be provided key information needed to fully inform their decisions on whether to approve a change, thus decreasing the risk of unwarranted cost increases and schedule delays; Status of GAO recommendation: Closed. GAO recommendation: 2. The Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to (1) stop performance of the IV&V function under the existing contract and (2) engage the services of an IV&V agent that is independent of all Navy ERP management, development, testing, and deployment activities that it may review; DOD action taken to address the recommendation: According to DOD, the Navy ERP program office terminated the IV&V functions under the existing contract on September 30, 2009, and awarded a new IV&V contract in September 2010; Status of GAO recommendation: Closed. Source: GAO analysis of data provided by DOD. [End of table] [End of section] Appendix IV: Assessments of Four DOD ERP Programs' Integrated Master Schedules: This appendix provides the results of our analysis of the extent to which the processes and methodologies used to develop and maintain the four ERP integrated master schedules meet the nine best practices associated with effective schedule estimating.[Footnote 71] Tables 11, 12, 13, 14, and 15 provide the detailed results of our analyses of the program schedules for DEAMS, ECSS, GFEBS, and GCSS-Army compared to the nine best practices. "Not met" means the program provided no evidence that satisfies any of the criterion. "Minimally" means the program provided evidence that satisfies a small portion of the criterion. "Partially" means the program provided evidence that satisfies about half of the criterion. "Substantially" means the program provided evidence that satisfies a large portion of the criterion. "Fully met" means the program provided evidence that completely satisfies the criterion. Table 11: Analysis of the Air Force's DEAMS Program Schedule: Best practice: 1. Capturing all activities; Explanation: The schedule should reflect all activities as defined in the project's work breakdown structure, which defines in detail the work necessary to accomplish a project's objectives, including activities to be performed by both the owner and contractors; Criterion met: Partially; GAO analysis: Our analysis found that the DEAMS program schedule is not fully integrated. While the DEAMS PMO maintains internal schedules that reflect government-only activities, these government schedules have no links to activities within the contractor schedule. We found that activities in the contractor schedule are mapped to contract line item numbers and assigned to integrated product teams, but many activities are missing contractor work breakdown structure mappings. PMO officials told us that because of the firm-fixed price (FFP) nature of the current contract, the prime contractor is not obligated to provide detailed insight into the contractor schedule. Instead, the PMO uses the contractor schedule as a starting point to develop more detailed internal tools, such as lower-level schedule information maintained in spreadsheets. But without government activities fully integrated with contractor activities, we cannot guarantee that the schedule has either adequately captured all key activities necessary for the program's completion or that the PMO can reliably estimate the finish date for the program. Best practice: 2. Sequencing all activities; Explanation: The schedule should be planned so that critical project dates can be met. To meet this objective, activities need to be logically sequenced--that is, listed in the order in which they are to be carried out. In particular, activities that must be completed before other activities can begin (predecessor activities), as well as activities that cannot begin until other activities are completed (successor activities), should be identified. This helps ensure that interdependencies among activities that collectively lead to the accomplishment of events or milestones can be established and used as a basis for guiding work and measuring progress; Criterion met: Minimally; GAO analysis: Our analysis of the DEAMS contractor schedule shows that 131 of the 273 remaining activities, or 48 percent, have missing predecessor or successor logic. Missing predecessors or successors reduce the credibility of the calculated dates. If an activity that has no logical successor slips, the schedule will not reflect the effect on the critical path, float, or scheduled start dates of downstream activities. In addition, we found that 42 remaining activities, or 15 percent, have "dangling" logic--that is, these activities whose start or finish dates are missing logic. Of these 42 activities with dangling logic, 37 activities are missing logic that would determine their start dates. Because their start dates are not determined by logic, these activities would have to start earlier in order to finish on time if they ran longer than their planned durations. The other 5 activities with dangling logic are missing successors off their finish date. In other words, these activities could continue indefinitely and not affect the start or finish dates of future activities. We found six remaining activities with start-to- finish links. Start-to-finish links are rarely, if ever, used because they have the odd effect of causing a successor to finish before its predecessor.[A] We also found 18 links to or from summary tasks. Summary tasks should not have dependencies because they take their start date, finish date, and duration from lower-level activities. In addition, we found 50 remaining activities (18 percent) with Start No Earlier Than constraints. These are considered "soft" date constraints in that they allow the activity to slip into the future based on what happens to their predecessor activities. While activities may be soft constrained, for example, to represent receipt of delivery of equipment, in general constraining an activity's start date prevents managers from accomplishing work as soon as possible and consumes flexibility in the project. Of the remaining activities, 47 activities are linked to their successor activities with lags, including lags that are greater than 100 days. Lags represent the passing of time between activities but are often misused to put activities on a specific date or to insert a buffer for risk. Lags should be justified because they cannot vary with risk or uncertainty. PMO officials noted that the contractor schedule follows a Data Item Description (DID) that details the preparation of the schedule. However, the schedule does not meet the requirements set forth in the DID. For example, the DID states that the schedule "shall be an integrated, logical network-based schedule" and that a key element of the schedule is the "relationship/dependency" of an activity. Without logically sequencing activities, the schedule cannot be used as a reliable basis for guiding work and measuring progress. Best practice: 3. Assigning resources to all activities; Explanation: The schedule should reflect what resources (e.g., labor, materials, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist; Criterion met: Fully met; GAO analysis: Because of the current FFP contractual arrangement, the government does not have insight into the contractor's efforts to assign resources to activities. However, contractor officials provided evidence that resources have been assigned to activities within their schedule. In addition, PMO officials assign and monitor individual government resources to lower-level activities that are updated in internal tools outside the delivered contractor schedule. Best practice: 4. Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used. Durations should be as short as possible and have specific start and end dates. The schedule should be continually monitored to determine when forecasted completion dates differ from planned dates; this information can be used to determine whether schedule variances will affect subsequent work; Criterion met: Substantially; GAO analysis: The majority of remaining activities in the contractor schedule meet best practices for durations. There are 50 activities (18 percent) with planned durations longer than 44 days, which exceeds the best practice for activity duration.[B] There are 7 (3 percent) level-of-effort activities with durations greater than 1,200 days. These level-of-effort activities drive the end date of the project and hence adversely affect the calculation of the critical path--the longest duration path through the sequenced list of activities. Level- of-effort activities, such as systems engineering and program management, should not define the critical path because they are nondiscrete support activities that do not produce a definite end product. Best practice: 5. Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link products and outcomes associated with other sequenced activities. These links are commonly referred to as "handoffs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that the dates for starting and completing activities in the integrated master schedule should be aligned with the dates for supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Criterion met: Minimally; GAO analysis: Vertical integration--that is, the ability to consistently trace work breakdown structure elements between detailed, intermediate, and master schedules--is demonstrated somewhat because of the efforts by the DEAMS PMO to enhance its insight into contractor effort despite the FFP contract environment. PMO officials stated that while the contractor is under no obligation to provide detailed activities in the contractor schedule, the government has broken down areas such as object development and testing into detailed activities with internal tools that allow for weekly monitoring and status checking. However, we could not fully establish the link between the internal updating of activities by the government in lower-level spreadsheets and the high-level schedule delivered by the contractor. Issues with missing dependencies, activities with dangling logic, overuse of lags, and critical level-of-effort activities prevent the contractor schedule from fully complying with the requirement of horizontal integration--that is, the overall ability of the schedule to depict relationships between different program elements and product handoffs. PMO officials stated that rather than using the high-level contractor schedule, government and contractor subject matter experts meet each week to discuss progress on ongoing activities using other internal management tools. If activities are delayed or accelerated, the experts discuss potential impacts to downstream activities and provide management with weekly to daily information on these impacts. But while subject matter experts may understand the impacts of delayed activities, senior decision makers may not be aware of near-critical activities that have the potential to significantly delay the project, nor do they have the proper insight into available float--the amount of time an activity can slip before it delays the finish date of the project--that can be used to mitigate the risk of critical or near- critical activities. Best practice: 6. Establishing the critical path for all activities; Explanation: Scheduling software should be used to identify the critical path, which represents the chain of dependent activities with the longest total duration. Establishing a project's critical path is necessary to examine the effects of any activity slipping along this path. Potential problems along or near the critical path should also be identified and reflected in scheduling the duration of high-risk activities; Criterion met: Minimally; GAO analysis: Our analysis could not determine a valid critical path within the DEAMS contractor schedule, particularly because over 60 percent of remaining activities have missing or incomplete logic, and because level-of-effort activities (over 1,200 days long) define the start and finish dates of the project. Level-of-effort activities, such as systems engineering and program management, should not define the critical path because they are nondiscrete support activities that do not produce a definite end product. PMO officials acknowledged that a critical path cannot be calculated within the schedule and stated that the contractor schedule is used only as a starting point for more detailed internal tracking tools such as spreadsheets. Detail is not available within the contractor schedule because of the current FFP contract with the contractor. PMO officials also stated that establishing a traditional critical path is not possible in a complex ERP environment because there is no one clear path through development or testing. Rather than use the high-level contractor schedule government and contractor subject matter experts meet on a weekly to daily basis to discuss progress on ongoing activities using other internal management tools. If activities are delayed or accelerated, the experts discuss potential impacts to downstream activities and provide management with weekly to daily information on these impacts. But senior decision makers may not be aware of near-critical activities nor have the proper insight into available float that can be used to mitigate the risks associated with these activities. In addition, PMO officials noted that the contractor schedule follows a DID that details the preparation of the schedule. However, the contractor schedule does not meet the requirements set forth in the DID. The DID states a critical path is a key element of the detailed schedule; that "the critical path and near-critical paths are calculated by the scheduling software"; and "the critical path shall be easily identified." Best practice: 7. Identifying reasonable float; Explanation: The schedule should identify the float--the amount of time by which a predecessor activity can slip before the delay affects successor activities--so that a schedule's flexibility can be determined. As a general rule, activities along the critical path have the least float. Total float is the total amount of time by which an activity can be delayed without delaying the project's completion, if everything else goes according to plan; Criterion met: Minimally; GAO analysis: Our analysis found that float calculations within the DEAMS contractor schedule are not reliable because of the improper linking of summary tasks. In addition, because the schedule is missing dependencies, float estimates will be miscalculated because float is directly related to the logical sequencing of events. PMO officials told us that internal activity tracking and monitoring tools used in lieu of the detailed contractor activities do not allow insight into float calculations. PMO officials noted that the contractor schedule follows a DID that details the preparation of the schedule. However, the contractor schedule does not meet the requirements set forth in the DID. The DID states total float is a key element of the detailed schedule to be delivered monthly. Without float estimates management may be unable to allocate resources from non-critical activities to activities that cannot slip without affecting the project finish date. Best practice: 8. Conducting a schedule risk analysis; Explanation: A schedule risk analysis should be performed using statistical techniques to predict the level of confidence in meeting a project's completion date. This analysis focuses not only on critical path activities but also on activities near the critical path, since they can affect the project's status; Criterion met: Minimally; GAO analysis: The program office has not performed a schedule risk analysis on the schedule because the schedule is not used as a primary tool for monitoring the status of the program. However, program officials stated that they have tied risks to what subject matter experts consider to be critical path activities. They stated that they proactively monitor risk on a weekly basis by assigning a probability to the risk, examining the potential impact of the risk on activities if it is realized, and developing mitigation plans to be executed if the risk is realized. However, the risk assessments cannot be used to calculate the overall probability of finishing the project on time. Since any task can become critical if it is delayed long enough, complete schedule logic and a comprehensive risk assessment are essential tools for decision makers. A schedule risk analysis can be used to determine a level of confidence in meeting the completion date or whether proper reserves have been incorporated into the schedule. A schedule risk analysis will calculate schedule reserve, which can be set aside for those activities identified as high risk. Without this reserve, the program faces the risk of delays to the scheduled completion date if any delays were to occur on critical path activities. In addition, PMO officials noted that the contractor schedule follows a DID that details the preparation of the schedule. However, the contractor schedule does not meet the requirements set forth in the DID. The DID states that a key element of the detailed schedule is a schedule risk analysis that "predicts the probability of project completion by contractual dates" using three-point estimates about the remaining durations of remaining activities. Best practice: 9. Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should be continuously updated using logic and durations to determine realistic start and completion dates for program activities. The schedule should be analyzed continuously for variances to determine when forecasted completion dates differ from planned dates. This analysis is especially important for those variations that impact activities identified as being in a project's critical path and can impact a scheduled completion date; Criterion met: Minimally; GAO analysis: Our analysis shows the contractor schedule does not have a status date (or data date), nor did the program office expect one. A status date denotes the date of the latest update to the schedule and thus defines the point in time at which completed work and remaining work are calculated. Officials stated that the status date is reflected by the month in the schedule file name; but because no day is given there is no indication whether the date reflects the beginning or end of the calendar month or beginning or end of the contractor accounting period. Regardless of the exact date, we found 31 activities that had actual starts in future months relative to the month in the file name. That is, according to the schedule, these activities had actually started in the future. For example, the schedule file name is November 2009, yet we found actual start dates for activities in December 2009, February 2010, and April 2010. PMO officials noted that the contractor schedule follows the DID that details the preparation of the schedule. However, the contractor schedule does not meet the requirements set forth in the DID. The DID states that "actual start and actual finish dates, as recorded, shall not be later than the status date." PMO officials stated that rather than use the high-level contractor schedule, which does not give the required activity detail, government and contractor subject matter experts meet on a weekly to daily basis to discuss progress on ongoing activities using other internal management tools. For example, the Testing Integrated Product Team meets daily to review tasks that have been performed that day. If deadline criteria are not met, senior decision makers are alerted to potential impacts to the schedule. However, the schedule should use logic and durations in order to reflect realistic start and completion dates for program activities. The schedule should also be continually monitored to determine when forecasted completion dates differ from the planned dates, which can be used to determine whether schedule variances will affect downstream work. Maintaining the integrity of the schedule logic is not only necessary to reflect true status, but is also required before conducting a schedule risk analysis. Source: GAO analysis based on data provided by the DEAMS PMO. [A] Activities need to have certain predecessor-successor relationships so the schedule gives the correct results when they are updated or when durations change. Two logic requirements have to be provided: (1) a finish-to-start or start-to-start predecessors, so that if the activity is longer than scheduled it does not just start earlier automatically, and (2) finish-to-start or finish-to-finish successors that will be "pushed" if they take longer or finish later. [B] The Naval Air Systems Command recommends keeping individual task durations to less than two calendar months (44 working days). The shorter the duration of the tasks in the schedule, the more often the Control Account Managers are compelled to update completed work which more accurately reflects the actual status of the tasks. When task durations are too long, management insight into the actual status of the activity is reduced. [End of table] Table 12: Analysis of the Air Force's ECSS Solutions Development Project Schedule: Best practice: 1. Capturing all activities; Explanation: The schedule should reflect all activities as defined in the project's work breakdown structure, which defines in detail the work necessary to accomplish a project's objectives, including activities to be performed by both the owner and contractors; Criterion met: Substantially; GAO analysis: While the PMO does have detailed schedules of government effort--a commendable best practice--these are not fully integrated into an integrated master schedule (IMS) with the contractor schedules. Our analysis found that the ECSS Solutions Development schedule contains 215 detail activities associated with government effort, representing dependencies between contractor and government activities. However, the government activities are not completely linked to government schedules maintained and updated by the government PMO. Our analysis found that activities in the Solutions Development workstream schedule are mapped to contractor work breakdown structure elements and can be traced to completion criteria and descriptions of associated work products. Best practice: 2. Sequencing all activities; Explanation: The schedule should be planned so that critical project dates can be met. To meet this objective, activities need to be logically sequenced--that is, listed in the order in which they are to be carried out. In particular, activities that must be completed before other activities can begin (predecessor activities), as well as activities that cannot begin until other activities are completed (successor activities), should be identified. This helps ensure that interdependencies among activities that collectively lead to the accomplishment of events or milestones can be established and used as a basis for guiding work and measuring progress; Criterion met: Partially; GAO analysis: Our analysis shows that 31 of the 1,901 remaining activities, or 2 percent, have missing predecessor or successor logic. This is a relatively low number for such a highly integrated schedule, but any number of missing predecessors or successors can reduce the credibility of calculated dates. If an activity that has no logical successor slips, the schedule will not reflect the effect on the critical path, float, or scheduled start dates of future activities. However, of those remaining activities that have logical predecessor and successor links, 259 activities (14 percent), have "dangling logic." Of these 259 activities with dangling logic, 229 activities are missing logic that would determine their start dates. Because their start dates are not determined by logic, these activities would have to start earlier in order to finish on time if they ran longer than their planned durations. The other 30 activities with dangling logic are missing successors off their finish date. In other words, these activities could continue indefinitely and not affect the start or finish dates of downstream activities. The schedule includes four Must Finish On constraints. A Must Finish On constraint is considered a "hard" date constraint because it prevents the activity from finishing earlier or later than its planned date. This renders the schedule rigid and prevents the schedule from being dynamic. A Must Finish On constraint is artificial and makes the scheduled activity appear to be on track to finish on time when it may not be. There are also 17 Start No Earlier Than constraints within the schedule. These are considered "soft" constraints in that they allow the activity to slip into the future based on what happens to their predecessor activities. Activities may be soft constrained, for example, to represent receipt of delivery of equipment. However, in general constraining an activity's start date prevents managers from accomplishing work as soon as possible and consumes flexibility in the project. We found 78 start-to-finish links within the schedule. Start- to-finish links are rarely, if ever, used in scheduling practice because they have the odd effect of causing a successor activity to finish when its predecessor starts. Moreover, we found that each of the 78 start-to-finish links have 3-to 7-day lags. That is, the schedule logic dictates that these successors must finish a set number of days before their predecessors begin. Start-to-finish logic is at best confusing and at worst incorrect; activities should be rearranged to find true predecessors and successors and linked with straightforward logic. Of the 1,901 remaining activities, 467 activities are linked to their successor activities with lags. Lags represent the passing of time between activities but are often misused to put activities on a specific date or to insert a buffer for risk. Lags should be justified because they cannot have risk or uncertainty. Without logically sequencing activities, the schedule cannot be used as a reliable basis for guiding work and measuring progress. Best practice: 3. Assigning resources to all activities; Explanation: The schedule should reflect what resources (e.g., labor, materials, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist; Criterion met: Minimally; GAO analysis: The ECSS PMO stated that the government is aware that the contractor assigns resources to activities, but the government has no detailed insight into the resources because of the current FFP contractual arrangement. However, the program office was not able to provide evidence that would confirm that the schedule is resource loaded. Resource information would assist the program office in forecasting the likelihood of activities being completed based on their projected end dates. If the current schedule does not allow for insight into current or projected over-allocation of resources, then the risk of the program slipping is significantly increased. Best practice: 4. Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used. Durations should be as short as possible and have specific start and end dates. The schedule should be continually monitored to determine when forecasted completion dates differ from planned dates; this information can be used to determine whether schedule variances will affect subsequent work; Criterion met: Substantially; GAO analysis: Eighty-eight percent of remaining activities meet best practices for durations, being less than 44 days (or two working months). Seventy activities (4 percent) have longer than 100-day durations; the PMO has identified the majority of these as level-of- effort support activities. Twenty-five of these level-of-effort activities span the start and end dates of the project and appear in the schedule as critical activities. Level-of-effort activities, such as systems engineering and program management, cannot define the critical path because they are nondiscrete support activities that do not produce a definite end product. Best practice: 5. Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link products and outcomes associated with other sequenced activities. These links are commonly referred to as "handoffs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that the dates for starting and completing activities in the integrated master schedule should be aligned with the dates for supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Criterion met: Partially; GAO analysis: We found that vertical integration--that is, the ability to consistently trace work breakdown structure elements between detailed, intermediate, and master schedules--is demonstrated because the overall ECSS schedule is made up of individual project schedules like the Solutions Development schedule. However, issues with reliance on hard date constraints, the overuse of lags, critical level-of- effort tasks, and instances of convoluted logic such as start-to- finish links keep this detailed schedule from fully complying with the requirement of horizontal integration--that is, the overall ability of the schedule to depict relationships between different program elements and product handoffs. Horizontal integration demonstrates that the overall schedule is rational, planned in a logical sequence, accounts for interdependencies between work and planning packages, and provides a way to evaluate current status. Best practice: 6. Establishing the critical path for all activities; Explanation: Scheduling software should be used to identify the critical path, which represents the chain of dependent activities with the longest total duration. Establishing a project's critical path is necessary to examine the effects of any activity slipping along this path. Potential problems along or near the critical path should also be identified and reflected in scheduling the duration of high-risk activities; Criterion met: Partially; GAO analysis: Our analysis could not determine a valid critical path-- the longest duration path through the sequenced list of activities-- because level-of -effort activities define the start and finish dates of the detail planning portion of the project. Level of effort activities should not drive the critical path because they only serve to support detail work activities. The PMO acknowledged that a critical path would be difficult to calculate within the schedule because the project schedules are team-oriented rather than product- oriented, which causes complex linking relationships. While a true critical path does not exist throughout all 46 project schedules, program management reviews a high-level, manually constructed "Critical Events" schedule that tracks the status of major program milestones. These major program milestones are linked to lower-level schedules, and their status is updated daily and reviewed each week by program management. However, it is important that the lower-level schedules include complete logic that addresses the relationships between predecessor and successor activities, because any activity can become critical under some circumstances. Without clear insight into a critical path at the project level, management will not be able to monitor critical or near-critical detail activities that may have a detrimental impact on downstream activities if delayed. Best practice: 7. Identifying reasonable float; Explanation: The schedule should identify the float--the amount of time by which a predecessor activity can slip before the delay affects successor activities--so that a schedule's flexibility can be determined. As a general rule, activities along the critical path have the least float. Total float is the total amount of time by which an activity can be delayed without delaying the project's completion, if everything else goes according to plan; Criterion met: Partially; GAO analysis: Most remaining tasks appear to have reasonable total float, but there are 587 activities (31 percent) with over 50 days (2 working months) of total float. In other words, according to the schedule, 587 remaining activities (20 percent) could be delayed by 2 working months and not delay the final activity in the Solutions Development schedule. Activities with large float values may indicate a lack of completeness in the schedule logic. The PMO stated that total float is monitored by management in higher-level milestone schedules, not lower-level project schedules. Incorrect float estimates will result in an invalid critical path, and will result in an inability to allocate resources from non-critical activities to activities that cannot slip without affecting the project finish date. Best practice: 8. Conducting a schedule risk analysis; Explanation: A schedule risk analysis should be performed using statistical techniques to predict the level of confidence in meeting a project's completion date. This analysis focuses not only on critical path activities but also on activities near the critical path, since they can affect the project's status; Criterion met: Not met; GAO analysis: PMO officials stated that while the program reviews the schedule on a weekly basis and assesses risks to the program, it has not performed a schedule risk analysis. Best practices suggest that a schedule risk analysis can be used to determine a level of confidence in meeting the completion date or whether proper reserves have been incorporated into the schedule. Such an analysis will calculate schedule reserve, which can be set aside for those activities identified as high risk. Without this reserve, the program faces the risk of delays to the scheduled completion date if any delays were to occur in critical path activities. Best practice: 9. Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should be continuously updated using logic and durations to determine realistic start and completion dates for program activities. The schedule should be analyzed continuously for variances to determine when forecasted completion dates differ from planned dates. This analysis is especially important for those variations that impact activities identified as being in a project's critical path and can impact a scheduled completion date; Criterion met: Partially; GAO analysis: The status date for the version of the schedule we analyzed is January 1, 2010, a federal holiday. A status date denotes the date of the latest update to the schedule and thus defines the point in time at which completed work and remaining work are calculated. The PMO could not confirm that this date was correct. Assuming the status date is correct, we found several date anomalies within the schedule, suggesting that management may need to review how and when the schedule is updated. We found 29 activities (2 percent) that should have started but have no actual start date; 24 activities (1 percent) that should have finished but have no actual finish date; and 9 milestone activities with actual finish dates in the future. In addition, we found 24 instances (1 percent) of out-of- sequence logic--that is, actual progress being recorded on activities that, according to schedule logic, should not have begun yet. This is a common occurrence in scheduling, as actual events often override planned logic. However, some of these successor activities are planned to begin 2 to 3 months in the future, suggesting that the schedule logic should be updated to reflect changes. Source: GAO analysis based on data provided by the ECSS PMO. Note: The ECSS program schedule consists of a master schedule with 46 embedded project schedules representing individual product teams, or workstreams. The 46 schedules include 2 high-level schedules, one dedicated to key date milestones and another to critical events. Two project schedules were chosen based on their importance to the program and the high amount of activity currently associated with the product team. [End of table] Table 13: Analysis of the Air Force's ECSS Reports, Interfaces, Conversions, and Extensions (RICE) Program Schedule: Best practice: 1. Capturing all activities; Explanation: The schedule should reflect all activities as defined in the project's work breakdown structure, which defines in detail the work necessary to accomplish a project's objectives, including activities to be performed by both the owner and contractors; Criterion met: Substantially; GAO analysis: While the PMO does have detailed schedules of government effort--a commendable best practice--these are not fully integrated into an integrated master schedule (IMS) with the contractor schedules. Our analysis found that the ECSS RICE schedule contains "touch points," or links between government and contractor activities, representing dependencies between contractor and government activities. However, the government activities are not completely linked to government schedules maintained and updated by the government PMO. Our analysis found that activities in the RICE workstream schedule are mapped to contractor work breakdown structure elements and can be traced to completion criteria and descriptions of associated work products. Best practice: 2. Sequencing all activities; Explanation: The schedule should be planned so that critical project dates can be met. To meet this objective, activities need to be logically sequenced--that is, listed in the order in which they are to be carried out. In particular, activities that must be completed before other activities can begin (predecessor activities), as well as activities that cannot begin until other activities are completed (successor activities), should be identified. This helps ensure that interdependencies among activities that collectively lead to the accomplishment of events or milestones can be established and used as a basis for guiding work and measuring progress; Criterion met: Partially; GAO analysis: Our analysis shows that 472 of the 4,433 remaining activities, or 11 percent, have missing logic. Missing predecessors or successors are usually a signal of broken logic and reduce the credibility of the calculated dates. If an activity that has no logical successor slips, the schedule will not reflect the effect on the critical path, float, or scheduled start dates of future activities. In addition, we found 820 remaining activities, or 19 percent, have "dangling" logic. Of these 820 activities with dangling logic, 241 activities are missing logic that would determine their start dates. Because their start dates are not determined by logic, these activities would have to start earlier in order to finish on time if they ran longer than their planned durations. The other 579 activities with dangling logic are missing successors off their finish date. In other words, these activities could continue indefinitely and not affect the start or finish dates of future activities. We found 277 Start No Earlier Than constraints (6 percent) within the schedule. These are considered "soft" date constraints in that they allow the activity to slip into the future based on what happens to their predecessor activities. Activities may be soft constrained, for example, to represent receipt of delivery of equipment. However, in general constraining an activity's start date prevents managers from accomplishing work as soon as possible and consumes flexibility in the project. Of the remaining activities, 91 activities (2 percent) are linked to their successor activities with lags, including a lag greater than 100 days. Lags represent the passing of time between activities but are often misused to put activities on a specific date or to insert a buffer for risk. Lags should be justified because they cannot have risk or uncertainty. Without logically sequencing activities, the schedule cannot be used as a reliable basis for guiding work and measuring progress. Best practice: 3. Assigning resources to all activities; Explanation: The schedule should reflect what resources (e.g., labor, materials, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist; Criterion met: Minimally; GAO analysis: The ECSS PMO stated that the government is aware that the contractor assigns resources to activities, but the government has no detailed insight into the resources because of the current FFP contractual arrangement. However, the program office was not able to provide evidence that would confirm the schedule is resource loaded. Resource information would assist the program office in forecasting the likelihood of activities being completed based on their projected end dates. If the current schedule does not allow for insight into current or projected over-allocation of resources, then the risk of the program slipping is significantly increased. Best practice: 4. Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used. Durations should be as short as possible and have specific start and end dates. The schedule should be continually monitored to determine when forecasted completion dates differ from planned dates; this information can be used to determine whether schedule variances will affect subsequent work; Criterion met: Substantially; GAO analysis: Ninety-seven percent of the remaining activities meet best practices for durations, being less than 44 days (or two working months). Sixty activities (1 percent) have longer than 100-day durations, which the PMO has identified as level-of-effort support activities. Forty-two of these level-of-effort activities span the start and end dates of the project and appear in the schedule as critical activities. Level-of-effort activities, such as systems engineering and program management, cannot define the critical path because they are nondiscrete support activities that do not produce a definite end product. Best practice: 5. Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link products and outcomes associated with other sequenced activities. These links are commonly referred to as "handoffs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that the dates for starting and completing activities in the integrated master schedule should be aligned with the dates for supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Criterion met: Partially; GAO analysis: We found that vertical integration--that is, the ability to consistently trace work breakdown structure elements between detailed, intermediate, and master schedules--is demonstrated because the overall ECSS schedule is made up of individual project schedules like the RICE schedule. However, issues with missing dependencies, activities with dangling logic, overuse of lags, and critical level-of- effort activities keep this detailed schedule from being fully compliant with the requirement of horizontal integration--that is, the overall ability of the schedule to depict relationships between different program elements and product handoffs. Horizontal integration demonstrates that the overall schedule is rational, planned in a logical sequence, accounts for interdependencies between work and planning packages, and provides a way to evaluate current status. Best practice: 6. Establishing the critical path for all activities; Explanation: Scheduling software should be used to identify the critical path, which represents the chain of dependent activities with the longest total duration. Establishing a project's critical path is necessary to examine the effects of any activity slipping along this path. Potential problems along or near the critical path should also be identified and reflected in scheduling the duration of high-risk activities; Criterion met: Partially; GAO analysis: Our analysis could not determine a valid critical path-- the longest duration path through the sequenced list of activities-- because nearly 30 percent of remaining activities have missing or incomplete logic, and because level-of-effort tasks (209 days long) define the start and finish dates of the project. Level-of-effort activities should not drive the critical path because they only serve to support detail work activities. The government PMO acknowledged that a critical path would be difficult to calculate within the schedule because the project schedules are team-oriented rather than product-oriented, which causes complex linking relationships. While a true critical path does not exist throughout all 46 project schedules, program management reviews a high-level, manually constructed Critical Events schedule that tracks the status of major program milestones. These major program milestones are linked to lower-level schedules, and their status is updated daily and reviewed each week by program management. However, it is important that the lower level schedules include complete logic that addresses the relationships between predecessor and successor activities, because any activity can become critical under some circumstances. Without clear insight into a critical path at the project level, management will not be able to monitor critical or near-critical detail activities that may have a detrimental impact on downstream activities if delayed. Best practice: 7. Identifying reasonable float; Explanation: The schedule should identify the float--the amount of time by which a predecessor activity can slip before the delay affects successor activities--so that a schedule's flexibility can be determined. As a general rule, activities along the critical path have the least float. Total float is the total amount of time by which an activity can be delayed without delaying the project's completion, if everything else goes according to plan; Criterion met: Partially; GAO analysis: We found that the schedule did not have a reasonable amount of float because 78 percent of remaining activities have zero days of total float. In other words, according to the schedule, 3,448 remaining activities cannot slip one day without delaying the finish date of the project by one day. The program lead scheduler stated that total float is monitored by management at the higher critical events schedules, not lower-level project schedules. However, incorrect float estimates in lower-level schedules will result in an invalid critical path, and will result in an inability to allocate resources from non- critical activities to activities that cannot slip without affecting the project finish date. Best practice: 8. Conducting a schedule risk analysis; Explanation: A schedule risk analysis should be performed using statistical techniques to predict the level of confidence in meeting a project's completion date. This analysis focuses not only on critical path activities but also on activities near the critical path, since they can affect the project's status; Criterion met: Not met; GAO analysis: PMO officials stated that while the program reviews the schedule on a weekly basis and assesses risks to the program, it has not performed a schedule risk analysis. Best practices suggest that a schedule risk analysis can be used to determine a level of confidence in meeting the completion date or to determine whether proper reserves have been incorporated into the schedule. Such an analysis will calculate schedule reserve, which can be set aside for those activities identified as high risk. Without this reserve, the program faces the risk of delays to the scheduled completion date if any delays were to occur on critical path activities. Best practice: 9. Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should be continuously updated using logic and durations to determine realistic start and completion dates for program activities. The schedule should be analyzed continuously for variances to determine when forecasted completion dates differ from planned dates. This analysis is especially important for those variations that impact activities identified as being in a project's critical path and can impact a scheduled completion date; Criterion met: Partially; GAO analysis: The status date for the version of the schedule we analyzed is January 1, 2010, a federal holiday. A status date denotes the date of the latest update to the schedule and thus defines the point in time at which completed work and remaining work are calculated. The PMO could not confirm that this date was correct. Assuming the status date is correct, we found several date anomalies within the schedule, suggesting that management may need to review how and when the schedule is updated. For example, we found 14 activities (less than 1 percent) that should have started but have no actual start date; 17 activities (less than 1 percent) that should have finished but have no actual finish date; and 155 activities (3 percent) that occurred in the past according to the schedule but are missing both actual start dates and actual finish dates. In addition, we found 22 (less than 1 percent) instances of out-of-sequence logic-- that is, actual progress being recorded on activities that, according to schedule logic, should not have begun yet. This is a common occurrence in scheduling, as actual events often override planned logic. However, schedule logic should be updated to reflect changes as much as possible. Source: GAO analysis based on data provided by the ECSS PMO. [End of table] Table 14: Analysis of the Army's GFEBS Program Schedule: Best practice: 1. Capturing all activities; Explanation: The schedule should reflect all activities as defined in the project's work breakdown structure, which defines in detail the work necessary to accomplish a project's objectives, including activities to be performed by both the owner and contractors; Initial result: Substantially; Final result: Substantially; GAO analysis: Initial Analysis: Our analysis found that while the Wave 4 deployment schedule captures both contractor and government activities, the program schedule is not fully integrated because individual deployment schedules for software releases are not related to activities within other program schedules. PMO officials stated that the while release and maintenance activities are integrated together in one schedule, and each deployment wave has its own schedule, the schedules are not linked to each other because the activities within each schedule are not related. However, a fully integrated master schedule would link government and contractor development, deployment, and subsequent maintenance activities. Activities in the program schedule are mapped to the program's integrated master plan, and deliverables in the Wave 4 schedule are mapped to the program's Quality Assurance Surveillance Plan through unique identification numbers. A large portion of the Wave 4 deployment schedule is made up of receiver milestones; that is, products the program needs to receive from external field sites before certain activities can be conducted. In addition to including government and contractor activities, the schedule also include tasks representing work being performed by external organizations; Updated analysis: No change to initial assessment. Best practice: 2. Sequencing all activities; Explanation: The schedule should be planned so that critical project dates can be met. To meet this objective, activities need to be logically sequenced--that is, listed in the order in which they are to be carried out. In particular, activities that must be completed before other activities can begin (predecessor activities), as well as activities that cannot begin until other activities are completed (successor activities), should be identified. This helps ensure that interdependencies among activities that collectively lead to the accomplishment of events or milestones can be established and used as a basis for guiding work and measuring progress; Initial result: Minimally; Final result: Partially; GAO analysis: Initial analysis: Our analysis found 18 activities of 2,150 remaining (less than 1 percent) within the schedule that have no successor links, and three activities (less than 1 percent) that have neither successor nor predecessor links. Activities without successor links do not affect any other future activity. That is, they can continue until the end of the project without affecting the finish date of the project. The schedule includes 24 (1 percent) Must Start On (MSO) constraints. An MSO constraint is considered a "hard" date constraint because it prevents the activity from starting earlier or later than its planned date. This renders the schedule rigid and prevents the schedule from being dynamic. An MSO constraint is artificial and makes the scheduled activity appear to be on track to finish on time when it may not be. PMO schedulers told us that of the 24 MSO-constrained tasks, 15 (less than 1 percent of all remaining) are associated with executive briefings that are now out of scope and should be removed from the schedule. Of the remaining 9 MSO- constrained tasks, 8 are used to force successor activities to start on exactly the first days of calendar months. While these constraints may make scheduling activities simpler, they have an adverse effect on the project's critical path. An activity with an MSO constraint automatically becomes critical within scheduling software regardless of whether it actually should be critical. A final MSO constraint is attached to the "Go Live" milestone, which prevents the project finish milestone from shifting because of completed or remaining effort on predecessor activities. PMO officials acknowledged that the MSO constraint should not have been applied to the finish milestone and stated that it would be removed in the next update of the schedule. Our analysis also found that 50 summary tasks (12 percent of remaining summary tasks) have predecessor links. PMO schedulers told us that these summary links are used in lieu of linking predecessors to the numerous lower-level tasks. Because many of the lower-level tasks begin on the same date, this makes updating the schedule simpler: an updated start date for the summary task will force that same date on all the unlinked lower-level tasks. While this indeed makes updating easier, this technique is not considered a best practice. First, summary tasks do not represent work and are simply used as grouping elements. As such, they should take their start and finish dates from lower-level activities; they should not dictate the start or finish of lower-level activities. Secondly, linking summary tasks obfuscates the logic of the schedule. That is, tracing logic through summary links does not impart to management the sequence in which lower-level activities should be carried out. Our analysis found that 358 activities (17 percent) are scheduled to occur on a Sunday. This is a consequence of a summary task linked to a constrained milestone-- constrained to start on the first day of a calendar month, which happened to be a Sunday and in turn causes a multitude of lower-level activities to also begin on a Sunday. PMO schedulers acknowledged that this was an error and the activities would be shifted to begin on a work day. There are 67 remaining activities (3 percent) that are linked to their successor activities with lags and 38 (2 percent) are linked with negative lags (or "leads"). Lags represent the passing of time between activities but are often misused to put activities on a specific date or to insert a buffer for risk. Lags should be justified because they cannot have risk or uncertainty. Without logically sequenced activities, the schedule cannot be used as a reliable basis for guiding work and measuring progress; Updated analysis: There are still 18 tasks within the schedule without successors (or less than 1 percent of remaining activities in the updated schedule). While these activities have finish dates in December 2009 and February 2010, they do not have actual finish dates and we therefore cannot determine if these activities are completed or have the potential to cause future activities to slip. The updated schedule now contains 7 MSO constraints (less than 1 percent): the 15 constraints associated with executive meetings have been removed; the MSO constraint on the "Go Live" milestone has been removed; 2 MSO constraints marking the beginning of months have occurred; and 1 new constraint has been added to mark the beginning of the month following deployment. The 358 activities unintentionally scheduled to begin on a Sunday have been altered by a 1-day lag to begin on a proper workday. However, lags should not be used in lieu of logic to force activities to start on a specified date. Additionally, the updated schedule corrects minor missing predecessor logic issues. Best practice: 3. Assigning resources to all activities; Explanation: The schedule should reflect what resources (e.g., labor, materials, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist; Initial result: Not met; Final result: Not met; GAO analysis: Initial analysis: GFEBS officials stated that because of the current FFP contractual arrangement, the government does not have insight into the contractor's efforts to assign resources to activities. They stated that while they are aware that activities in previous schedule releases were assigned resources by the contractor, the current schedule is not resource loaded. Resource information would assist the program office in forecasting the likelihood of activities being completed based on their projected end dates. If the current schedule does not allow for insight into current or projected over-allocation of resources, then the risk of the program slipping is significantly increased; Updated analysis: No change to initial assessment. Best practice: 4. Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used. Durations should be as short as possible and have specific start and end dates. The schedule should be continually monitored to determine when forecasted completion dates differ from planned dates; this information can be used to determine whether schedule variances will affect subsequent work; Initial result: Fully met; Final result: Fully met; GAO analysis: Initial analysis: Seventy-two percent of remaining activities meet best practices for duration, being less than 44 days (or 2 working months). Activities with excessive durations (more than 100 days) represent effort being performed by organizations outside of the program office. Representing effort in the schedule that is performed by outside organizations is considered a best practice because it keeps management informed of ongoing work that might easily be forgotten until the deliverable is due, and the impact on future activities if the deliverable is behind schedule; Updated analysis: No change to initial assessment. Best practice: 5. Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link products and outcomes associated with other sequenced activities. These links are commonly referred to as "handoffs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that the dates for starting and completing activities in the integrated master schedule should be aligned with the dates for supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Initial result: Minimally; Final result: Minimally; GAO analysis: Initial analysis: The GFEBS program schedule includes detailed information on release, deployment, and maintenance government and contractor activities. However, our analysis of the schedule concludes that vertical integration--that is, the ability to consistently trace work breakdown structure elements between detailed, intermediate, and master schedules--is not fully demonstrated because none of the activities within the deployment schedules are related to activities associated with release, maintenance, or other wave schedules. PMO officials stated that while the release and maintenance activities are integrated together in one schedule, and each deployment wave has its own schedule, the schedules are not linked to each other because the activities within each schedule are not related. However, it is unlikely that deployment activities are unrelated to release or maintenance activities. Without vertically integrating the schedules, lower-level schedules cannot be clearly traced to upper-tiered milestones. Issues with reliance on hard date constraints, lags, and instances of convoluted logic such as linked summary tasks, keep the schedule from fully complying with the requirement of horizontal integration--that is, the overall ability of the schedule to clearly depict relationships between different program elements and product handoffs. Horizontal integration demonstrates that the overall schedule is rational, planned in a logical sequence, accounts for interdependencies between work and planning packages, and provides a way to evaluate current status; Updated analysis: No change to initial assessment. Best practice: 6. Establishing the critical path for all activities; Explanation: Scheduling software should be used to identify the critical path, which represents the chain of dependent activities with the longest total duration. Establishing a project's critical path is necessary to examine the effects of any activity slipping along this path. Potential problems along or near the critical path should also be identified and reflected in scheduling the duration of high-risk activities; Initial result: Minimally; Final result: Partially; GAO analysis: Initial analysis: Our analysis could not determine a valid critical path--the longest duration path through the sequenced list of activities--because the "Go Live" finish milestone is constrained with an MSO constraint. An MSO constraint is considered a "hard" date constraint because it prevents the activity from starting earlier or later than its planned date. This renders the schedule rigid and prevents the schedule from being dynamic. An MSO constraint is artificial and makes the scheduled activity appear to be on track to finish on time when it may not be. When the constraint is removed, the "Go Live" milestone slips two months from its constrained date of January 3, 2011 to March 7, 2011. In addition, our analysis found that without the MSO constraint, the nearest driving activity to the "Go Live" milestone (that is, the activity determining the date of the "Go Live" activity) is in October 2010. In other words, according to the schedule, no activity starting in November or December 2010 is critical to determining the "Go Live" date. Without clear insight into a critical path at the project level, management will not be able to monitor critical or near-critical detail activities that may have a detrimental impact on downstream activities if delayed; Updated analysis: The updated schedule has altered the predecessors on the "Go Live" finish milestone and the MSO constraint has been removed. The "Go Live" date is now scheduled to occur in February 2011. However, the critical path, as measured by the path with the lowest available float, shows only five activities from the "Go Live" date in February 2011 to the "Site Visit Activities Complete" milestone completed in March 2010. Because so few activities are on the current critical path, no activities scheduled within 1 or 2 months of deployment are currently driving the project finish date. In addition, the earliest critical activity on the path appears to be a functional survey scheduled for April 1, 2010, that has yet to actually start. Best practice: 7. Identifying reasonable float; Explanation: The schedule should identify the float--the amount of time by which a predecessor activity can slip before the delay affects successor activities--so that a schedule's flexibility can be determined. As a general rule, activities along the critical path have the least float. Total float is the total amount of time by which an activity can be delayed without delaying the project's completion, if everything else goes according to plan; Initial result: Minimally; Final result: Minimally; GAO analysis: Initial analysis: We found that the Wave 4 Deployment schedule displays unrealistic total float values. For example, 1,273 activities (59 percent) within the schedule are showing negative float. That is, these activities are one to 242 days behind schedule. Other tasks display an unrealistic amount of positive float: 49 tasks (59 percent) are showing 100 to more than 300 days of total float. In other words, according to the schedule, 49 remaining activities could be delayed by more than 4 working months and not delay the final activity in the Wave 4 schedule. As a general rule, activities along the critical path have the least amount of float. Activities with large float values may indicate some lack of completeness in the schedule logic. Incorrect float estimates will result in an invalid critical path and an inability to allocate resources from noncritical activities to activities that cannot slip without affecting the project finish date; Updated analysis: The updated schedule continues to reflect unrealistic float. For example, 172 remaining activities (8 percent) have from 90 to 252 days of negative float, while 25 remaining activities (1 percent) have 104 to 310 days of float. Best practice: 8. Conducting a schedule risk analysis; Explanation: A schedule risk analysis should be performed using statistical techniques to predict the level of confidence in meeting a project's completion date. This analysis focuses not only on critical path activities but also on activities near the critical path, since they can affect the project's status; Initial result: Not met; Final result: Not met; GAO analysis: Initial analysis: The PMO has not performed a schedule risk analysis. GFEBS officials stated that while schedule risks have been discussed in team meetings, the PMO has not performed a formal schedule risk analysis. However, officials stated that they are open to improving in the area of schedule risk analysis. Best practices suggest that a schedule risk analysis can be used to determine a level of confidence in meeting the completion date or to determine whether proper reserves have been incorporated into the schedule. Such an analysis will calculate schedule reserve, which can be set aside for those activities identified as high-risk. Without this reserve, the program faces the risk of delays to the scheduled completion date if any delays were to occur on critical path activities; Updated analysis: No change to initial assessment. Best practice: 9. Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should be continuously updated using logic and durations to determine realistic start and completion dates for program activities. The schedule should be analyzed continuously for variances to determine when forecasted completion dates differ from planned dates. This analysis is especially important for those variations that impact activities identified as being in a project's critical path and can impact a scheduled completion date; Initial result: Minimally; Final result: Partially; GAO analysis: Initial analysis: The status date for the version of the Wave 4 schedule we analyzed is May 3, 2010. A status date denotes the date of the latest update to the schedule and thus defines the point in time at which completed work and remaining work are calculated. As of this date, we found a relatively large number of date anomalies within the schedule, suggesting that management may need to review how and when the schedule is updated. For example, we found 247 activities (11 percent) that should have started but have no actual start date and 200 activities (9 percent) that should have finished but have no actual finish date. Moreover, we found 7 activities (less than 1 percent) that have actual finish dates in the future. Schedule logic should be updated to reflect actual progress so that management is aware of the latest plan and the impacts to the project if activity planned dates are not met; Updated analysis: The updated schedule no longer includes activities that have actual finish dates beyond the status date. However, the schedule contains 44 activities (2 percent) that should have started but have no actual start date; 22 activities (1 percent) that should have finished but have no actual finish date; and 109 activities (5 percent) that should have started and finished but have neither an actual start nor actual finish date. Source: GAO analysis based on data provided by the GFEBS PMO. Note: The initial analysis reflects our assessment of the schedule originally submitted by the GFEBS PMO for our review. In response to limitations that we identified and shared with the GFEBS PMO, the program office enacted several formal changes to their existing schedule. The updated analysis reflects our review of the revised schedule. [End of table] Table 15: Analysis of the Army's GCSS-Army Program Schedule: Best practice: 1. Capturing all activities; Explanation: The schedule should reflect all activities as defined in the project's work breakdown structure, which defines in detail the work necessary to accomplish a project's objectives, including activities to be performed by both the owner and contractors; Criterion met: Partially; GAO analysis: We found that the GCSS-Army program schedule is not fully integrated. While the program schedule contains detailed contractor activities, it only contains some major government milestones. Other government activities, such as testing events and future milestones beyond December 2010, are displayed in isolated, high-level illustrated documents rather than in dynamic scheduling documents. We also found that contractor activities within the program schedule are assigned contractor work package numbers and can be traced to individual control account plans and contractor work breakdown structure elements. Activities are also assigned integrated product teams and individual control account managers. However, without fully integrating government activities with contractor activities, DOD cannot guarantee the schedule has either adequately captured all key activities necessary for the program's completion or that it can reliably estimate the finish date for the program. Best practice: 2. Sequencing all activities; Explanation: The schedule should be planned so that critical project dates can be met. To meet this objective, activities need to be logically sequenced--that is, listed in the order in which they are to be carried out. In particular, activities that must be completed before other activities can begin (predecessor activities), as well as activities that cannot begin until other activities are completed (successor activities), should be identified. This helps ensure that interdependencies among activities that collectively lead to the accomplishment of events or milestones can be established and used as a basis for guiding work and measuring progress; Criterion met: Partially; GAO analysis: We found that only 2 of 2,255 activities (less than 1 percent) are missing dependencies and 19 activities (less than 1 percent) have "dangling" logic--that is, activities whose start or finish dates are missing logic. These activities with dangling logic have no successor from their finish date, meaning they can carry on indefinitely without affecting the start date of any other activity. While dependencies within the schedule are generally sound, 60 percent of the activities (1,360) have Start No Earlier Than constraints. Start No Earlier Than constraints are considered "soft" date constraints because they allow an activity to slip into the future if their predecessor activity is delayed, but the activity cannot begin earlier than its constraint date. Program officials stated that Start No Earlier Than constraints are used to manually allocate resources and to coordinate data tests, which rely on coordination with outside partners. Officials further stated that individual control account managers monitor these constraints. However, we found that 87 percent of the constraints were actively affecting the start date of their activities. That is, without the constraint, the activity may be able to start sooner. If these activities cannot start earlier, then their dates and dependencies should be updated to reflect reality. Constraining over half of all activities to start on or after specific dates defeats the purpose of a dynamic scheduling tool and greatly reduces to ability of the program to take advantage of possible time savings. We also found 143 Finish No Earlier Than constraints (6 percent). These are also considered "soft" date constraints because they prevent activities from finishing earlier than their constraint date. Program officials stated that these were erroneously created in the schedule during an internal file conversion process and would be removed in the next version of the program schedule. Without logically sequenced activities, the schedule cannot be used as a reliable basis for guiding work and measuring progress. Best practice: 3. Assigning resources to all activities; Explanation: The schedule should reflect what resources (e.g., labor, materials, and overhead) are needed to do the work, whether all required resources will be available when needed, and whether any funding or time constraints exist; Criterion met: Substantially; GAO analysis: While the integrated master schedule is not resource loaded, scheduled activities can be traced to control account plans which have resources laid out by month by labor category. Budgets are assigned at the control account level and resources are accounted for in monthly updates to the program's earned value management system. Best practice: 4. Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used. Durations should be as short as possible and have specific start and end dates. The schedule should be continually monitored to determine when forecasted completion dates differ from planned dates; this information can be used to determine whether schedule variances will affect subsequent work; Criterion met: Fully met; GAO analysis: Ninety-eight percent of remaining activities meet the best practice for activity duration, being less than 44 days. Only two remaining activities have durations that exceed the best practice, extending beyond 80 days. Best practice: 5. Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link products and outcomes associated with other sequenced activities. These links are commonly referred to as "handoffs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that the dates for starting and completing activities in the integrated master schedule should be aligned with the dates for supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Criterion met: Partially; GAO analysis: The schedule is vertically integrated, with low-level tasks and milestones being traceable to higher-level summary tasks. While the schedule has a relatively small number of missing dependencies and activities with dangling logic, the use of date constraints on more than 60 percent of remaining activities, prevent the schedule from being completely horizontally integrated. That is, the date constraints limit the overall ability of the schedule to depict dynamic relationships between different program elements and product handoffs. Horizontal integration demonstrates that the overall schedule is rational, planned in a logical sequence, accounts for interdependencies between work and planning packages, and provides a way to evaluate current status. Best practice: 6. Establishing the critical path for all activities; Explanation: Scheduling software should be used to identify the critical path, which represents the chain of dependent activities with the longest total duration. Establishing a project's critical path is necessary to examine the effects of any activity slipping along this path. Potential problems along or near the critical path should also be identified and reflected in scheduling the duration of high-risk activities; Criterion met: Partially; GAO analysis: We found that a reliable and realistic critical path could not be determined within the program schedule, and program officials agreed with our assessment. Program officials stated that the schedule is constructed to increase the visibility of each software object's development, and as a consequence of this amount of detail, a critical path cannot be shown. The schedule displays the detailed development life cycle for hundreds of objects and depending on the order in which objects are completed, the dependencies between objects, and the constant reallocation of resources, a traditional critical path may be too volatile to be useful. However, officials stated that a higher-level summary type schedule, which would display a valid critical path, would not allow management the proper insight into the risks underlying the development of each object. In lieu of a traditional critical path, program management monitors object development weekly and program officials stated that they are fully aware of which activities are behind or ahead of schedule. It is commendable that the schedule includes the necessary amount of complexity and detail to track lower-level, high-risk development activities. However, our analysis found that a critical path could not be derived because of artificial date constraints rather than complex object development detail. Program officials stated that three major milestones are being tracked: Critical Design Review, DTOE 1.1, and Build/Design Phase Completion. We found that critical paths do not exist for any of these milestones because of artificial date constraints on activities unrelated to detailed object development. As a result, we cannot determine a critical path to any major milestone based on actual effort related to object development. In this respect, the schedule cannot reliably forecast completion dates for Critical Design Review, DTOE, Build/Design Completion, or, as a consequence, Milestone C. Best practice: 7. Identifying reasonable float; Explanation: The schedule should identify the float--the amount of time by which a predecessor activity can slip before the delay affects successor activities--so that a schedule's flexibility can be determined. As a general rule, activities along the critical path have the least float. Total float is the total amount of time by which an activity can be delayed without delaying the project's completion, if everything else goes according to plan; Criterion met: Substantially; GAO analysis: The majority of remaining tasks in the GCSS-Army contractor schedule appear to have reasonable total float values, varying from 0 to 30 days. Program office officials stated that they believe the schedule portrays accurate float. However, our analysis found 338 (15 percent) remaining activities with over 100 days of total float. In other words, according to the schedule, 338 remaining activities could be delayed by 4 months and not delay the final project date. Activities with large float values may indicate a lack of completeness in the schedule logic. Best practice: 8. Conducting a schedule risk analysis; Explanation: A schedule risk analysis should be performed using statistical techniques to predict the level of confidence in meeting a project's completion date. This analysis focuses not only on critical path activities but also on activities near the critical path, since they can affect the project's status; Criterion met: Minimally; GAO analysis: Program office officials stated that a schedule risk analysis is not routinely performed, and that there is currently no requirement for the contractor to do so. While no detailed risk analysis has been performed on the schedule, the contractor recently conducted a high-level Monte Carlo risk analysis on two major milestones for an integrated master schedule management review meeting. This high-level risk analysis shows the probability of completing the key milestones on time and identifies mitigating actions to prevent delays. Program officials stated they are interested in periodic risk analysis and intend to include a schedule risk analysis requirement in the contract within the next few months. A schedule risk analysis can be used to determine a level of confidence in meeting the completion date or whether proper reserves have been incorporated into the schedule. Such an analysis will calculate schedule reserve, which can be set aside for those activities identified as high risk. Without this reserve, the program faces the risk of delays to the scheduled completion date if any delays were to occur on critical path activities. Best practice: 9. Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should be continuously updated using logic and durations to determine realistic start and completion dates for program activities. The schedule should be analyzed continuously for variances to determine when forecasted completion dates differ from planned dates. This analysis is especially important for those variations that impact activities identified as being in a project's critical path and can impact a scheduled completion date; Criterion met: Substantially; GAO analysis: We found no instances of illogical dates, such as actual start or actual finish dates in the future. We found 112 instances (5 percent) of out-of-sequence logic; that is, actual progress recorded on activities that, according to schedule logic, should not have started yet. This is a common occurrence in scheduling, as actual events often override planned logic. However, a large number of out-of- sequence activities may indicate that the schedule is not being thoroughly updated to reflect reality on a periodic basis. Source: GAO analysis based on data provided by the GCSS-Army PMO. [End of table] [End of section] Appendix V: Assessments of Four DOD ERP Program Cost Estimates: This appendix provides the results of our analysis of the extent to which the processes and methodologies used to develop and maintain the four ERP cost estimates meet the characteristics of high-quality cost estimates.[Footnote 72] The four characteristics of high-quality estimates are explained and mapped to the 12 steps of such estimates in table 16. Table 16: The 12 Steps of High-Quality Cost Estimating, Mapped to the Steps of a High-Quality Cost Estimate: Characteristic: Well-documented; Explanation: The documentation should address the purpose of the estimate, the program background and system description, its schedule, the scope of the estimate (in terms of time and what is and is not included), the ground rules and assumptions, all data sources, estimating methodology and rationale, the results of the risk analysis, and a conclusion about whether the cost estimate is reasonable. Therefore, a good cost estimate--while taking the form of a single number--is supported by detailed documentation that describes how it was derived and how the expected funding will be spent in order to achieve a given objective. For example, the documentation should capture in writing such things as the source data used and their significance, the calculations performed and their results, and the rationale for choosing a particular estimating method or reference. Moreover, this information should be captured in such a way that the data used to derive the estimate can be traced back to and verified against their sources, allowing for the estimate to be easily replicated and updated. Finally, the cost estimate should be reviewed and accepted by management to ensure that there is a high level of confidence in the estimating process and the estimate itself; Step: Step 1: Define the estimate's purpose, scope, and schedule; Step 3: Define the program characteristics; Step 5: Identify ground rules and assumptions; Step 6: Obtain the data; Step 10: Document the estimate; Step 11: Present the estimate to management for approval. Characteristic: Comprehensive; Explanation: The cost estimates should include both government and contractor costs of the program over its full life cycle, from inception of the program through design, development, deployment, and operation and maintenance to retirement of the program. They should also completely define the program, reflect the current schedule, and be technically reasonable. Comprehensive cost estimates should provide a level of detail appropriate to ensure that cost elements are neither omitted nor double counted, and they should document all cost- influencing ground rules and assumptions. Establishing a product- oriented work breakdown structure (WBS) is a best practice because it allows a program to track cost and schedule by defined deliverables, such as a hardware or software component; Step: Step 2: Develop the estimating plan; Step 4: Determine the estimating structure; Step 5: Identify ground rules and assumptions[A]. Characteristic: Accurate; Explanation: The cost estimates should provide for results that are unbiased, and they should not be overly conservative or optimistic. Estimates are accurate when they are based on an assessment of most likely costs, adjusted properly for inflation, and contain few, if any, minor mistakes. In addition, the estimates should be updated regularly to reflect material changes in the program, such as when schedules or other assumptions change, and actual costs so that the estimate is always reflecting current status. Among other things, the estimate should be grounded in documented assumptions and a historical record of cost estimating and actual experiences on other comparable programs; Step: Step 7: Develop the point estimate[B]; Step 12: Update the estimate to reflect actual costs and changes. Characteristic: Credible; Explanation: The cost estimates should discuss any limitations of the analysis because of uncertainty or biases surrounding data or assumptions. Major assumptions should be varied, and other outcomes recomputed to determine how sensitive they are to changes in the assumptions. Risk and uncertainty analysis should be performed to determine the level of risk associated with the estimate. Further, the estimate's results should be crosschecked, and an independent cost estimate conducted by a group outside the acquiring organization should be developed to determine whether other estimating methods produce similar results. For management to make good decisions, the program estimate must reflect the degree of uncertainty, so that a level of confidence can be given about the estimate. Having a range of costs around a point estimate is more useful to decision makers because it conveys the level of confidence in achieving the most likely cost and also informs them on cost, schedule, and technical risks; Step: Step 7: Compare the point estimate to an independent cost estimate[C]; Step 8: Conduct sensitivity analysis; Step 9: Conduct risk and uncertainty analysis. Source: GAO-09-3SP. [A] This step applies to two of the characteristics--well-documented and comprehensive. [B] A point estimate is a single cost estimate number representing the most likely cost. [C] This step applies to two of the characteristics--credible and accurate. [End of table] Tables 17, 18, 19, and 20 provide the detailed results of our analysis of the program cost estimates for DEAMS, ECSS, GFEBS and GCSS-Army. "Not met" means the program provided no evidence that satisfies any of the criterion. "Minimally" means the program provided evidence that satisfies a small portion of the criterion. "Partially" means the program provided evidence that satisfies about half of the criterion. "Substantially" means the program provided evidence that satisfies a large portion of the criterion. "Fully met" means the program provided evidence that completely satisfies the criterion. Table 17: Analysis of the Air Force's DEAMS Cost Estimate: Four characteristics of high-quality cost estimates: Well-documented; Criterion met: Substantially; Key examples of rationale for assessment: The purpose, scope and schedule of the estimate were clearly defined. Further, the estimate identified all the ground rules and assumptions as well as the estimating methodology. The PMO presented evidence of receiving approval of the estimate through briefings to management. The sources of data the estimate was based on were also documented. However, we found inconsistencies when comparing the program requirements found in the Cost Analysis Requirements Document (CARD) with the requirements contained in the cost estimate. For example, commercial-off-the-shelf (COTS) software licenses requirements outlined in the CARD do not match the assumptions used in the cost estimate. Four characteristics of high-quality cost estimates: Comprehensive; Criterion met: Fully met; Key examples of rationale for assessment: The program provided supporting documentation that showed the ground rules and assumptions. The estimate is based on a cost element structure as stated in the Department of Defense Automated Information System Economic Analysis Guide. The program also provided an estimating plan that included the cost estimating schedule. Four characteristics of high-quality cost estimates: Accurate; Criterion met: Fully met; Key examples of rationale for assessment: The DEAMS cost model details the calculations and inflation indexes underlying the estimated costs. Calculations within the model can be traced back to supporting documentation. In addition, the cost model is updated annually to incorporate actual costs expended in prior fiscal years. Four characteristics of high-quality cost estimates: Credible; Criterion met: Fully met; Key examples of rationale for assessment: An independent cost estimate was developed by the Air Force Cost Analysis Agency. The PMO and Air Force Cost Analysis Agency also conducted analyses to identify the cost elements with the greatest degree of uncertainty, and determine the cost drivers for the program, and performed analyses to determine the impact of changing major ground rules and assumptions. For example, during the reconciliation process, there was debate as to the best estimate for the total number of DEAMS users. The PMO performed a sensitivity analysis on this parameter to illustrate the total life cycle cost impact of changing this assumption. The PMO submitted several supporting documents that detail the risk and uncertainty analysis performed on the cost estimates. In addition to the risk and uncertainty analysis, the PMO implemented a risk management process at the inception of the program and is planned to continue throughout the program's life. Source: GAO analysis based on data provided by the DEAMS PMO. [End of table] Table 18: Analysis of the Air Force's ECSS Cost Estimate: Four characteristics of high-quality cost estimates: Well-documented; Criterion met: Substantially; Key examples of rationale for assessment: The purpose, scope and schedule of the estimate were clearly defined. The PMO presented evidence of receiving approval of the estimate through briefings to management. The data sources were also documented. The PMO also provided ample descriptions of the methodology used to derive the estimates. However, our analysis found inconsistencies between requirements found in the CARD and assumptions used to calculate the estimate. For example, personnel requirements and the number of reports, interfaces, conversions, and extensions were different between the two documents. Four characteristics of high-quality cost estimates: Comprehensive; Criterion met: Fully met; Key examples of rationale for assessment: The program provided supporting documentation that showed the ground rules and assumptions. The estimate is prepared in accordance with the Office of the Secretary of Defense ERP work breakdown structure as stated in draft DOD guidance.[A] The program also provided an estimating plan that included the cost estimating schedule. Four characteristics of high-quality cost estimates: Accurate; Criterion met: Substantially; Key examples of rationale for assessment: The ECSS cost model details the calculations and inflation indexes underlying the estimated costs. Calculations within the model can be traced back to supporting documentation. However, our analysis found minor inconsistencies when cross-checking costs that were presented to management and the underlying calculations within the model. For example, estimates for data migration, data cleansing, and help desk within the cost model do not match the cost estimates presented to management. ECSS PMO officials stated they cannot compare actual costs to the cost estimate because they do not yet have an approved baseline. However, these officials stated the program has a Baseline Change Board that holds monthly Resource Board meetings during which officials review the program baseline to assess the potential impacts of proposed changes to all aspects of the program's life cycle. Four characteristics of high-quality cost estimates: Credible; Criterion met: Partially; Key examples of rationale for assessment: An independent cost estimate was created by the Air Force Cost Analysis Agency. ECSS PMO officials stated that the cost estimate was adjusted based on sensitivity analyses. However, the cost estimate model does not include evidence of a sensitivity analysis. Because the Air Force did not conduct a sensitivity analysis to identify the effects of uncertainties associated with different assumptions, there is an increased risk that decisions will be made without a clear understanding of the possible impact on cost and benefit estimates. The ECSS PMO performed a cost risk and uncertainty analysis. This analysis shows that the service cost position is at the 60 percent confidence level-meaning there is a 40 percent chance of a cost overrun. In addition to the risk and uncertainty analysis, the PMO has implemented a risk management process to identify and mitigate schedule, cost, and performance risks. Source: GAO analysis based on data provided by the ECSS PMO. [A] MIL-HDBK-881. [End of table] Table 19: Analysis of the Army's GFEBS Cost Estimate: Four characteristics of high-quality cost estimates: Well-documented; Criterion met: Fully met; Key examples of rationale for assessment: The purpose, scope and schedule of the estimate were clearly defined. Further, the documentation identified all the ground rules and assumptions as well as the estimating methodology. The PMO presented evidence of receiving approval of the estimate through briefings to management. The sources of data the estimate was based on were also documented. Four characteristics of high-quality cost estimates: Comprehensive; Criterion met: Fully met; Key examples of rationale for assessment: The program provided supporting documentation that showed the ground rules and assumptions underlying the cost estimate. The estimate is based on a cost estimating structure as dictated by the Department of the Army Economic Analysis Manual. The program also provided an estimating plan that included the cost estimating schedule. Four characteristics of high-quality cost estimates: Accurate; Criterion met: Substantially; Key examples of rationale for assessment: The GFEBS cost estimate details the calculations and inflation indexes underlying the estimated costs. Calculations within the model can be traced back to supporting documentation. In addition, evidence was provided that shows how estimated costs were derived based on actual costs incurred to date. For example, the estimated cost for program management is based on actual historical program management costs. However, because a cost uncertainty analysis has not been performed, DOD cannot guarantee that the estimate represents most likely costs to be incurred. Four characteristics of high-quality cost estimates: Credible; Criterion met: Minimally; Key examples of rationale for assessment: An independent cost estimate was created by the Office of the Deputy Assistant Secretary of the Army for Cost and Economics. However, the estimate does not include either a sensitivity or risk and uncertainty analysis. The GFEBS PMO stated that it has adequately accounted for risks in the cost estimate based on the maturity of the program and the reconciliation process between the PMO estimate and the independent cost estimate. However, because the Army did not conduct a sensitivity analysis to identify the effect of uncertainties associated with different assumptions, there is an increased risk that decisions will be made without a clear understanding of the possible impact on cost and benefit estimates. Source: GAO analysis based on data provided by the GFEBS PMO. [End of table] Table 20: Analysis of the Army's GCSS-Army Cost Estimate: Four characteristics of high-quality cost estimates: Well-documented; Criterion met: Substantially; Key examples of rationale for assessment: The purpose, scope and schedule of the cost estimate were clearly defined. The program has a current technical baseline document, and the PMO presented evidence of receiving approval of the estimate through briefings to management. However, the Economic Analysis documentation describing the cost estimate presents costs at a high level but does not provide details on lower level cost elements. Four characteristics of high-quality cost estimates: Comprehensive; Criterion met: Substantially; Key examples of rationale for assessment: The GCSS-Army PMO uses a "hybrid" work breakdown structure for the program based on its collaboration with the Office of the Secretary of Defense Cost and Resource Center. This hybrid work breakdown structure while not entirely product-oriented, standardizes the vocabulary for cost elements for automated information systems. Because there is currently no standardized work breakdown structure in use by DOD that corresponds to the implementation of an ERP system, the PMO worked closely with the Office of the Secretary of Defense Cost and Resource Center to develop a mutually acceptable work breakdown structure that meets best practices. In addition, the program provided supporting documentation that showed the ground rules and assumptions used to generate the cost estimate. However, our analysis shows that not all ground rules and assumptions were used to develop the cost risk and uncertainty analysis. For example, there are several assumptions associated with the number of software licenses, yet the risk and uncertainty analysis does not reflect any risk associated with these assumptions. Four characteristics of high-quality cost estimates: Accurate; Criterion met: Partially; Key examples of rationale for assessment: The cost estimate model shows the methodology and calculations used to prepare the estimate. However, because the PMO did not provide supporting documentation that details the use of actual costs to derive cost estimates, we are unable to verify the quality of the cost estimates. Programs should be monitored continuously for their cost-effectiveness by comparing planned and actual performance against the approved program baseline. The estimates should be updated with actual costs so that it is always relevant and current. This results in a higher-quality cost estimate and provides an opportunity to incorporate lessons learned. Four characteristics of high-quality cost estimates: Credible; Criterion met: Partially; Key examples of rationale for assessment: An independent cost estimate was created by the Army Cost Review Board Working Group. However, the cost estimate does not include a sensitivity analysis. Because the Army did not conduct a sensitivity analysis to identify the effects of uncertainties associated with different assumptions, there is an increased risk that decisions will be made without a clear understanding of the possible impact on cost and benefit estimates. The supporting documentation shows risk-adjusted costs, which were generated by applying probability distributions to cost elements within the cost model. However, the probability distributions applied throughout the model to account for risks are generalized and do not make a distinction in how specific risks may affect specific cost elements differently. While the GCSS-Army PMO has a risk process to identify, analyze, plan, track, control, and communicate risks, our analysis found that the PMO did not adequately link risks to the cost estimate. For example, data cleansing and data migration are noted as high-risks within the risk register but they are not accounted for in the risk and uncertainty analysis. Without a realistic risk and uncertainty analysis, the PMO can neither quantify the level of confidence in achieving a program within a certain funding level, nor determine a defensible amount of contingency reserve to quickly mitigate risk. Source: GAO analysis based on data provided by the GCSS-Army PMO. Note: The focus of our GCSS-Army cost assessment is the "ratified" GCSS-Army Cost Position dated November 2006 because the ratified Army Cost Position represents a more detailed approach to the program's cost estimating process compared to the current "federated" approach estimate for Milestone B. The current federated estimate, which reflects the federated ERP integration strategy for GCSS-Army and General Funds Enterprise Business Systems (GFEBS), was developed within 40 days as mandated by the Department of the Army. The PMO plans to implement a more detailed cost estimating process for their federated Army Cost Position in preparation for Milestone C in February 2011. [End of table] [End of section] Appendix VI: GAO Contacts and Staff Acknowledgments: GAO Contacts: Asif A. Khan, (202) 512-9095 or khana@gao.gov: Nabajyoti Barkakati, (202) 512-4499 or barkakatin@gao.gov: Staff Acknowledgments: In addition to the contacts named above, the following individuals made key contributions to this report: J. Christopher Martin, Senior- Level Technologist; Darby Smith, Assistant Director; Evelyn Logue, Assistant Director; Karen Richey, Assistant Director; F. Abe Dymond, Assistant General Counsel; Beatrice Alff; Tyler Benson; Michael Bird; Jennifer Echard; Maxine Hattery; Jason Kelly; Jason Kirwan; Crystal Lazcano; Jason Lee; Len Ogborn; and Vanessa Virtudazo. [End of section] Footnotes: [1] DOD's business systems are information systems, including financial and nonfinancial systems that support DOD business operations, such as civilian personnel, finance, health, logistics, military personnel, procurement, and transportation. [2] GAO, High-Risk Series: An Update, [hyperlink, http://www.gao.gov/products/GAO-09-271] (Washington, D.C.: January 2009). [3] An ERP solution is an automated system using commercial off-the- shelf (COTS) software consisting of multiple, integrated functional modules that perform a variety of business-related tasks such as general ledger accounting, payroll, and supply chain management. [4] These areas were designated as high risk in 2005, 1995, and 1990, respectively. [5] The 10 ERPs are as follows: Army--General Fund Enterprise Business System (GFEBS), Global Combat Support System-Army (GCSS-Army), and Logistics Modernization Program (LMP); Navy--Navy Enterprise Resource Planning (Navy ERP) and Global Combat Support System-Marine Corps (GCSS-MC); Air Force--Defense Enterprise Accounting and Management System (DEAMS) and Expeditionary Combat Support System (ECSS); Defense--Service Specific Integrated Personnel and Pay Systems and Defense Agencies Initiative (DAI); and Defense Logistics Agency-- Business System Modernization (BSM). According to DOD, BSM was fully implemented in July 2007. [6] GAO, Defense Logistics: Actions Needed to Improve Implementation of the Army Logistics Modernization Program, [hyperlink, http://www.gao.gov/products/GAO-10-461] (Washington, D.C.: Apr. 30, 2010); DOD Business Systems Modernization: Navy Implementing a Number of Key Management Controls on Enterprise Resource Planning System, but Improvements Still Needed, [hyperlink, http://www.gao.gov/products/GAO-09-841] (Washington, D.C.: Sept. 15, 2009); DOD Business Systems Modernization: Important Management Controls Being Implemented on Major Navy Program, but Improvements Needed in Key Areas, [hyperlink, http://www.gao.gov/products/GAO-08-896] (Washington, D.C.: Sept. 8, 2008); DOD Business Transformation: Air Force's Current Approach Increases Risk That Asset Visibility Goals and Transformation Priorities Will Not Be Achieved, [hyperlink, http://www.gao.gov/products/GAO-08-866] (Washington, D.C.: Aug. 8, 2008); DOD Business Systems Modernization: Key Marine Corps System Acquisition Needs to Be Better Justified, Defined, and Managed, [hyperlink, http://www.gao.gov/products/GAO-08-822] (Washington, D.C.: July 28, 2008); and DOD Business Transformation: Lack of an Integrated Strategy Puts the Army's Asset Visibility System Investments at Risk, [hyperlink, http://www.gao.gov/products/GAO-07-860] (Washington, D.C.: July 27, 2007). [7] [hyperlink, http://www.gao.gov/products/GAO-10-461], [hyperlink, http://www.gao.gov/products/GAO-09-841], [hyperlink, http://www.gao.gov/products/GAO-08-896], [hyperlink, http://www.gao.gov/products/GAO-08-866], [hyperlink, http://www.gao.gov/products/GAO-08-822], and [hyperlink, http://www.gao.gov/products/GAO-07-860]. [8] We reviewed the Army's GFEBS and GCSS-Army and the Air Force's DEAMS and ECSS. [9] GAO, GAO Cost Estimating and Assessment Guide Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington, D.C.: March 2009). [10] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [11] [hyperlink, http://www.gao.gov/products/GAO-08-822] and [hyperlink, http://www.gao.gov/products/GAO-08-896]. [12] The reported amounts are not audited. In November 2009, the DOD Inspector General reported that because of long-standing internal control weaknesses, DOD's annual financial statements, which included these reported amounts, were not accurate and reliable. [13] DOD excludes from its business systems those designated as national security systems under Section 2222 (j) of Title 10, United States Code. National security systems are intelligence systems, cryptologic activities related to national security, military command and control systems, and equipment that is an integral part of a weapon or weapons system or is critical to the direct fulfillment of military or intelligence missions. [14] DOD Directive 5000.01, The Defense Acquisition System (Nov. 20, 2007). [15] There are five tiers of business systems. Tier 1 systems include all large, expensive system programs classified as a major automated information system (MAIS) or a major defense acquisition program (MDAP) and subject to the most extensive statutory and regulatory reporting requirements. Tier 2 systems include those with modernization efforts of $10 million or greater but that are not designated as MAIS or MDAP or programs that have been designated as investment review board programs of interest because of their effect on DOD transformation objectives. Tier 3 systems include those with modernization efforts that have anticipated costs greater than $1 million but less than $10 million. Tier 4 includes systems with development/modernization cost of $1 million or less. Tier 5 includes systems in operation and maintenance or sustainment. [16] The five IRBs are (1) financial management established by the Under Secretary of Defense (Comptroller); (2) weapon systems life- cycle management and materiel supply and services management established by the Under Secretary of Defense (Acquisition, Technology and Logistics); (3) real property and installations life-cycle management established by the Under Secretary of Defense (Acquisition, Technology and Logistics); (4) human resources management established by the Under Secretary of Defense for Personnel and Readiness; and (5) Department of Defense Chief Information Officer established by the Assistant Secretary of Defense (Networks and Information Integration)/DOD Chief Information Officer. [17] Ronald W. Reagan National Defense Authorization Act for Fiscal Year 2005, Pub. L. No. 108-375, § 332, 118 Stat. 1811, 1851-1856 (Oct. 28, 2004), codified in part at 10 U.S.C. § 2222, directs that DOD may not obligate appropriated funds for a defense business system modernization with a total cost of more than $1 million unless, the approval authority--that is the appropriate IRB--certifies that the business system modernization either (1) complies with the department's business enterprise architecture, (2) is necessary to achieve a critical national security capability or address a critical requirement in an area such as safety or security, or (3) is necessary to prevent a significant adverse effect on an essential project in consideration of alternative solutions. This certification must also be approved by the DBSMC. Also, as of October 28, 2009, the fiscal year 2010 National Defense Authorization Act, Pub. L. No. 111-84, §1072, 123 Stat. 2190, 2470 (Oct. 28, 2009), amended this requirement. This amendment requires the chief management officer of the military services, or for defense agencies, the DOD DCMO, to assess whether (1) the business process that the system supports will be as streamlined and efficient as possible and (2) the need to tailor commercial-off- the-shelf systems to meet unique requirements or incorporate unique interfaces has been eliminated or reduced to the maximum extent practicable. This assessment is required both as a precondition of the approval of any new business system modernization with a cost over $1 million, and as a review of any previously approved business system modernization with a cost over $100 million. [18] Pub. L. No. 110-417, div. A, title IX; §908, 122 Stat. 4356, 4569 (Oct. 14, 2008). [19] The general fund can be defined as the fund into which receipts are deposited, except those from specific sources required by law to be deposited into other designated funds and from which appropriations are made by Congress to carry on the general and ordinary operations of the government. [20] According to the GFEBS PMO, once the system is fully operational the Army will assess the feasibility of GFEBS becoming the system of record for the Corps of Engineers. [21] The six Navy commands are the Naval Air Systems Command, the Naval Supply Systems Command, the Space and Naval Warfare Systems Command, the Naval Sea Systems Command, the Strategic Systems Program, and the Office of Naval Research and Strategic Systems Planning. [22] The military services integrated personnel and pay system is a replacement for the Defense Integrated Military Human Resources System that was intended to provide a joint, integrated, standardized personnel and pay system for all military personnel. [23] Full deployment means with respect to a major automated information system program, the fielding of an increment of the program in accordance with the terms of a full deployment decision-- the final decision made by the MDA authorizing an increment of the program to deploy software for operational use. Pub. L. No. 111-84, div. A, §841, 123 Stat. 2190, 2418 (Oct. 28, 2009), the National Defense Authorization Act for Fiscal Year 2010, directed that the terminology be changed from full operational capability to full deployment. [24] A life-cycle cost estimate provides an accounting of all resources and associated cost elements required to develop, produce, deploy, and sustain a particular program. The life-cycle cost estimate encompasses all past, present, and future costs for every aspect of the program, regardless of funding source. [25] ERPs are developed in accordance with various models using terminology that varies among defense organizations and in some cases even within a given military service. For example, the Army's GFEBS refers to a scheduled segment as a "release," and within a release, there are "waves." The Air Force's DEAMS program refers to scheduled segments as "increments" and within increments, there are "spirals." For the purposes of this report, we refer generally to scheduled segments of implementation as "phases." [26] Master data are that persistent, non-transactional data that defines a business entity for which there is, or should be, an agreed upon view across the organization. This key business information may include data about customers, products, employees, materials, and suppliers. Master data are often used by several functional groups and stored in different data systems across an organization and may or may not be referenced centrally; therefore, the possibility exists for duplicate master data and/or inaccurate master data. [27] IOT&E are conducted on production or production-representative articles to determine whether systems are operationally effective and suitable. [28] Conditions or issues needing resolution may be placed upon the ERPs by the MDA, the IRB, or the DBSMC during the business system's funding certification and acquisition decision review milestone process. These conditions are generally noted in a memorandum. [29] U.S. Army Test and Evaluation Command, Operational Test Agency Evaluation Report for the General Fund Enterprise Business System (Alexandria, Va.: Dec. 16, 2009). [30] According to the Software Engineering Institute, requirements management is a process that establishes a common understanding between the customer and the software project manager regarding the customer's business needs that will be addressed by a project. A critical part of this process is to ensure that the requirement development portion of the effort documents, at a sufficient level of detail, the problems that need to be solved and the objectives that need to be achieved. [31] GAO, DOD Business Systems Modernization: Billions Continued to be Invested with Inadequate Management Oversight and Accountability, [hyperlink, http://www.gao.gov/products/GAO-04-615] (Washington, D.C.: May 27, 2004). [32] GAO, Army Depot Maintenance: Ineffective Oversight of Depot Maintenance Operations and System Implementation Efforts, [hyperlink, http://www.gao.gov/products/GAO-05-441] (Washington, D.C.: June 30, 2005). [33] [hyperlink, http://www.gao.gov/products/GAO-10-461]. [34] At the time LMP was designated as a MAIS program in December 2007, it was required to comply with the DOD guidance for MAIS programs. This guidance requires, among other things, that a MAIS program have a completed and approved acquisition program baseline-- the baseline description of the program, including the life-cycle cost estimate--prior to Milestone B approval. The $2.6 billion is the only life-cycle cost estimate that has been developed for the program. [35] [hyperlink, http://www.gao.gov/products/GAO-08-896]. [36] According to DOD, the purpose of the IUID is to facilitate asset accountability and tracking, including the identification and aggregation of related costs to derive the full cost of a contract deliverable. [37] BTA defines SFIS as a comprehensive "common business language" that supports information and data requirements for budgeting, financial accounting, cost/performance management, and external reporting across the DOD enterprise. [38] January 2013 is the estimated full deployment date in the proposed acquisition program baseline submitted for MDA approval in February 2010. [39] [hyperlink, http://www.gao.gov/products/GAO-08-822]. [40] According to the May 10, 2004, analysis of alternatives, this estimate was a "rough order of magnitude" for research and development, procurement and operations and support from fiscal years 2004 through 2011. [41] According to the July 15, 2005, economic analysis, program costs are estimated from fiscal years 2005 through 2018, in base year 2005 dollars, and exclude $9.6 million associated with supporting and maintaining legacy systems during GCSS-MC development and $11.9 million in fiscal year 2004 sunk costs. [42] The acquisition program baseline is an important document for program management and should reflect the approved program being executed. In this regard, the acquisition program baseline formally documents the program's estimated cost, schedule, and performance goals. [43] [hyperlink, http://www.gao.gov/products/GAO-08-866]. [44] [hyperlink, http://www.gao.gov/products/GAO-08-866]. [45] [hyperlink, http://www.gao.gov/products/GAO-08-866]. [46] The program objective memorandum details planned resource allocation 6 years in the future. [47] Each military department refers to its respective personnel and pay system by a different name--the Integrated Personnel and Pay System--Army, the Navy Future Pay and Personnel Solution, and the Air Force Integrated Personnel and Pay System. For purposes of this report, we are collectively referring to these efforts as the Service Specific Integrated Personnel and Pay Systems--a name used by DOD. [48] A critical path is the longest duration path through a sequenced list of activities within a schedule. A schedule risk analysis uses statistical techniques to predict a level of confidence in meeting a completion date. [49] [hyperlink, http://www.gao.gov/products/GAO-08-822] and [hyperlink, http://www.gao.gov/products/GAO-08-896]. [50] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]; OMB Revised Circular No. A-94, Guidelines and Discount Rates for Benefit-Cost Analysis of Federal Programs (Oct. 29, 1992); and DOD Instruction 7041.3, Economic Analysis of Decisionmaking (Nov. 7, 1995). [51] [hyperlink, http://www.gao.gov/products/GAO-08-822] and [hyperlink, http://www.gao.gov/products/GAO-08-896]. [52] See, for example, [hyperlink, http://www.gao.gov/products/GAO-09-3SP]; and OMB Capital Programming Guide V 2.0, Supplement to Office of Management and Budget Circular A- 11, Part 7: Planning, Budgeting, and Acquisition of Capital Assets (Washington, D.C.: June 2006). [53] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [54] A constraint predefines the start, finish, or both dates of an activity. The schedule should use logic and durations in order to reflect realistic start and completion dates for activities. [55] These unusual links are known as start-to-finish links, and they are rarely, if ever, used in scheduling. Schedules should contain a predominance of finish-to-start logical relationships so that one can know which activities must finish before others begin. [56] Summary activities summarize the effort of multiple lower-level tasks. [57] Float is the amount of time by which a predecessor activity can slip before the delay affects successor activities. [58] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [59] An independent cost estimate is another estimate based on the same technical information that is used to validate and cross-check the baseline estimate, but is prepared by a person or organization that has no stake in the approval of the project. [60] [hyperlink, http://www.gao.gov/products/GAO-07-860]. [61] OMB Circular No. A-11, Preparation, Submission, and Execution of the Budget (June 2006); OMB Circular No. A-130, Revised, Management of Federal Information Resources (Nov. 28, 2000); and Office of Management and Budget, Capital Programming Guide: Supplement to Circular A-11, Part 7, Preparation, Submission, and Execution of the Budget (June 2000). [62] Pub. L. No. 104-106, div. E, title LI, § 5123, 110 Stat. 679, 683- 84 (Feb. 10, 1996), codified, as amended, at 40 U.S.C. § 11313. [63] GAO, Tax Administration: IRS Needs to Further Refine Its Tax Filing Season Performance Measures, [hyperlink, http://www.gao.gov/products/GAO-03-143] (Washington, D.C.: Nov. 22, 2002). [64] [hyperlink, http://www.gao.gov/products/GAO-10-461]. [65] This engagement focused on nine ERP efforts that DOD considers critical to transforming its business operations and resolving some of the department's high-risk areas such as business transformation, business system modernization, financial management, and supply chain management. [66] [hyperlink, http://www.gao.gov/products/GAO-10-461], [hyperlink, http://www.gao.gov/products/GAO-09-841], [hyperlink, http://www.gao.gov/products/GAO-08-896], [hyperlink, http://www.gao.gov/products/GAO-08-866], [hyperlink, http://www.gao.gov/products/GAO-08-822], and [hyperlink, http://www.gao.gov/products/GAO-07-860]. [67] [hyperlink, http://www.gao.gov/products/GAO-08-822] and [hyperlink, http://www.gao.gov/products/GAO-08-896]. [68] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [69] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [70] United States Army, 2009 United States Army Report to Congress (Arlington, Va.), and 2010 Army Report to Congress on Business Transformation (Arlington, Va.: Mar. 1, 2010); Department of the Navy, Congressional Report, NDAA 2009, Section 908, Business Transformation Initiatives for the Military Departments (Arlington, Va.), and Department of the Navy Fiscal Year 2010 Business Transformation Report Update (Arlington, Va.); and United States Air Force, Initial Report on Implementation of NDAA 2009, Business Transformation Initiatives for the Military Departments (Sec 908) (Arlington, Va.: July 2009), and March 2010 Follow-up Report on Implementation of NDAA 2009, Business Transformation Initiatives for the Military Departments (Sec 908) (Arlington, Va.: March 2010). [71] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [72] [hyperlink, http://www.gao.gov/products/GAO-09-3SP]. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO’s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO’s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548: