This is the accessible text file for GAO report number GAO-06-84 
entitled 'Military Readiness: Navy's Fleet Response Plan Would Benefit 
from a Comprehensive Management Approach and Rigorous Testing' which 
was released on November 22, 2005. 

This text file was formatted by the U.S. Government Accountability 
Office (GAO) to be accessible to users with visual impairments, as part 
of a longer term project to improve GAO products' accessibility. Every 
attempt has been made to maintain the structural and data integrity of 
the original printed product. Accessibility features, such as text 
descriptions of tables, consecutively numbered footnotes placed at the 
end of the file, and the text of agency comment letters, are provided 
but may not exactly duplicate the presentation or format of the printed 
version. The portable document format (PDF) file is an exact electronic 
replica of the printed version. We welcome your feedback. Please E-mail 
your comments regarding the contents or accessibility features of this 
document to Webmaster@gao.gov. 

This is a work of the U.S. government and is not subject to copyright 
protection in the United States. It may be reproduced and distributed 
in its entirety without further permission from GAO. Because this work 
may contain copyrighted images or other material, permission from the 
copyright holder may be necessary if you wish to reproduce this 
material separately. 

Report to Congressional Committees: 

United States Government Accountability Office: 

GAO: 

November 2005: 

Military Readiness: 

Navy's Fleet Response Plan Would Benefit from a Comprehensive 
Management Approach and Rigorous Testing: 

GAO-06-84: 

GAO Highlights: 

Highlights of GAO-06-84, a report to congressional committees: 

Why GAO Did This Study: 

The Navy has been transforming itself to better meet 21st century 
needs. Since 2000, the Congress has appropriated about $50 billion 
annually for the Navy to operate and maintain its forces and support 
around 376,000 military personnel. In recognizing that the Navy faces 
affordability issues in sustaining readiness within its historical 
share of the defense budget, the Chief of Naval Operations announced a 
concept called the Fleet Response Plan to enhance its deployment 
readiness status. The Fleet Response Plan is designed to more rapidly 
prepare and sustain readiness in ships and squadrons. 

GAO evaluated the extent to which the Navy has (1) employed a sound 
management approach in implementing the Fleet Response Plan and (2) 
tested and evaluated the effectiveness of the plan and shared results 
to improve implementation. 

What GAO Found: 

In establishing the Fleet Response Plan, the Navy has embraced a major 
change in the way it manages its forces. However, the Navy’s management 
approach in implementing the Fleet Response Plan has not fully 
incorporated sound management practices needed to guide and assess 
implementation. These practices include (1) establishing a coherent 
mission and strategic goals, including resource commitments; (2) 
setting implementation goals and a timeline; and (3) establishing a 
communication strategy. While the Navy has taken a number of positive 
actions to implement the plan, it has not provided readiness goals for 
units other than carrier strike groups; resource and maintenance goals; 
performance measures and timelines; or a communications strategy. Sound 
management practices were not fully developed because senior leaders 
wanted to quickly implement the plan in response to changes in the 
security environment. However, without an overall management plan 
containing all of these elements, it may be difficult for the Navy to 
determine whether its efforts to improve the fleet’s readiness are 
achieving the desired results, adequately measuring overall progress, 
or identifying what resources are needed to implement the Fleet 
Response Plan. 

The Navy has not fully tested and evaluated the Fleet Response Plan or 
developed lessons learned to identify the effectiveness of its 
implementation and success over time. Systematic testing and evaluation 
of new concepts is an established practice to gain insight into how 
systems and capabilities will perform in actual operations. However, 
instead of methodically conducting realistic tests to evaluate the 
Fleet Response Plan, the Navy has tried to demonstrate the viability of 
the plan by relying on loosely linked events that were not part of an 
overall test and evaluation strategy. This approach could impair the 
Navy’s ability to validate the plan and evaluate its success over time. 
In addition, the Navy has not used its lessons learned system to share 
the results of its Fleet Response Plan events or as an analytical tool 
to evaluate the progress of the plan and improve implementation, which 
limits the Navy’s ability to identify and correct weaknesses across the 
fleet. 

The U.S.S. George Washington Carrier Strike Group Sailing to 
Participate in an Exercise: 

[See PDF for image] 

[End of figure] 

What GAO Recommends: 

To facilitate implementation of the Fleet Response Plan, GAO recommends 
that the Navy develop a comprehensive management plan with goals and 
performance measures. GAO also recommends that the Navy develop a 
comprehensive testing and evaluation plan to help determine whether the 
plan has been successful. The Department of Defense generally agreed 
with GAO’s recommendations and described efforts to address them. 

www.gao.gov/cgi-bin/getrpt?GAO-06-84. 

To view the full product, including the scope and methodology, click on 
the link above. For more information, contact Janet St. Laurent at 
(202) 512-4402 or stlaurentj@gao.gov. 

[End of section] 

Contents: 

Letter: 

Results in Brief: 

Background: 

Fleet Response Plan Does Not Fully Incorporate Sound Management 
Practices: 

Navy Has Not Fully Tested and Evaluated the Fleet Response Plan or 
Developed Lessons Learned: 

Conclusions: 

Recommendations for Executive Action: 

Agency Comments and Our Evaluation: 

Appendix I: Scope and Methodology: 

Appendix II: Comments from the Department of Defense: 

Appendix III: GAO Contact and Staff Acknowledgments: 

Figures: 

Figure 1: The U.S.S. George Washington Carrier Strike Group Sailing to 
Participate in an Exercise: 

Figure 2: Fleet Response Plan Readiness Cycle Is Intended to Provide 
Eight Carriers for Surge Deployments: 

Figure 3: Best Practice Steps for a Methodical Study Plan Process: 

United States Government Accountability Office: 

Washington, DC 20548: 

November 22, 2005: 

Congressional Committees: 

A key goal of Navy senior leadership is to transform the Navy to better 
meet 21st century security challenges. Since 2000, Congress has 
appropriated about $50 billion annually for the Navy to operate and 
maintain its forces and support around 376,000 active military 
personnel. Nonetheless, the Navy recognizes it is facing affordability 
issues related to sustaining readiness while developing and procuring 
several types of new ships within its historical share of the defense 
budget. One area where the Navy has made significant changes is in its 
operational posture. In March 2003, the Chief of Naval Operations 
initiated the development of a concept, which became known as the Fleet 
Response Plan,[Footnote 1] to enhance the Navy's deployment readiness 
status. The Fleet Response Plan, as implemented by Fleet Forces Command 
in May 2003, is designed to more rapidly prepare and then sustain 
readiness in ships and squadrons. To achieve this capability, the plan 
alters prior manning, maintenance, and training practices to allow for 
a more responsive and ready naval force. The Navy expects this new 
readiness approach will enable its forces to provide not only presence 
and engagement in forward areas, but also surge a greater number of 
ships on short-notice in response to significant crises without 
increasing the readiness budget. The Fleet Response Plan modifies the 
Navy's pre-2001 rotational deployment policy, replacing 6-month routine 
deployments with more flexible deployment options for as many as eight 
carrier strike groups when and where needed. 

The Fleet Response Plan represents a major change in the way the Navy 
manages its forces. Implementing large-scale change management 
initiatives, such as organizational transformations, can be a complex 
endeavor. Our prior work shows that failure to adequately address--and 
often even consider--a wide variety of management issues is at the 
heart of unsuccessful transformations. We have identified a number of 
key best practices and lessons learned from major public and private 
sector organizational mergers, acquisitions, and 
transformations.[Footnote 2] These sound management practices include, 
for example, establishing a coherent mission and integrated strategic 
goals to guide the transformation, including resource commitments; 
setting implementation goals and a timeline to build momentum and show 
progress from day one; and establishing a communication strategy to 
create shared expectations and report related progress. 

We prepared this report under the Comptroller General's authority and 
are providing it to you because of your oversight of defense issues. We 
have previously reported on the maintenance aspects of the Navy's Fleet 
Response Plan.[Footnote 3] This report focuses on the following two 
questions: (1) To what extent has the Navy employed a sound management 
approach in implementing the Fleet Response Plan? (2) To what extent 
has the Navy tested and evaluated the effectiveness of its Fleet 
Response Plan and shared results to improve its implementation? 

To assess the Navy's management approach in implementing the Fleet 
Response Plan, we obtained and analyzed key messages, briefings, and 
instructions on the Fleet Response Plan and interviewed Department of 
Defense (DOD) and Navy headquarters and fleet officials, and compared 
the Navy's approach with best practices for transformations of large 
organizations. To assess the extent to which the Navy has tested the 
effectiveness of the Fleet Response Plan and shared results to improve 
its implementation, we obtained briefings from and interviewed Navy 
officials, reviewed and queried the Navy Lessons Learned System to 
determine relevant lessons recorded, and examined Navy guidance on test 
and evaluation efforts. We reviewed and validated the Navy Lessons 
Learned System data and determined the data were sufficiently reliable 
for our analysis. We conducted our review from January 2005 through 
August 2005 in accordance with generally accepted government auditing 
standards. The scope and methodology used in our review are described 
in further detail in appendix I. 

Results in Brief: 

While the Navy has taken a number of positive actions to establish the 
Fleet Response Plan, the Navy has not fully developed a comprehensive 
management approach to effectively guide, monitor, and assess 
implementation. Sound management practices that provide a framework for 
implementing and managing programs include (1) establishing a coherent 
mission and integrated strategic goals to guide the transformation, 
including resource commitments; (2) establishing a communication 
strategy to create shared expectations and report related progress; and 
(3) setting implementation goals and a timeline to build momentum and 
show progress. The Navy's implementation of the Fleet Response Plan has 
included some aspects of these practices. For example, the Navy has 
established strategic goals for progressive readiness levels for 
carrier strike groups. However, the Navy has not established specific 
readiness goals for the rest of the fleet or determined the resources 
needed to achieve its goals, although it has stated the plan will be 
budget neutral. The Navy also does not have an official written 
definition of the Fleet Response Plan or communications strategy that 
clearly establishes a coherent mission and integrated strategic goals 
to guide the transformation, including resource commitments. These 
sound management practices were not fully implemented because senior 
leaders wanted to implement the Fleet Response Plan as quickly as 
possible in response to the Chief of Naval Operations' direction. 
Although Navy officials recently tasked the Center for Naval Analyses 
to conduct the study to identify potential goals and performance 
measures, it is not clear what the study will recommend or how long it 
will take for the Navy to take action. Until an overall management plan 
is developed, neither the Navy nor Congress may be able to determine 
whether the Fleet Response Plan has effectively achieved its goals, 
measure the plan's overall progress, or determine what resources are 
needed to implement the plan. 

In addition, the Navy has not fully developed a comprehensive set of 
plans to test and evaluate the Fleet Response Plan and has not 
developed formal lessons learned from past exercises to evaluate the 
plan's effectiveness. DOD has long recognized the importance of testing 
new concepts by using war games and experimentation, and recent Navy 
guidance stresses the importance of establishing a long-range plan for 
testing complex and novel problems. The Navy has identified three 
loosely linked events that Navy officials say demonstrate the viability 
of the plan. However, none of the three events cited by the Navy were 
part of an overall test and evaluation strategy to assess the value of 
the plan in increasing readiness. The Navy has not developed an 
overarching test and evaluation plan because Navy officials believe 
existing readiness reports provide adequate information to assess the 
Fleet Response Plan. However, readiness reports do not produce 
information on important factors such as costs, long-term maintenance 
implications, or quality of life issues. Additionally, the Navy did not 
analyze and evaluate the results of these three events and submit 
formal lessons learned to the Navy Lessons Learned System. Without 
systematic testing and evaluation and use of the lessons learned 
system, the Navy's ability to validate a complex change like the Fleet 
Response Plan, identify and correct problem areas, and disseminate 
lessons learned throughout the fleet is limited. This not only prevents 
ship and command staffs from learning from the experiences of others, 
but it also prevents the Navy Lessons Learned System from possibly 
identifying problems and patterns across the fleet that may require a 
high-level, Navy-wide response. 

To facilitate implementation of the Fleet Response Plan, we recommend 
that the Navy develop a comprehensive management plan with goals and 
performance measures. We also recommend that the Navy develop a 
comprehensive testing and evaluation plan to help determine whether the 
Fleet Response Plan has been successful. In its comments on a draft of 
this report, DOD generally concurred with the report's recommendations. 
DOD concurred with our recommendation to develop a comprehensive 
management plan with goals and performance measures, citing several 
actions it has underway or planned. DOD partially concurred with our 
recommendation to test and evaluate the Fleet Response Plan. However, 
DOD does not plan to conduct no-notice surges as we recommended because 
it views such exercises as unnecessary and costly. We continue to 
believe that no-notice surges are important because they can serve as 
an effective means of gauging whether the Navy is ready to respond to 
real world events, which can occur with little notice. DOD comments and 
our evaluation are discussed on pages 21 and 22. 

Background: 

Composition of a Carrier Strike Group: 

Carrier strike groups are typically centered around an aircraft carrier 
and its air wing, and also include a guided missile cruiser; two guided 
missile destroyers; a frigate; an attack submarine; and one or more 
supply ships with ammunition, fuel, and supplies (such as food and 
spare parts). These groups are formed and disestablished on an as 
needed basis, and their compositions may differ though they contain 
similar types of ships. Figure 1 shows a carrier strike group sailing 
in a group formation as it prepares to participate in an exercise. 

Figure 1: The U.S.S. George Washington Carrier Strike Group Sailing to 
Participate in an Exercise: 

[See PDF for image] 

[End of figure] 

Origin of the Fleet Response Plan: 

Prior to the September 11, 2001, terrorist attacks, only those Navy 
ships and air squadrons at peak readiness were deployed overseas, 
usually for 6 months at a time. Most of the Navy's remaining units were 
not available because they were in early stages of their maintenance or 
training cycles, or because the Navy did not have good visibility of 
the readiness of these units. This prompted the Chief of Naval 
Operations in March 2003 to task the Commander, Fleet Forces Command, 
to develop the Fleet Response Plan concept to enhance the Navy's surge 
capability. The Chief of Naval Operations approved the concept in May 
2003 and further directed the Commander, Fleet Forces Command, to be 
responsible and accountable for effectively implementing the plan. 

Fleet Response Plan Believed to Provide Increased and Flexible 
Readiness: 

The Fleet Response Plan emphasizes an increased level of readiness and 
the ability to quickly deploy naval forces to respond to crises, 
conflicts, or homeland defense needs. The plan applies broadly to the 
entire fleet; however, it only sets specific requirements for carrier 
strike groups. For example, the plan calls for eight carrier strike 
groups to be ready to deploy within 90 days of notification. Six of 
them would be available to deploy within 30 days and the other two 
within 90 days. This is commonly referred to as the 6 + 2 goal. Under 
the Fleet Response Plan, the Navy has developed a surge capability 
schedule that it uses to manage and identify the level of training a 
ship has completed and its readiness to deploy. The schedule contains 
three progressive readiness goals: emergency surge, surge-ready, and 
routine deployable status.[Footnote 4] Each readiness goal specifies 
phases of training that must be completed to achieve the goal. To be 
placed in emergency surge status, a ship or an air squadron needs to 
have completed its unit-level phase training. Achieving surge-ready 
status requires the completion of integrated phase training. Attaining 
routine deployable status requires achievement of all necessary 
capabilities, completion of underway sustainment phase training, and 
certification of the unit for forward deployed operations.[Footnote 5] 

The surge capabilities schedule provides a readiness snapshot for each 
ship, allowing decision makers to quickly determine which ships are 
available to meet the needs of the mission. Figure 2 illustrates how 
the Navy notionally identifies the eight aircraft carriers available 
for surge deployments. The carriers numbered 1 through 6 are expected 
to be ready to deploy within 30 days notice. The carriers labeled "+1" 
and "+2" are expected to able to surge within 90 days notice. The six 
surge-ready carriers include two carriers on deployment (numbered 3 and 
4), one carrier that is part of the forward deployed naval force based 
in Japan (number 6), and three carriers in the sustainment phase 
(numbered 1, 2, and 5). These six carriers are expected to have 
completed postdeployment depot-level maintenance and their unit-level 
phase training. The two additional surge carriers are expected to have 
completed depot-level maintenance but not to have completed unit-level 
phase training. The remaining four carriers are in the maintenance 
phase or deep maintenance.[Footnote 6] 

Figure 2: Fleet Response Plan Readiness Cycle Is Intended to Provide 
Eight Carriers for Surge Deployments: 

[See PDF for image] 

[End of figure] 

Revised Fleet Response Plan Is Being Developed: 

Based on the Navy's experiences during the past 2 years, Fleet Forces 
Command has convened a cross-functional working group to develop a 
refined version of the Fleet Response Plan. This update, known as Fleet 
Response Plan-Enhanced, is intended to further define the Fleet 
Response Plan, modify terminology for progressive readiness states to 
better reflect their meaning, tie in elements such as a human capital 
strategy, and expand the focus of the plan beyond carrier strike groups 
to the entire Navy. It may also extend the Fleet Response Plan's 
current employment cycle length of 27 months. The Fleet Response Plan- 
Enhanced is still under development at this time. 

Fleet Response Plan Does Not Fully Incorporate Sound Management 
Practices: 

The Navy's management approach in establishing the Fleet Response Plan 
as its new readiness construct has not fully incorporated sound 
management practices needed to effectively guide, monitor, and assess 
implementation.[Footnote 7] Studies by several organizations have shown 
that successful organizations in both the public and private sectors 
use sound management practices to assist agencies in measuring 
performance, reporting results, and achieving desired outcomes. These 
practices provide management with a framework for effectively 
implementing and managing programs and shift program management focus 
from measuring program activities and processes to measuring program 
outcomes. Sound management practices include (1) establishing a 
coherent mission and integrated strategic goals to guide the 
transformation, including resource commitments; (2) setting 
implementation goals and a timeline to build momentum and show progress 
from day one; and (3) establishing a communication strategy to create 
shared expectations and report related progress. 

The Navy's implementation of the Fleet Response Plan has included some 
aspects of these practices. For example, the Navy has established some 
strategic goals needed to meet the intent of the plan, such as the 
progressive readiness levels of emergency surge, surge-ready, and 
routine deployable status. The Navy also has established specific 
training actions to support these goals, such as that carrier strike 
groups must complete unit-level training to be certified as emergency 
surge-ready. However, other actions taken by the Navy do not fully 
incorporate these practices. For example, the Navy has identified the 6 
+ 2 surge capability as a readiness goal and performance measure for 
carrier strike groups, but no such goal was established for the rest of 
the fleet. The Navy also has some unofficial goals and performance 
measures regarding manning and maintenance, but these unofficial goals 
and performance measures have not been formally established. For 
example, briefings on the Fleet Response Plan state that the Navy 
desires and needs fully manned ships (i.e., manning at 100 percent of a 
ship's requirement) for the program to be successful. Moreover, 
according to Navy officials, the Navy has not established milestones 
for achieving its results. 

In addition, 2 years after initiating implementation of the Fleet 
Response Plan, the Navy still does not have an official written 
definition of the Fleet Response Plan that clearly establishes a 
coherent mission and integrated strategic goals to guide the 
transformation, including resource commitments. This definition would 
describe the Fleet Response Plan's total scope and contain guidance 
with formal goals and performance measures. The Navy recently has taken 
some action to address this area. In February 2005, the Navy directed 
the Center for Naval Analyses to conduct a study to develop formal 
definitions and guidance as well as identify goals and performance 
measures for the plan. However, it remains to be seen whether this 
study will be completed as planned by November 2005; if it will 
recommend developing and implementing sound management practices, such 
as goals, measures, milestones, and timelines; and whether any 
management improvement recommendations made in the study will be 
implemented by the Fleet Forces Command, the Navy command responsible 
for implementing the Fleet Response Plan. Without goals, performance 
measures, timelines, milestones, benchmarks, and guidance to help 
effectively manage implementation of the Fleet Response Plan and 
determine if the plan is achieving its goals, the Navy may find it more 
difficult to implement the Fleet Response Plan across the entire naval 
force. 

Moreover, despite the Navy's unofficial goal that the Fleet Response 
Plan be budget neutral, as articulated in briefings and by senior 
leaders, the Navy has not yet clearly identified the resources needed 
to achieve its goals or provided a rationale for how these resources 
will contribute to achieving the expected level of performance. Navy 
officials have said that current operations and maintenance funding 
levels, as well as manning at 100 percent of required positions, have 
contributed to successful implementation of the Fleet Response Plan. 
However, officials do not know what level of manning or funding is 
actually required for program success over the long term to avoid any 
unintended consequences, such as greater amounts of deferred 
maintenance. According to Navy officials, it is difficult to attribute 
costs to the plan because there is no single budget line item that 
tracks the costs associated with the Fleet Response Plan. Without 
knowing the funding needed, the Navy may not be able to assess the 
impact of possible future changes in funding on implementing the plan. 
Furthermore, without a comprehensive plan that links costs with 
performance measures and outcomes, neither the Navy nor Congress may be 
able to determine if the Fleet Response Plan is actually achieving its 
unofficial goal of being budget neutral. 

Finally, the Navy also has not developed a comprehensive communications 
strategy that reaches out to employees, customers, and stakeholders and 
seeks to genuinely engage them in a two-way exchange, which is a 
critical step in successfully implementing cultural change or 
transformation. We looked for formal mechanisms that communicated the 
details of the Fleet Response Plan and spoke with personnel from 
carrier strike groups, aircraft carriers, air wings and an air 
squadron, one surface combatant ship, and other command staff.[Footnote 
8] We found that while the Fleet Response Plan was communicated 
extensively to senior-level officers, and the Navy provided numerous 
briefings and messages related to the plan, communication and 
understanding of the plan did not flow through to the lower ranks. 
While the concept of the Fleet Response Plan is generally understood by 
some senior-level officials, many of the lower grade personnel on these 
ships were unaware of the scope, goals, and other aspects of the plan. 
In the absence of clear communication throughout the fleet via an 
overall communications strategy that could increase employee awareness 
of the Fleet Response Plan, its successful implementation could be 
impeded. 

Sound management practices, such as those noted above, were not fully 
used by the Navy because senior leaders wanted to quickly implement the 
Fleet Response Plan in response to the Chief of Naval Operations' 
desires. However, without an overall management plan containing all of 
these elements to guide the implementation of such a major change, it 
may be difficult for the Navy and Congress to determine the extent to 
which the Fleet Response Plan is achieving the desired results, measure 
its overall progress, or determine the resources needed to implement 
the plan. 

Navy Has Not Fully Tested and Evaluated the Fleet Response Plan or 
Developed Lessons Learned: 

The Navy has not fully tested and evaluated the Fleet Response Plan or 
developed lessons learned to identify the effectiveness of its 
implementation and success over time. The methodical testing, 
exercising, and evaluation of new doctrines and concepts is an 
established practice throughout the military to gain insight into how 
systems and capabilities will perform in actual operations. However, 
instead of methodically conducting realistic tests to evaluate the 
Fleet Response Plan, the Navy has tried to demonstrate the viability of 
the plan by relying on loosely linked events that were not part of an 
overall test and evaluation strategy, which impairs the Navy's ability 
to validate the plan and evaluate its success over time. In addition, 
the Navy has not used its lessons learned system to share the results 
of its Fleet Response Plan tests or as an analytical tool to evaluate 
the progress of the plan and improve implementation, which limits the 
Navy's ability to identify and correct weaknesses across the fleet. 

Methodical Tests and Evaluations of New Concepts Are Important: 

Methodically testing, exercising, and evaluating new doctrines and 
concepts is an important and established practice throughout the 
military. DOD has long recognized the importance of using tabletop 
exercises, war games, and experimentation[Footnote 9] to explore 
military doctrine, operational concepts, and organizational 
arrangements. Collectively, these tests and experiments can provide 
important insight into how systems and capabilities will perform in 
actual operations. U.S. Joint Forces Command, which has lead 
responsibility for DOD experimentation on new concepts of operation and 
technologies, states that its experimental efforts aim to foster 
military innovation and improvement by exploring, developing, and 
transferring new concepts and organizational ideas into operational 
reality. 

Particularly large and complex issues may require long-term testing and 
evaluation that is guided by study plans. Joint Forces Command's Joint 
Warfighting Center has an electronic handbook that provides guidance 
for conducting exercises and lays out the steps in an exercise life 
cycle: design; planning; preparation; execution; and analysis, 
evaluation, and reports. The Army also has well-established 
guidance[Footnote 10] governing service studies, analyses, and 
evaluations that the Navy feels is representative of best practices for 
military operations research. This provides an important mechanism 
through which problems pertaining to critical issues and other 
important matters are identified and explored to meet service needs. As 
shown in figure 3, the Army's process involves six major steps that 
create a methodical process for developing, conducting, documenting, 
and evaluating a study. Following a formal study process enables data 
evaluation and development of lessons learned that could be used to 
build on the existing knowledge base. In a roundtable discussion with 
the Fleet Forces Command on the rationale behind Summer Pulse 2004, the 
Navy's major exercise for the Fleet Response Plan, a senior Navy 
official stated, "From the concept, … you need to exercise, … you need 
to practice, … you need to demonstrate it to know you got it right and 
what lessons are there to learn from how we did it."[Footnote 11] 

Figure 3: Best Practice Steps for a Methodical Study Plan Process: 

[See PDF for image] 

[End of figure] 

Other governmental agencies, like GAO, and the private sector also rely 
on detailed study plans, or data collection and analysis plans, to 
guide the development of studies and experiments and the collection and 
analysis of data, and to provide a feedback loop that links the 
outcomes of the study or experiment event and subsequent analysis to 
the original goals and objectives of the study or event. GAO guidance 
states that data collection and analysis plans "should carry forward 
the overall logic of the study so that the connection between the data 
that will be collected and the answers to the study questions will 
become evident."[Footnote 12] 

Recent Navy guidance also recognizes the need for a thorough evaluation 
of complex initiatives. In April 2005, the Navy issued a Study Planning 
and Conduct Guide assembled by the Navy Warfare Development 
Command.[Footnote 13] This guide stresses the importance of 
establishing a long-range plan for complex and novel problems and lays 
out the rationale for detailed study plans for exercises and 
experiments, as they establish a structure in which issues are explored 
and data are collected and analyzed in relation to the established 
goals or objectives for the event. Furthermore, the Navy's guide notes 
that random, inadequately prepared events and a determination just to 
study the problem do not lead to successful resolution of problems that 
may arise in programs and concepts that the Navy is testing and 
evaluating. 

Navy Events to Show Viability of the Fleet Response Plan Have Lacked 
Methodical Testing and Evaluation: 

The Navy has not methodically conducted realistic tests of the Fleet 
Response Plan to demonstrate the plan's viability and evaluate its 
progress and success over time, instead relying on loosely linked 
events and some routine data to demonstrate the viability of the plan. 
The events identified by the Navy as successful tests of the Fleet 
Response Plan are Summer Pulse 2004, the emergency deployment of the 
U.S.S. Abraham Lincoln, and Global War on Terrorism Surge 2005, but of 
these events only Summer Pulse 2004 was driven by the Fleet Response 
Plan with the intent of demonstrating that large numbers of ships could 
be surged. In addition, these events were not part of an overall test 
and evaluation strategy that yielded specific information from which to 
assess the value of the plan in increasing readiness and meeting the 
new 6 + 2 surge capability goal for carrier strike groups. 

Summer Pulse 2004 encompassed a number of previously scheduled 
deployments, exercises, and training events that took place between 
June and August of 2004. The intent of Summer Pulse 2004 was to 
demonstrate the Fleet Response Plan's new readiness construct and the 
Navy's ability to deploy multiple carrier strike groups of varying 
levels of readiness. However, Summer Pulse 2004 was not a methodical 
and realistic test of the Fleet Response Plan for three reasons. First, 
Summer Pulse 2004 did not follow best practices regarding study plans 
and the ability to evaluate the impact and outcomes of the plan. The 
Navy did not develop a formal study plan identifying study objectives, 
data collection requirements, and analysis, or produce a comprehensive 
after-event report describing the study's findings. Navy officials have 
stated that the elements of a formal study plan were there for the 
individual deployments, exercises, and training events constituting 
Summer Pulse 2004, but were not brought together in a single package. 
While the Navy may have had the study elements present for the 
individual exercises, they were not directly linked to testing the 
Fleet Response Plan. Without such a comprehensive study plan and 
overall evaluation, there is no ability to discern potential impacts on 
fleet readiness, maintenance, personnel, and other issues that are 
critical to the Fleet Response Plan's long-term success. Second, Summer 
Pulse 2004 was not a realistic test because all participating units had 
several months' warning of the event. As a result, five carriers were 
already scheduled to be at sea and only two had to surge. Because six 
ships are expected to be ready to deploy with as little as 30 days' 
notice under the plan and two additional carriers within 90 days, a 
more realistic test of the Fleet Response Plan would include no-notice 
or short-notice exercises.[Footnote 14] Such exercises conducted 
without advance notification to the participants would provide the 
highest degree of challenge and realism. Without such exercises, the 
Navy might not be able to realistically practice and coordinate a full 
surge deployment. Third, Summer Pulse 2004 was not a sufficient test 
because the Navy involved only seven carriers instead of the eight 
carriers called for in the plan. Therefore, it did not fully test the 
Navy's ability to meet deployment requirements for the expected force. 

Another event cited by the Navy as evidence of the Fleet Response 
Plan's success is the deployment of the U.S.S. Abraham Lincoln carrier 
strike group while it was in surge status in October 2004. Originally 
scheduled to deploy in the spring of 2005, the Lincoln was deployed 
early to support operations in the Pacific Command area of operation 
and provide aid to areas devastated by a tsunami in the Indian Ocean in 
December 2004. Navy officials said that the Fleet Response Plan enabled 
the Navy to identify a carrier to send to the Pacific and to quickly 
tailor its training package based on its progressive readiness status. 
The Navy touted this rapid response relief work by a strike group 
deployed during surge status as a Fleet Response Plan success story. We 
agree that the Lincoln carrier strike group was able to respond 
quickly. However, the extent to which this event realistically tested 
the Fleet Response Plan's expectations for surging one carrier strike 
group is not known. As with Summer Pulse 2004, the Lincoln deployment 
was not a methodical test of the Fleet Response Plan because there was 
no plan to systematically collect or analyze data that would evaluate 
the outcomes of the Lincoln deployment against Fleet Response Plan- 
related study goals. 

The Navy also pointed to a third event, its recent Global War on 
Terrorism Surge 2005,[Footnote 15] as an indicator that the Fleet 
Response Plan works. The Global War on Terrorism surge was a response 
to a request for forces[Footnote 16] from which the Navy is looking to 
glean Fleet Response Plan-related information about what did and did 
not work when the ships return. However, this is not a good test of the 
Fleet Response Plan because there is no plan showing what specific data 
are being collected or what analytical approaches are being employed to 
assess the ships' experiences. As of September 2005, no other events 
had been scheduled to further test and evaluate the Fleet Response 
Plan. 

The Navy has not developed the kind of comprehensive plans to test and 
evaluate the Fleet Response Plan as recommended by DOD and Navy 
guidance and best practices because Navy officials have stated that 
existing readiness reporting processes effectively evaluate the Fleet 
Response Plan's success on a daily basis. They said after-action 
reports[Footnote 17] from training exercises and the Joint Quarterly 
Readiness Review[Footnote 18] assist with this function. Navy officials 
explained that they implemented the Fleet Response Plan the same way 
they had implemented the Inter-Deployment Training Cycle, the 
predecessor to the Fleet Response Plan's Fleet Readiness Training Plan. 
While this may be true, the Inter-Deployment Training Cycle was focused 
on the specific training needed to prepare units for their next 
deployment, not for implementing a new readiness construct that 
emphasized surge versus routine deployments. Furthermore, the Inter- 
Deployment Training Cycle did not contain stated goals whose validity 
the Navy needed to test. In addition, ongoing readiness reports do not 
provide information on important factors such as costs, long-term 
maintenance implications, and quality of life issues. 

The Summer Pulse 2004, Lincoln surge deployment, and Global War on 
Terrorism Surge 2005 testing events were not part of a methodical test 
and evaluation approach. Therefore, the Navy is unable to convincingly 
use these events to evaluate the Fleet Response Plan and determine 
whether the plan has been successful in increasing readiness or 
achieving other goals. Moreover, without effective evaluation of the 
Fleet Response Plan, the Navy may be unable to identify and correct 
potential problem areas across the fleet. Without a comprehensive long- 
range plan that establishes methodical and realistic testing of the 
Fleet Response Plan, the Navy may be unable to validate the Fleet 
Response Plan operational concept, evaluate its progress and success 
over time, and ensure that it can effectively meet Navy goals over the 
long term without any adverse, unintended consequences for maintenance, 
quality of life, and fleet readiness. 

Navy Lessons Learned System's Repository and Analytic Resources Have 
Not Been Used to Catalog and Share Fleet Response Plan Lessons: 

The formal Navy repository for lessons learned, the Navy Lessons 
Learned System, has not been used to disseminate Fleet Response Plan- 
related lessons learned or to analyze test results to evaluate the 
progress of the plan and improve implementation. The Navy Lessons 
Learned System has been designated by the Chief of Naval Operations as 
the singular Navy program for the collection, validation, and 
distribution of unit feedback as well as the correction of problems 
identified and derived from fleet operations, exercises, and 
miscellaneous events. However, there are no mechanisms or requirements 
in place to force ships, commands, and numbered fleet staffs to submit 
all lessons learned to the Navy Lessons Learned System, although such 
mechanisms exist for the submission of port visit and other reports. 
For the events that the Navy cites as tests of the Fleet Response Plan, 
it did not analyze and evaluate the results and produce formal lessons 
learned to submit to the Navy Lessons Learned System for recordation 
and analysis. Any evaluation done of the testing events has not been 
incorporated into the Lessons Learned System, preventing comprehensive 
analyses of lessons learned and identification of problems and patterns 
across the fleet that may require a high-level, Navy-wide response. 

Some ship and carrier strike group staff informed us that they prefer 
informal means of sharing lessons learned, because they feel the 
process through which ships and commands have to submit lessons learned 
for validation and inclusion in the database can be complex and 
indirect. This may prevent ship and command staffs across the fleet 
from learning from the experiences of others, but it also prevents the 
Navy Lessons Learned System from performing comprehensive analyses of 
the lessons learned and possibly identifying problems and patterns 
across the fleet that may require a high-level Navy-wide response. In 
addition, the lessons learned are recorded by mission or exercise 
(e.g., Operation Majestic Eagle) and not by operational concept (e.g., 
the Fleet Response Plan), making identification of Fleet Response Plan- 
specific lessons learned difficult and inconsistent. 

Over the last 10 years, we have issued several reports related to 
lessons learned developed by the military. We have found that service 
guidance does not always require standardized reporting of lessons 
learned[Footnote 19] and lessons learned are not being used in training 
or analyzed to identify trends and performance weaknesses.[Footnote 20] 
We emphasized that effective guidance and sharing of lessons learned 
are key tools used to institutionalize change and facilitate efficient 
operations. We found that despite the existence of lessons learned 
programs in the military services and the Joint Staff, units repeat 
many of the same mistakes during major training exercises and 
operations. Our current review indicates that the Navy still does not 
include all significant information in its lessons learned database. 
Therefore, Navy analysts cannot use the database to perform 
comprehensive analyses of operational concepts like the Fleet Response 
Plan to evaluate progress and improve implementation. 

Officials from the Navy Warfare Development Command stated that the 
Navy is currently drafting a new Chief of Naval Operations Instruction 
governing the Navy Lessons Learned System that will address some of 
these issues.[Footnote 21] Navy Warfare Development Command officials 
hope that the new instruction will result in several improvements over 
the current system. First, they would like to see a dual reporting 
system, so that lessons learned are simultaneously sent to the Navy 
Lessons Learned System for preliminary evaluation when they are 
submitted to the numbered fleets for validation. This would allow Navy 
Lessons Learned analysts to look at unvarnished data for patterns or 
issues of interest to the Chief of Naval Operations, without taking 
away the numbered fleets' validation processes. In addition, officials 
would like to establish deadlines for the submission of lessons learned 
to ensure timeliness. Not only will these changes add value to the data 
stored in the Navy Lessons Learned System, but they will keep the data 
flowing while ensuring that data are actually submitted and not lost as 
they move up the chain of command. According to Navy Lessons Learned 
officials, other branches of the military already allow operators in 
the field to submit lessons learned directly to their lessons learned 
systems, enabling value-added analysis and the timely posting of 
information. By addressing these issues, the Navy can help ensure that 
the lessons learned process will become more efficient, be a command 
priority, and produce actionable results. 

Conclusions: 

Two years after implementing a major change in how it expects to 
operate in the future, the Navy has not taken all of the steps needed 
to enable the Navy or Congress to assess the effectiveness of the Fleet 
Response Plan. As the Navy prepares to implement the Fleet Response 
Plan across the entire naval force, it becomes increasingly important 
that the Navy effectively manages this organizational transformation so 
that it can determine if the plan is achieving its goals. The absence 
of a more comprehensive overarching management plan to implement the 
Fleet Response Plan has left essential questions about definitions, 
goals, performance measures, guidance, timelines, milestones, 
benchmarks, and resources unanswered, even though sound management 
practices recognize the need for such elements to successfully guide 
activities and measure outcomes. The absence of these elements could 
impede effective implementation of the Fleet Response Plan. 
Furthermore, without a comprehensive plan that links costs with 
performance measures and outcomes, neither the Navy nor Congress may be 
able to determine if the Fleet Response Plan is budget neutral. More 
effective communications throughout the fleet using an overall 
communications strategy could increase employee awareness of the plan 
and help ensure successful implementation. 

The Navy also has not developed a comprehensive long-range plan for 
testing and evaluating the Fleet Response Plan. Without a well- 
developed plan and methodical testing, the Navy may not be aware of all 
of the constraints to successfully surging its forces to crises in a 
timely manner. Moreover, the absence of an overarching testing and 
evaluation plan that provides for data collection and analysis may 
impede the Navy's ability to use its testing events to determine 
whether the Fleet Response Plan has been successful in increasing 
readiness and to identify and correct problem areas across the fleet. 
Failure to document and record the results of testing and evaluation 
efforts in the Navy Lessons Learned System could limit the Navy's 
ability to validate the value of the concept, identify and correct 
performance weaknesses and trends across the fleet, perform 
comprehensive analyses of lessons learned, and disseminate these 
lessons and analyses throughout the fleet. 

Recommendations for Executive Action: 

To facilitate successful implementation of the Fleet Response Plan and 
enhance readiness and ensure the Navy can determine whether the plan 
has been successful in increasing readiness and is able to identify and 
correct performance weaknesses and trends across the fleet, we 
recommend that the Secretary of Defense take the following two actions: 

* Direct the Secretary of the Navy to develop a comprehensive 
overarching management plan based on sound management practices that 
will clearly define goals, measures, guidance, and resources needed for 
implementation of the Fleet Response Plan, to include the following 
elements: 

* establishing or revising Fleet Response Plan goals that identify what 
Fleet Response Plan results are to be expected and milestones for 
achieving these results, 

* developing implementing guidance and performance measures based on 
these goals, 

* identifying the costs and resources needed to achieve each 
performance goal, and: 

* communicating this information throughout the Navy. 

* Direct the Secretary of the Navy to develop a comprehensive plan for 
methodical and realistic testing and evaluation of the Fleet Response 
Plan. Such a comprehensive plan should include a description of the 
following elements: 

* how operational tests, exercises, war games, experiments, 
deployments, and other similar events will be used to show the 
performance of the new readiness plan under a variety of conditions, 
including no-notice surges; 

* how data will be collected and analyzed for these events and 
synthesized to evaluate program success and improvements; and: 

* how the Navy Lessons Learned System will collect and synthesize 
lessons from these events to avoid repeating mistakes and improve 
future operations. 

Agency Comments and Our Evaluation: 

In written comments on a draft of this report, DOD generally concurred 
with our recommendations and cited actions it will take to implement 
the recommendations. 

DOD concurred with our recommendation that the Navy should develop a 
comprehensive overarching management plan based on sound management 
practices that would clearly define the goals, measures, guidance, and 
resources needed for successful implementation of the Fleet Response 
Plan, including communicating this information throughout the Navy. DOD 
noted that the Navy has already taken action or has plans in place to 
act on this recommendation, and described several specific 
accomplishments and ongoing efforts in this regard. DOD also noted that 
the Navy intends to communicate through message traffic, white papers, 
instructions, lectures, and meetings with Navy leadership. We agree 
that these means of communication are an important part of an effective 
communication strategy; however, we do not believe that these methods 
of communication constitute a systemic strategy to ensure communication 
at all personnel levels. We believe the Navy would benefit from a 
comprehensive communication strategy that builds on its ongoing 
efforts, but encompasses additional actions to ensure awareness of the 
plan throughout the Navy. 

DOD partially concurred with our recommendation to test and evaluate 
the Fleet Response Plan. DOD noted that it plans to use a variety of 
events and war games to evaluate the Fleet Response Plan, but it does 
not see a need to conduct no-notice surges to test the Fleet Response 
Plan. DOD stated that it believes no-notice surges are expensive and 
unnecessary and could lead to penalties on overall readiness and the 
ability to respond to emergent requirements. DOD also noted that the 
Navy has surged single carrier strike groups, expeditionary strike 
groups, and individual ships or units under the Fleet Response Plan, 
and it cited several examples of such surges. We commend the Navy's 
plans to use a variety of events to evaluate the Fleet Response Plan 
and its use of the Navy Lessons Learned System to report and evaluate 
the lessons learned in the Global War on Terrorism Surge 2005 exercise 
held earlier this year. However, we continue to believe that no-notice 
surges are critical components of realistic testing and evaluation 
plans and that the benefits of such exercises can outweigh any 
additional costs associated with conducting such tests on a no-notice 
basis. Both we and Congress have long recognized the importance of no- 
notice exercises. For example, in a 1989 report, we noted that DOD was 
instituting no-notice exercises to assess the preparedness of combatant 
commands' state of training of their staffs and components.[Footnote 
22] In addition, in 1990 the Department of Energy conducted no-notice 
tests of security personnel in response to our work and out of 
recognition that such tests are the best way to assess a security 
force's ability at any given time.[Footnote 23] Furthermore, in recent 
years, the Department of Homeland Security, Department of Energy, and 
others have conducted no-notice exercises because they add realism and 
demonstrate how well organizations are actually prepared to respond to 
a given situation. Despite the importance of no-notice exercises, the 
Navy has not conducted no-notice exercises to test and evaluate the 
centerpiece surge goal of 6 + 2 for carrier strike groups. We believe 
that the smaller surges cited by DOD can provide insights into the 
surging process, but we do not believe that such surges can effectively 
test the Navy's readiness for a full 6 + 2 carrier strike group surge. 

DOD also provided technical and editorial comments, which we have 
incorporated as appropriate. DOD's comments are reprinted in appendix 
II of this report. 

We are sending copies of this report to other interested congressional 
committees; the Secretary of Defense; the Secretary of the Navy; and 
the Director, Office of Management and Budget. We will make copies 
available to others upon request. In addition, the report will be 
available at no charge on the GAO Web site at http://www.gao.gov. 

If you or your staff have any questions about this report, please 
contact me at (202) 512-4402 or stlaurentj@gao.gov. Contact points for 
our Offices of Congressional Relations and Public Affairs may be found 
on the last page of this report. GAO staff who made major contributions 
to this report are listed in appendix III. 

Signed by: 

Janet St. Laurent: 
Director, Defense Capabilities and Management: 

List of Committees: 

The Honorable John Warner: 
Chairman: 
The Honorable Carl Levin: 
Ranking Minority Member: 
Committee on Armed Services: 
United States Senate: 

The Honorable Ted Stevens: 
Chairman: 
The Honorable Daniel K. Inouye: 
Ranking Minority Member: 
Subcommittee on Defense: 
Committee on Appropriations: 
United States Senate: 

The Honorable Duncan L. Hunter: 
Chairman: 
The Honorable Ike Skelton: 
Ranking Minority Member: 
Committee on Armed Services: 
House of Representatives: 

The Honorable C. W. Bill Young: 
Chairman: 
The Honorable John P. Murtha 
Ranking Minority Member: 
Subcommittee on Defense: 
Committee on Appropriations: 
House of Representatives: 

[End of section] 

Appendix I: Scope and Methodology: 

To assess the extent to which the Navy has employed a sound management 
approach in implementing the Fleet Response Plan, we interviewed Navy 
headquarters and fleet officials; received briefings from relevant 
officials; and reviewed key program documents. In the absence of a 
comprehensive planning document, we compared best practices for 
managing and implementing major efforts to key Navy messages, 
directives, instructions, and briefings, including, but not limited to, 
the Culture of Readiness message sent by the Chief of Naval Operations 
(March 2003); the Fleet Response Concept message sent by the Chief of 
Naval Operations (May 2003); the Fleet Response Plan Implementation 
message sent by the Commander, Fleet Forces Command (May 2003); the 
Fleet Response Plan Implementation Progress message sent by the 
Commander, Third Fleet (September 2003); and the U.S. Fleet Forces 
Command's Fleet Training Strategy instruction (May 2002 and an undated 
draft). We also conducted meetings with several of the commanding 
officers, executive officers, and department heads of selected carrier 
strike groups, aircraft carriers, and air wings to obtain information 
on how the plan had been communicated, how the plan had changed their 
maintenance and training processes, the impact on their quality of 
life, the cost implications of the plan, and other factors. 

To assess the extent to which the Navy has tested the effectiveness of 
the Fleet Response Plan and shared results to improve its 
implementation, we obtained briefings; interviewed Navy headquarters 
and fleet officials; and reviewed test and evaluation guidance for both 
the Navy and other federal agencies. To evaluate the three Fleet 
Response Plan demonstrations identified by the Navy, we interviewed 
officials from the Fleet Forces Command and the Navy Warfare 
Development Command, reviewed existing documentation on the 
demonstrations, queried the Navy Lessons Learned System for lessons 
learned from the demonstrations, and compared our findings to accepted 
best practices for tests and evaluations. Further, we reviewed Navy 
Lessons Learned System instructions and queried the system to determine 
recorded lessons learned pertaining to the Fleet Response Plan. 

We validated the Navy Lessons Learned System data and determined the 
data were sufficiently reliable for our analysis. We conducted our 
review from January 2005 through August 2005 in accordance with 
generally accepted government auditing standards at the following 
locations: 

* The Joint Staff, Washington, D.C. 

* U.S. Pacific Command, Camp H. M. Smith, Hawaii: 

* Offices of the Chief of Naval Operations, Washington, D.C. 

* Naval Sea Systems Command, Washington, D.C. 

* U.S. Fleet Forces Command, Norfolk, Virginia: 

* Offices of the Fleet Forces Command: 

* Commander, U.S. Second Fleet: 

* Commander, Naval Air Forces: 

* Commander, Submarine Forces: 

* Commander, Naval Surface Force: 

* U.S. Marine Corps Forces: 

* Afloat Training Group: 

* Navy Warfare Development Command, Newport, Rhode Island: 

* Commander, U.S. Pacific Fleet, Pearl Harbor, Hawaii: 

* Offices of the U.S. Pacific Fleet: 

* Commander, Naval Submarine Force: 

* Asia-Pacific Center for Security Studies, Honolulu, Hawaii: 

We held group discussions with selected personnel such as commanding 
officers, executive officers/chief of staffs, department heads, and 
crew members from the following units, all located in the Norfolk, 
Virginia, area: 

* U.S.S. Bulkeley: 

* U.S.S. Enterprise: 

* U.S.S. Theodore Roosevelt: 

* U.S.S. Harry S. Truman: 

* Carrier Air Wing 3: 

* Carrier Air Wing 8: 

* Carrier Strike Group 2: 

* Carrier Strike Group 10: 

[End of section] 

Appendix II: Comments from the Department of Defense: 

OFFICE OF THE UNDER SECRETARY OF DEFENSE: 
PERSONNEL AND READINESS: 
4000 DEFENSE PENTAGON: 
WASHINGTON, D.C. 20301-4000: 

OCT 31 2005: 

Janet A. St. Laurent: 
Director, Defense Capabilities and Management: 
U.S. Government Accountability Office: 
441 G Street, N.W.: 
Washington, DC 20548: 

Dear Ms. St. Laurent: 

This is the Department of Defense (DoD) response to the GAO draft 
report, GAO-06-84, "MILITARY READINESS: Navy's Fleet Response Plan 
Would Benefit From a Comprehensive Management Approach and Rigorous 
Testing" (GAO Code 350625)," dated September 27, 2005. 

DoD appreciates the opportunity to comment on the draft report. DoD 
concurs with Recommendation 1 and partially concurs with Recommendation 
2. Detailed comments on the GAO recommendations and report are 
enclosed: 

Sincerely, 

Signed by: 

Joseph J. Angello: 
Director: 
Readiness Programming & Assessment: 

Enclosure: As Stated: 

GAO DRAFT REPORT - DATED September 27, 2005 GAO CODE 350625/GAO-06-84: 

"MILITARY READINESS: Navy's Fleet Response Plan Would Benefit From a 
Comprehensive Management Approach and Rigorous Testing" 

DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS: 

RECOMMENDATION 1: The GAO recommended that the Secretary of Defense 
direct the Secretary of the Navy to develop a comprehensive overarching 
management plan based on sound management practices that will clearly 
define goals, measures, guidance, and resources needed for 
implementation of the Fleet Response Plan, to include the following 
elements: 

* establish or revise Fleet Response Plan goals that identify what 
Fleet Response Plan results are to be expected and milestones for 
achieving these results; 

* develop implementing guidance and performance measures based on these 
goals; 

* identify the costs and resources needed to achieve each performance 
goal; and: 

* communicate this information throughout the Navy. 

DoD RESPONSE: Concur. The Navy has either already developed or is in 
process of developing guidance, performance measures, and a methodology 
to capture Fleet Response Plan (FRP) costs and efficiencies. 
Additionally, the Navy leadership is continuously working to educate 
and inform personnel, at all levels, on the dynamics, requirements, and 
benefits of the FRP. 

Since its very inception, the FRP is designed as a dynamic and flexible 
operational process that affords all levels of command an ability to 
adapt to changing world environments and requirements. Therefore, the 
management and process tracking of FRP will also change in concert with 
those emergent requirements. 

Specific accomplishments and ongoing efforts are described below: 

* FRP goals and expectations were initially established and 
communicated to the Fleet via formal message traffic and guidance from 
subordinate commanders. CNO's "personal for" Commanders, Commanding 
Officers, and Officers in Charge dated 22 May 2003 outlined the 
requirement to develop a process that would improve Navy's speed of 
response. It announced the approval of the Fleet Response Concept and 
directed CFFC to develop FRP to make the concept a reality. 
Supplemental guidance from CFFC explained FRP and delineated key 
implementation milestones. More recent and expanded guidance can be 
seen in the draft CFFC Instruction 3501.3A titled Fleet Training 
Continuum. This document discusses and identifies in-depth guiding 
principles, roles and responsibilities, business rules, and critical 
training enablers. 

* Overarching guidance in the form of a CNO FRP instruction is in work 
with OPNAV N43 as lead agent. This instruction will address all facets 
of FRP and will include references and descriptions of performance 
based measures and consolidate implementation guidance. Additionally, 
significant progress has been made to standardize performance measures 
via a capabilities-based approach, using the Navy Mission Essential 
Task Lists (NMETLs). These are the basic building blocks used in 
determining Fleet training requirements, plans, execution, and 
readiness assessment. 

* Identification of costs and resources needed to efficiently operate 
the FRP has been undertaken by several organizations under the 
cognizance of CFFC. The Center for Naval Analysis (CNA) will deliver 
the results of a comprehensive study on the "Resource Implications of 
the FRP". This study is currently expected to be published in November 
2005. Fleet Forces Command is also developing the Fleet Training 
Capability Cost System (FTCCS) which is an Activity Based 
Costing/Management system (ABC/M) designed to trace Fleet resource 
utilization and costs. 

* As previously described, Navy leadership has worked diligently to 
communicate evolving changes to Fleet plans to prepare and respond to 
our changing security environment. This work will continue through 
message traffic, white papers, instructions, and face-to-face lectures 
and meetings with Navy leadership. 

RECOMMENDATION 2: The GAO recommended that the Secretary of Defense 
direct the Secretary of the Navy to develop a comprehensive plan for 
methodical and realistic testing and evaluation of the Fleet Response 
Plan. Such a comprehensive plan should include a description of the 
following elements: 

* how operational tests, exercises, war games, experiments, 
deployments, and other similar events will be used to show the 
performance of the new readiness plan under a variety of conditions, 
including no-notice surges; 

* how data will be collected and analyzed for these events and 
synthesized to evaluate program success and improvements, and: 

* how the Navy Lessons Learned System will collect and synthesize 
lessons from these events to avoid repeating mistakes and improve 
future operations. 

DoD RESPONSE: Partially Concur. Continual evaluation is a key tenet of 
the revised Fleet Response Plan (FRP) and the Navy plans on using a 
variety of events to include deployments and war games to evaluate FRP. 
However, conducting no-notice surges to "test' 'the Fleet Response Plan 
is an expensive and unnecessary means to evaluate the FRP, and could 
lead to penalties on overall readiness and ability to respond to future 
emergent requirements. The FRP is an operational construct which is 
subject to methodical and realistic "testing and evaluation" during 
ongoing deployments and operations. To expand: 

The Navy's mission is global response and providing our nation and 
allies dominant naval power wherever and whenever required. Our 
maritime forces do this through persistence, precision, reach, speed, 
and agility. The FRP is the operational framework that capitalizes on 
investments that have been made to readiness accounts and leverages 
force provider capability to meet global Combatant Commander demand 
signals for traditional roles, e.g. forward presence, and new emerging 
mission areas such as Detention Facility Security, JTF HOA, and Air 
Ambulance. The FRP is an on-going mission-driven response plan that 
provides the right readiness at the right time, enables responsive 
forward presence, and drives our ability to answer the Combatant 
Commanders' demand signals. With FRP the Navy has deployed and 
developed a more agile, flexible and scalable naval force capable of 
surging quickly to deal with unexpected threats, humanitarian disasters 
and contingency operations. 

* A key objective in the FRP evolution is the continual evaluation and 
re-evaluation of mission essential tasks, in response to Combatant 
Commander demands, to create an adaptive capability earlier in the 
training cycle. The Navy has and will continue to use both Unit Level 
Training (ULT) and the ability to respond to Requests For Forces (RFF) 
to evaluate both current training funding levels and FRP sustainment 
funding levels. By aligning key training milestones in accordance with 
national strategy, and Component and Combatant Commander 
feedback/lessons learned our naval forces will deliver a tailored 
combat ready force in the highest state of readiness. 

* To specifically address the no-notice surge requirement, any FRP 
surge is in fact no notice. Should world events require a large no- 
notice surge, the Navy will capture those lessons learned and 
incorporate improvements into the FRP. For example, Global War on 
Terrorism Surge 05 (European and Central Command surges) was no notice. 
The demand signal and supply signal were both generated in May 05 and 
ships deployed 27 May 05, NWDC is currently analyzing. Other surge 
examples: 

- HSV surge in support of PACOM Tsunami relief: 

- Hurricanes Katrina and Rita Navy response (HARRY S TRUMAN underway 96 
hours after notification): 

* Salvage/Recovery ship/EOD/Divers: 

* Mine Countermeasure ships: 

* Hospital ship (surge from reduced operational status): 

* SEABEES/Cargo Handling/Helicopters/Tactical Air Control/AEW/Maritime 
Preposition: 

* Expeditionary medical facilities: 

* Causeway transport: 

- SAIPAN Expeditionary Strike Group readiness to surge in 96 hours to 
support Haiti operations (JCS did not execute): 

- SAIPAN Expeditionary Strike Group 96 hour readiness to surge to PACOM 
in support of Tsunami relief operations (JCS did not execute): 

- THEODORE ROOSEVELT Carrier Strike Group surged 7 days early from 
planned deployment to meet CENTCOM increased demand signal for Iraqi 
elections. 

- LINCOLN Carrier Strike Group surge (from Emergency Surge status) to 
cover for KITTY HAWK in maintenance: 

- BOXER/BATAAN and KEARSARGE surge in support of Marine Corps Air 
Combat Element transport to Iraq. 

* The FRP was demonstrated in response to Operation Iraqi Freedom, 
where 6 carriers where deployed to Iraq, and again during SUMMER PULSE 
2004. During SUMMER PULSE over 40 Broad categories of data were 
captured. Prior to conduct of the surge, analysis was conducted to 
predict performance, and was then compared to SUMMER PULSE results, all 
of which were in the expected (fiscally, flight hour, weapons 
expenditure and steaming days constrained) and predicted readiness 
levels. Navy analyzed: 

- Cost/sorties/flight hours; 
- Combat readiness/Mobility ratings; 
- Steaming days/PASSEX events; 
- Mil to Mil exchanges/Conduct of Argentinean and Brazilian aircraft 
aboard USS REAGAN; 
- Training events conducted/weapons expended/ranges utilized (TAN TAN 
in Morocco noteworthy); 
- Anti-submarine warfare time; 
- Logistics/maintenance; 
- Personnel qualification gains; 
- Information Operations/Public Affairs; 
- FRP total ordnance posture; 
- Strategic/Operational and Tactical gains (Theater Security 
Cooperation) with International organizations and foreign nations. 

The Navy will continue to capitalize on the investments to the 
readiness accounts under the Fleet Response Plan framework/operational 
construct. FRP provides a superb base to meet the challenges faced in 
the Global War on Terror and transforming, trans-national threats while 
maintaining significant presence abroad; to assure our allies, dissuade 
our adversaries and respond with scalable Major Combat Operations 
capable forces. It provides the framework to rapidly adapt to new 
mission requirements and meet the demands of Combatant Commanders. It 
is an evolving operational construct that incorporates lessons learned, 
adapts to the changing landscape and maintains the capability to manage 
risk and fiscal responsibilities. 

[End of section] 

Appendix III: GAO Contact and Staff Acknowledgments: 

GAO Contact: 

Janet St. Laurent, (202) 512-4402 or stlaurentj@gao.gov: 

Acknowledgments: 

In addition to the contact named above, Richard Payne, Assistant 
Director; Renee Brown; Jonathan Clark; Nicole Collier; Dawn Godfrey; 
David Marroni; Bethann Ritter; Roderick Rodgers; John Van Schaik; and 
Rebecca Shea made significant contributions to this report. 

FOOTNOTES 

[1] The Fleet Response Plan is also known as the Fleet Readiness 
Program, or simply "FRP." 

[2] See GAO, Highlights of a GAO Forum on Mergers and Transformation: 
Lessons Learned for a Department of Homeland Security and Other Federal 
Agencies, GAO-03-293SP (Washington, D.C.: Nov. 14, 2002); and Results- 
Oriented Cultures: Implementation Steps to Assist Mergers and 
Organizational Transformations, GAO-03-669 (Washington, D.C.: July 2, 
2003). 

[3] See GAO, Defense Logistics: GAO's Observations on Maintenance 
Aspects of the Navy's Fleet Response Plan, GAO-04-724R (Washington, 
D.C.: June 18, 2004). 

[4] Emergency surge status means a unit can be employed in case of 
urgent need but does so at levels of operational risk correlating to 
the level of capability achieved at the time of emergency surge. Surge- 
ready status means that units are ready to be employed at more 
acceptable levels of operational risk commensurate with the level of 
capability achieved at the time of the requirement to surge. Routine 
deployable status means a unit has achieved all required capabilities, 
completed underway training requirements, and is certified for forward 
deployed operations. 

[5] Unit-level phase training focuses on completion of unit-level 
training requirements, including team training both on board and 
ashore; unit-level exercises in port and at sea; and unit inspections, 
assessments, certifications, and qualifications. During this phase, a 
unit becomes proficient in all required capabilities, meets the 
training commander's certification criteria, and becomes ready for more 
complex integrated training events. Integrated phase training brings 
individual units together to conduct strike-group-level integrated 
training and operations in a challenging operational environment as a 
foundation for performing their anticipated deployed mission. 
Sustainment phase training exercises units and staffs in multimission 
planning and execution, including the ability to interoperate 
effectively in a wartime environment. 

[6] The maintenance phase consists of depot-level shipyard maintenance 
for a period of 6 to 12 months. Deep maintenance is for a period of 2 
to 3 years for a nuclear-refueling overhaul. 

[7] The Congress enacted the Government Performance and Results Act of 
1993 to provide for, among other things, the establishment of strategic 
planning and performance measurement in the federal government. Pub. L. 
No. 103-62 (1993). 

[8] We met with people individually and in groups. Personnel included 
commanding officers, executive officers, and senior and junior enlisted 
personnel. Additionally, we met with command master chiefs and command 
career counselors. 

[9] Tabletop exercises are analytical tools that require fewer 
resources than full-fledged live exercises. They provide a means to 
develop both immediate and long-term solutions among functional areas, 
to develop standardization and interoperability of procedures, and to 
document best practices for others to utilize. 

[10] Army Regulation 5-5, Army Studies and Analyses (June 30, 1996). 

[11] Rear Admiral John D. Stufflebeem, "Roundtable: Summer Pulse 
Discussions With Rear Admiral John D. Stufflebeem," Summer Pulse '04 
News Archive (July 8, 2004), 
http://www.cffc.navy.mil/summerpulse04/stufflebeem-transcript.htm. 

[12] See GAO, Quantitative Data Analysis: An Introduction, GAO/PEMD- 
10.1.11, ch. 7 (Washington, D.C.: May 1992). 

[13] The Navy Warfare Development Command's responsibilities include 
being a champion for Navy warfare innovation, operating concepts, and 
concept of operations development in a naval, joint, and coalition 
environment; coordinating the planning and implementation of the Navy's 
experimentation process; managing development, approval, and rapid 
dissemination of naval, joint, and allied doctrine; and managing the 
Navy's Lessons Learned Program. 

[14] No-notice exercises demonstrate participants' ability to rapidly 
respond to unexpected situations. This type of exercise is valued 
because it can lead to improvements in procedures by exercising 
participants in a near-real-world context. 

[15] In the spring of 2005, the Navy surged five ships in support of 
the Global War on Terrorism to work with allies to detect, disrupt, and 
deny international terrorist organizations the use of the maritime 
environment. These ships will also work to build regional security and 
long-term stability. The five ships are the U.S.S. Saipan (LHA 2), 
U.S.S. Nashville (LPD 13), U.S.S. Nicholas (FFG 47), the U.S.S. Gunston 
Hall (LSD 44), and the U.S.S. Philippine Sea (CG 58). 

[16] A request for forces is a special request by a geographic 
combatant commander through the Chairman, Joint Chiefs of Staff, for 
forces in addition to the normal, preapproved force deployments or 
forces assigned. 

[17] An after-action report is a professional discussion of an event, 
focused on performance standards, that enables participants to discover 
for themselves what happened, why it happened, and how to sustain 
strengths and address weaknesses. It is a tool that leaders, trainers, 
and units can use to get maximum benefit from every mission or task. 

[18] The Joint Quarterly Readiness Review is a quarterly readiness 
assessment that identifies capability shortfalls and risks in mission 
execution and identifies appropriate measures for risk reduction. 

[19] See GAO, Chemical and Biological Defense: Army and Marine Corps 
Need to Establish Minimum Training Tasks and Improve Reporting for 
Combat Training Centers, GAO-05-8 (Washington, D.C.: Jan. 28, 2005), 
and Force Structure: Navy Needs to Fully Evaluate Options and Provide 
Standard Guidance for Implementing Surface Ship Rotational Crewing, 
GAO- 05-10 (Washington, D.C.: Nov. 11, 2004). 

[20] See GAO, Military Training: Potential to Use Lessons Learned to 
Avoid Past Mistakes Is Largely Untapped, GAO/NSIAD-95-152 (Washington, 
D.C.: Aug. 9, 1995). 

[21] Chief of Naval Operations Instruction 3500.37D, currently being 
drafted, would replace Chief of Naval Operations Instruction 3500.37C, 
March 19, 2001, Navy Lessons Learned System. 

[22] See GAO, Defense Reorganization: Progress and Concerns at JCS and 
Combatant Commands, GAO/NSIAD-89-83 (Washington, D.C.: Mar. 1, 1989). 

[23] See GAO, Nuclear Safety: Potential Security Weaknesses at Los 
Alamos and Other DOE Facilities, GAO/RCED-91-12 (Washington, D.C.: Oct. 
11, 1990). 

GAO's Mission: 

The Government Accountability Office, the investigative arm of 
Congress, exists to support Congress in meeting its constitutional 
responsibilities and to help improve the performance and accountability 
of the federal government for the American people. GAO examines the use 
of public funds; evaluates federal programs and policies; and provides 
analyses, recommendations, and other assistance to help Congress make 
informed oversight, policy, and funding decisions. GAO's commitment to 
good government is reflected in its core values of accountability, 
integrity, and reliability. 

Obtaining Copies of GAO Reports and Testimony: 

The fastest and easiest way to obtain copies of GAO documents at no 
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains 
abstracts and full-text files of current reports and testimony and an 
expanding archive of older products. The Web site features a search 
engine to help you locate documents using key words and phrases. You 
can print these documents in their entirety, including charts and other 
graphics. 

Each day, GAO issues a list of newly released reports, testimony, and 
correspondence. GAO posts this list, known as "Today's Reports," on its 
Web site daily. The list contains links to the full-text document 
files. To have GAO e-mail this list to you every afternoon, go to 
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order 
GAO Products" heading. 

Order by Mail or Phone: 

The first copy of each printed report is free. Additional copies are $2 
each. A check or money order should be made out to the Superintendent 
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or 
more copies mailed to a single address are discounted 25 percent. 
Orders should be sent to: 

U.S. Government Accountability Office 

441 G Street NW, Room LM 

Washington, D.C. 20548: 

To order by Phone: 

Voice: (202) 512-6000: 

TDD: (202) 512-2537: 

Fax: (202) 512-6061: 

To Report Fraud, Waste, and Abuse in Federal Programs: 

Contact: 

Web site: www.gao.gov/fraudnet/fraudnet.htm 

E-mail: fraudnet@gao.gov 

Automated answering system: (800) 424-5454 or (202) 512-7470: 

Public Affairs: 

Jeff Nelligan, managing director, 

NelliganJ@gao.gov 

(202) 512-4800 

U.S. Government Accountability Office, 

441 G Street NW, Room 7149 

Washington, D.C. 20548: