[House Hearing, 111 Congress]
[From the U.S. Government Printing Office]
[H.A.S.C. No. 111-63]
MEASURING PERFORMANCE:
DEVELOPING GOOD
ACQUISITION METRICS
__________
HEARING
BEFORE THE
PANEL ON DEFENSE ACQUISITION REFORM
OF THE
COMMITTEE ON ARMED SERVICES
HOUSE OF REPRESENTATIVES
ONE HUNDRED ELEVENTH CONGRESS
FIRST SESSION
__________
HEARING HELD
MAY 19, 2009
[GRAPHIC(S) NOT AVAILABLE IN TIFF FORMAT]
U.S. GOVERNMENT PRINTING OFFICE
50-297 WASHINGTON : 2010
-----------------------------------------------------------------------
For sale by the Superintendent of Documents, U.S. Government Printing
Office Internet: bookstore.gpo.gov Phone: toll free (866) 512-1800; DC
area (202) 512-1800 Fax: (202) 512-2104 Mail: Stop IDCC, Washington, DC
20402-0001
PANEL ON DEFENSE ACQUISITION REFORM
ROBERT ANDREWS, New Jersey, Chairman
JIM COOPER, Tennessee K. MICHAEL CONAWAY, Texas
BRAD ELLSWORTH, Indiana DUNCAN HUNTER, California
JOE SESTAK, Pennsylvania MIKE COFFMAN, Colorado
Andrew Hunter, Professional Staff Member
John Wason, Professional Staff Member
Alicia Haley, Staff Assistant
C O N T E N T S
----------
CHRONOLOGICAL LIST OF HEARINGS
2009
Page
Hearing:
Tuesday, May 19, 2009, Measuring Performance: Developing Good
Acquisition Metrics............................................ 1
Appendix:
Tuesday, May 19, 2009............................................ 23
----------
TUESDAY, MAY 19, 2009
MEASURING PERFORMANCE: DEVELOPING GOOD ACQUISITION METRICS
STATEMENTS PRESENTED BY MEMBERS OF CONGRESS
Andrews, Hon. Robert, a Representative from New Jersey, Chairman,
Panel on Defense Acquisition Reform............................ 1
Coffman, Hon. Mike, a Representative from Colorado, Panel on
Defense Acquisition Reform..................................... 2
WITNESSES
Fitch, David P., USN (Retired), Director, AT&L Leadership
Learning Center of Excellence, Defense Acquisition University.. 7
Nussbaum, Dr. Daniel A., Visiting Professor, Department of
Operations Research, Naval Postgraduate School................. 9
Patterson, J. David, USAF (Retired), Executive Director, National
Defense Business Institute, University of Tennessee............ 5
APPENDIX
Prepared Statements:
Andrews, Hon. Robert......................................... 27
Coffman, Hon. Mike........................................... 28
Dillard, John T., Senior Lecturer, Graduate School of
Business and Public Policy, Naval Postgraduate School...... 77
Fitch, David P............................................... 57
Nussbaum, Dr. Daniel A....................................... 66
Patterson, J. David.......................................... 32
Documents Submitted for the Record:
[There were no Documents submitted.]
Witness Responses to Questions Asked During the Hearing:
Mr. Andrews.................................................. 83
Questions Submitted by Members Post Hearing:
[There were no Questions submitted post hearing.]
MEASURING PERFORMANCE: DEVELOPING GOOD ACQUISITION METRICS
----------
House of Representatives,
Committee on Armed Services,
Defense Acquisition Reform Panel,
Washington, DC, Tuesday, May 19, 2009.
The panel met, pursuant to call, at 8:00 a.m., in room
2212, Rayburn House Office Building, Hon. Robert Andrews
(chairman of the panel) presiding.
OPENING STATEMENT OF HON. ROBERT ANDREWS, A REPRESENTATIVE FROM
NEW JERSEY, CHAIRMAN, PANEL ON DEFENSE ACQUISITION REFORM
Mr. Andrews. Good morning, ladies and gentlemen. The panel
will come to order. Appreciate your attendance. I especially
want to thank the witnesses for their very diligent
preparation. We have had the opportunity to review the
testimony ahead of time. It looks like we are going to have a
very engaging and meaningful discussion this morning.
This is the last of our series of hearings looking at the
first question that we are going to be looking at in our
inquiry. The members will recall that we are proceeding on a
series of questions, the first of which is, ``Can we design a
series of metrics that accurately measure the difference, if
any, between the price paid by the taxpayers and the value
received by the taxpayers and the warfighters for the systems
and services that we are buying?''
We are going to proceed after today's hearing with our
second category of questions, which is really hypotheses as to
what has gone wrong. We are going to have a series of panels
talk about their theories and analyses of why we have a
difference between the price paid and the value received.
The first of our two panels on this metrics question dealt
with the sort of measuring the orthodox algorithms that are
used. In the major weapons system side, we had a panel on that,
and on the services side, we had a panel on that.
The purpose of today's panel is to bring in some people who
we think perhaps look at the whole question through a different
prism. They are willing to give us some new perspectives
through which we can analyze the difference between the price
we pay and the value that we receive.
We have three witnesses with a wealth and breadth of
experience in the acquisition field, but also who I think
benefit from what I would call a healthy distance from the
daily responsibilities for that function, so they can give us a
perspective that marries experience with a fresh perspective.
And hopefully, we will be able to use the testimony of these
witnesses to go forward, draw some conclusions about the best
metrics to use in our work, and then proceed with our work in
analyzing the various hypotheses given for why we have suffered
these cost overruns.
I also want to take a moment and thank Mr. Conaway, who
can't be with us this morning but who is ably represented by
Mr. Coffman, and all the members of the panel for their
excellent contribution to the work on the Weapons Acquisition
System Reform Through Enhancing Technical Knowledge and
Oversight (WASTE TKO) bill, which has passed the House. We will
be meeting with the Senate in formal conference at 4:30 this
afternoon to, I believe, put the finishing touches on that
conference report so we will be able to proceed to the floor in
short order on that, get the bill on the President's desk. Each
member of the panel has made a very significant and welcome
contribution to that effort.
Having said that, as the chairman and Mr. McHugh have said,
we think that that effort covers about 20 percent of the
landscape. It looks at major weapons acquisition, which
certainly needed to be examined. But there are so many other
areas that fall outside of that that the panel has to pursue,
as well as, frankly, reviewing the early stage implementation
of the law that we believe the President will sign this week.
So we are by no means at the conclusion of our work. We are
really at the outset of it. And one thing we would ask the
witnesses to do this morning is to think about the fact that,
although the Congress is about to pass major legislation
dealing with major weapons systems, we have not yet addressed
hardware that doesn't fall into the major weapons system
category and the 55 to 60 percent of acquisition that is
services that we have to look at on behalf of those who wear
the uniform and on behalf of the taxpayers.
We are glad that you are here. And at this time, I want to
turn to Mr. Coffman for his opening statement, and we will
proceed with the witnesses after that.
[The prepared statement of Mr. Andrews can be found in the
Appendix on page 27.]
STATEMENT OF HON. MIKE COFFMAN, A REPRESENTATIVE FROM COLORADO,
PANEL ON DEFENSE ACQUISITION REFORM
Mr. Coffman. Good morning. Mr. Chairman, ladies and
gentlemen, I would like to extend a welcome on behalf of
Ranking Member Conaway, as well. He is sorry he could not be
with us today. I would like to thank the chairman for allowing
me to take a few introductory remarks in Representative
Conaway's place.
Today's hearing is an appropriate follow on to the panel's
two hearings which focus on how the Department of Defense (DOD)
and the Government Accountability Office (GAO) currently assess
performance on major weapons systems and service contracts. The
purpose of this hearing is to think outside the box on how we
should measure performance and about what we should be
measuring and less about the how we are doing it today.
Our previous hearings revealed that current measures of
performance tend to break down if the program baseline is
unrealistic. We want to know if the program can be corrected--
if the problem can be corrected. Furthermore, are there metrics
beyond cost and schedule performance that are of value, such as
how closely does the delivery system meet actual warfighter
needs. Does the time of delivery of operational capabilities
satisfy warfighter needs? How do we determine whether optional
or tradable capabilities requested by the warfighter are
affordable?
Mr. Chairman, we have a distinguished panel of witnesses in
front of us today. I have looked at their written testimony,
and I look forward to their testimony. There are a couple of
points I would like to highlight. Although Mr. Dillard is not
testifying today, he did submit a written statement that I
would like to briefly comment on.
[The prepared statement of Mr. Dillard can be found in the
Appendix on page 77.]
Mr. Coffman. He states that, ``The proliferation of
autonomous fiefdoms within the department continues to
increase, with each being a stovepipe of oversight expertise,
imposing unique reporting requirements, assessments and
reviews.''
He goes on to state that, in regards to adding additional
acquisition workforce professionals, that, ``These new people
should not be housed in the Pentagon, but instead where the
execution of programs occur.'' Mr. Chairman, I think this is an
important point that I believe we need to follow closely.
Another point of observation I would like to make is in
regards to requirements in joint programs. We have many
programs out there that have the word ``joint'' in front of
them, but they are joint in name only. I understand that the
current requirements generation process is called the Joint
Capabilities Integration and Development System, or JCIDS.
It seems to me that, in order to have a truly joint program
that the requirements for that program must be borne jointly.
Yet, I do not believe our current system, or as Mr. Dillard
describes, current fiefdoms, foster such an environment. So I
would be interested in hearing from our witnesses in this
regard.
I also encourage our witnesses to share their views on
existing laws and regulations that are particularly helpful or
not helpful to the Department's efforts to obtain the best
value and capability for our warfighters. We have heard from
the Department in two separate hearings about how they
currently measure performance and value on contracts. What we
haven't heard enough about is how they should be measuring
value. Consequently, your input will be greatly appreciated.
With that, I conclude. And again, thank you, fellow
members, and thank you, Mr. Chairman. I look forward to the
witnesses' testimony.
Mr. Andrews. Mr. Coffman, thank you.
Without objection, the statement of any other member of the
panel will be included in the record. And without objection,
the statement of each of the three witnesses, plus the
supplemental floor statement, will be included in the record of
the hearing.
[The prepared statement of Mr. Coffman can be found in the
Appendix on page 28.]
Mr. Andrews. I think each of you are veterans at this
process. You know that we generally have a five-minute rule
where we ask for succinct summaries of your written testimony.
We are going to be a little bit more liberal with that this
morning, but then we are going to turn as quickly as we can to
the questions from the panel and get into the give-and-take.
I am going to read a brief biography of each witness, and
then, Mr. Patterson, we are going to begin with you.
Mr. Dave Patterson is the Executive Director of the
National Defense Business Institute, which he is establishing
at the University of Tennessee in the College of Business
Administration. It is an institution inspiring business
innovation for both government and industry by providing
practical, sound assistance and creating economically efficient
and effective defense business and acquisition programs.
Prior to taking his current duties, he was the Principal
Deputy Under Secretary of Defense Comptroller. As the Principal
Deputy, he was directly responsible for advising and assisting
the Under Secretary of Defense with development, execution and
oversight of the DOD budget exceeding $515 billion, with annual
supplemental requests of more than $160 billion.
From August 2003 to June 2005, Mr. Patterson held duties as
the Special Assistant to the Deputy Secretary of Defense. In
that capacity, he was responsible for managing the Deputy
Secretary of Defense's personal staff, as well as providing
direction and advice to the Office of the Secretary of Defense
staff on a wide range of national security operations and
policy subjects.
He served in the Air Force from 1970 to 1993, retiring in
the rank of colonel. During that time, he held responsible
leadership and management positions, with assignments at the
air wing level as the C5-A aircraft Commander and Deputy
Operations Group Commander at Major Command Headquarters, U.S.
Air Force, the Office of the Chairman, Joint Chiefs of Staff,
the Office of the Secretary of Defense and Inspector General.
In 1986, he was the Air Force Fellow at the American Enterprise
Institute and served in Vietnam, flying 02-As as forward air
controller.
Thank you for your service to our country, Mr. Patterson.
Glad to have you with us.
Mr. Fitch, Mr. David P. Fitch enlisted in the Navy in 1966.
He was commissioned in 1968 after graduation from San Jose
State College and Aviation Officer Candidate School in
Pensacola, Florida.
In 1969, after fixed and rotary wing flight training, he
was designated a naval aviator and a distinguished naval
graduate. He retired from the Navy in 1998 as a captain,
following a career that included three operational and major
acquisition command assignments.
Mr. Fitch culminated his tour in the Navy as the major
program manager for the international and joint multi-
functional information distribution system. He was a 1997
recipient of both the David Packard Excellence in Acquisition
Award and the Department of Defense Value Engineering Award.
In 2001, after nearly three years in the defense industry,
Mr. Fitch joined the faculty of the Defense Acquisition
University, teaching and consulting acquisition executives and
program management and other acquisition disciplines.
In 2006, he led a major independent study of the Coast
Guard Deep Water program. In July of 2008, he assumed the
position of Director, Acquisition, Technology and Logistics
(AT&L) Leadership Learning Center of Excellence after nearly
seven years as the Dean of the Defense Systems Management
College.
Mr. Fitch holds degrees in business and industrial
management from San Jose State college, an M.S. in education
from the University of Southern California, and he is a
graduate with highest distinction from the Naval War College in
Newport, Rhode Island. Thank you, Mr. Fitch, for your service
to our country, and we are glad you are with us this morning.
Dr. Daniel Nussbaum, from 2004 to the present, has been a
visiting professor at the Naval Postgraduate School operations
research department. From 1999 to 2004, he was a principal at
Booz Allen Hamilton, responsible for a broad range of costs,
financial and economic analyses with clients across the
government and commercial spectrum.
From 1996 to 1999, he was the director of the Naval Center
for Cost Analysis at the Office of the Assistant Secretary of
the Navy for Financial Management and Comptroller here in D.C.
From 1987 to 1996, he was division head of the Naval Center for
Cost Analysis. From 1982 to 1986, he was Deputy Director and
Acting Director for operations research and cost analysis
divisions of the Naval Air Systems Command in Washington, D.C.
Again, a very distinguished servant of our country, three
excellent people with experience and fresh insight. We are
happy to have you with us.
Mr. Patterson, we would like you to begin with your
testimony.
STATEMENT OF J. DAVID PATTERSON, USAF (RETIRED), EXECUTIVE
DIRECTOR, NATIONAL DEFENSE BUSINESS INSTITUTE, UNIVERSITY OF
TENNESSEE
Mr. Patterson. Thank you, Mr. Chairman, members of the
Defense Acquisition Reform Panel. I am very pleased to be here
this morning to participate in a discussion of a question that
has clearly captured the attention of the current
Administration and Congress: How should Congress assist the
Department of Defense in improving its acquisition of weapons
and services so that it can meet the needs of the warfighter in
the field while still being a good steward of the taxpayer's
dollars?
The first consideration for judging the success of an
acquisition program is whether it fielded a weapons system or
information system or service in time to make a positive impact
for the warfighter. A system or service fielded too late to
need may as well not have been bought at all. The phrase, ``too
little, too late,'' can mean lost lives.
But before we look at the measures of the acquisition
system merit, there is another consideration central to this
discussion. When Secretary Gates made his budget announcement
on April 6th, 2009, I believe he was speaking from frustration
that was as much about what has been the persistent problem
with the acquisition system that we depend on that is simply
not responsive to the immediate warfighter needs, as it was
about winnowing bloated, failed and unnecessary programs.
Implicit in that expression of frustration is a clear lack
of confidence in a system that actually produces programs with
uncertainty and instability. The most dramatic improvement
metric will be when the senior leadership in the
Administration, Congress and the Department of Defense have
seen such improvements, results, not words, that they can say
they have renewed confidence in the stability, predictability
and effectiveness of the defense acquisition system.
The Defense Acquisition Performance Assessment's Report
contended that program stability and predictability were
singularly and uniquely crucial to managing programs that were
on cost, on schedule and performing. To that end, in the time I
have, allow me to describe two areas of improvement for
measuring program effectiveness worthy of attention.
First, Major Defense Acquisition Programs, or MDAPs, often
start at milestone B, the beginning of engineering and
manufacturing development, with critical staff positions
vacant. Percentage of critical staff positions filled at
milestone B is an easy and important metric to be observed. It
makes little difference to implement programs to raise the
level of skills of the program staff if they are missing in
action.
Second, the acquisition strategy document, that is to lay
out how the weapon system is to be acquired, the initial
roadmap, if you will, is often flawed in that it focuses more
on presenting a case for required capabilities and quantities
than on laying out the reasoning for acquisition competition
methodologies.
For example, how the prime contractor participants in an
MDAP competition will select subcontractors and how the winner
of the competition will manage the subcontractors to gain
improved efficiencies and effectiveness are generally given
little consideration. Creation of the acquisition strategy
document is one of, if not the most, important tasks the
government acquisition program management can undertake.
The strategy should establish the template for all the
activities that will take place throughout the source selection
process, engineering manufacturing development, and follow-on
production and fielding. More important, it establishes how the
program management team is thinking about the numerous events
and activities that a program will encounter.
The defense acquisition executive should establish a common
set of strategy elements that all the military department
service acquisition executives must include in MDAP acquisition
strategy documents. Additionally, a set of standards or metrics
by which the strategy elements can be evaluated as effective
must be part of that process.
In closing, I would be remiss if I didn't acknowledge the
progress that has been made by the Department in improving the
acquisition system over the last four years. Though it is the
General Accountability Office's headline that the 96 major
acquisition programs have grown in cost by $296 billion that
gets attention, those numbers belie an equally worthy but
overlooked statistic published in the very same report.
The average increase in unit cost of the 28 MDAP programs
with less than 5 years since development started is only 1
percent. Compared with an average of 55 percent increase in
acquisition unit cost of 25 programs in the group with 5 to 9
years since program development start, there has been important
improvement that should be recognized.
With that, it is my privilege to be here. Thank you very
much, and I would be happy to take any questions that you have.
[The prepared statement of Mr. Patterson can be found in
the Appendix on page 32.]
Mr. Andrews. Thank you, Mr. Patterson.
Mr. Fitch.
STATEMENT OF DAVID P. FITCH, USN (RETIRED), DIRECTOR, AT&L
LEADERSHIP LEARNING CENTER OF EXCELLENCE, DEFENSE ACQUISITION
UNIVERSITY
Mr. Fitch. Chairman Andrews, Congressman Coffman, members
of the panel, thank you for the opportunity to appear here
today. I will address the subject of acquisition performance
metrics and your questions about how to increase the realism of
program baselines, making trades between affordability and
performance and how to assess the value of systems that do not
necessarily--please recognize that my opinions do not
necessarily reflect the views of the Defense Acquisition
University (DAU), the Department of Defense or the
Administration.
Mr. Andrews. That is why we are interested.
Mr. Fitch. I suspected as much.
Measurement of acquisition performance must encompass both
strategic and tactical elements. As emphasized in a recent
Defense Science Board report titled, ``Creating a DOD Strategic
Acquisition Platform,'' the management, execution and oversight
of acquisition programs is moot if we aren't spending taxpayer
dollars to buy the right capabilities if we aren't
demonstrating strategic choice. It is as important to decide
what capabilities we won't buy as well as what we will buy.
I believe one of the root causes of funding instability is
what I described as too many programs chasing too few dollars.
Too many programs chasing too few dollars is one of the root
causes, I believe, of overly optimistic cost estimates. The
recently implemented Materiel Development Decision (MDD)
process provides a framework for making strategic investment
decisions.
The MDD is the formal point of inquiry into the acquisition
process. MDD will increase integration of the three major
acquisition support systems, requirements, resources, and
acquisition.
Improving requirements management, an initiative supported
by the Congress, includes providing training to requirements
writers and managers to ensure that they have a sufficient
understanding of the critical elements of acquisition, such as
systems engineering and testing. The goal is to improve
collaboration between acquisition and requirements communities
to exploit cost and performance trades and improve acquisition
outcomes.
Having a formal requirements and acquisition process is not
unique to the DOD. We can learn important lessons from
commercial industry. If you compare the DOD acquisition system
with the process to develop electronic games, there are marked
similarities and differences.
Notably, the process to get games on the shelves for the
December holiday season starts with a precise clarity of what
will be developed by when, and includes a corporate commitment
to the resources required for the project. Precise clarity is a
result of intense interaction between the people that divine
the capabilities of the game and the people that will design
and test the software.
Turning to the subject of tactical acquisition metrics, the
most effective tools and templates incorporate metrics, both
quantitative and qualitative. The question was raised, ``Are
there metrics beyond cost and schedule performance that are of
value?'' The answer is yes, and an ongoing example is a
probability of program success metrics, POPS, that are
currently being developed and deployed across the services and
other federal agencies, such as the Department of Homeland
Security.
Starting with a blank sheet of paper, a group of DAU
faculty comprised of experienced program managers and other
functional experts ask themselves questions, including, ``What
conditions facilitate the success of programs? What metrics are
leading indicators of derailment?''
The resulting tool, POPS, uses a structured process to
continually assess and display key elements of planning,
resourcing, execution and external influences that promote or
negatively impact program success. Still evolving, POPS is
being used in the Army, Air Force, Navy, Marine Corps and Coast
Guard. Timely, accurate and transparent metrics integrated in a
management and oversight process, will produce better program
outcomes.
Another question that has been asked, ``Can we ensure
improved realistic baselines.'' Again, I believe the answer is
yes, and there are ongoing initiatives that will yield more
realistic baselines. These include increased emphasis on
technology readiness before starting major development,
emphasis on improved cost estimating, which I am sure that we
will talk about today, and competitive prototyping, which is
included in the new DOD 5000.2 and pending legislation.
Prototyping increases the opportunity to identify and
assess affordability and capability traits. Competitive
prototyping also allows the government to observe the
performance of competing industry teams before making a down-
select for engineering and manufacturing development. No matter
how thoughtfully we plan and discipline source selection, a
paper-only source selection process is never as good as
demonstrated performance.
The ultimate assessment of whether we have delivered value
and needed capability to the warfighter is feedback from the
field from the warfighter. At various times, I have seen photos
of messages written with felt-tip pens on trucks and personnel
carriers that have been returned to the government or industry
depots. One of those signed messages--a picture is in my
written testimony--reads, ``This truck saved my life, as well
as five others, 2-April-2008 at 2300 Lima, Basra, Iraq.'' Those
kinds of testimonies, those kinds of results, are what those of
us that are in the acquisition aim to achieve on a daily basis
by our efforts and commitment.
Before equipment is fielded, it undergoes rigorous levels
of development and operational testing, and the new DOD 5000
has increased emphasis on earlier testing.
Mr. Chairman, thank you very much for the opportunity to
participate in today's discussion. As the Secretary of Defense
has said, there is no silver bullet. It will be a combination
of initiatives and collaboration amongst all the stakeholders
to create an acquisition system that consistently produces
successful acquisition programs.
I look forward to your questions.
[The prepared statement of Mr. Fitch can be found in the
Appendix on page 57.]
Mr. Andrews. Mr. Fitch, thank you very much for your
testimony.
Dr. Nussbaum.
STATEMENT OF DR. DANIEL A. NUSSBAUM, VISITING PROFESSOR,
DEPARTMENT OF OPERATIONS RESEARCH, NAVAL POSTGRADUATE SCHOOL
Mr. Nussbaum. Mr. Chairman, distinguished members of the
panel, I would like to thank you for this opportunity to
discuss my thoughts on how to improve acquisition and cost
estimating processes in the Department of Defense. These ideas
are mine alone.
As the chairman said, I am a member of the faculty at the
Naval Postgraduate School in Monterey, California. And I have
spent the last 30 years mainly doing, and more recently,
teaching and researching in the defense acquisition management
system with a focus on cost estimating. I was a previous
director of the Naval Center for Cost Analysis and past
president of the Society of Cost Estimating and Analysis.
All my experiences in cost estimating confirm that three
things are necessary for sound cost estimating: Acceptance of
the underlying uncertainties in predicting the future, accurate
and plentiful historical data, and professionally trained and
certified personnel.
On uncertainty, there is intrinsic uncertainty in all
estimates. It derives from several sources, mainly that we are
usually designing or building or operating something that is
substantively different from what we did before. The difference
could be in the product or in the economic conditions or the
programmatic conditions.
An estimate reflects our knowledge at a point in time when
we freeze the problem and base the cost on the configuration
and programmatics as they are understood at that time. From
that baseline, many things can change that can also change the
cost estimate, including labor rates, overhead rates,
schedules, enhancements of the capabilities or quantities,
changes when a particular technical solution to a problem
doesn't work as planned. And we need an alternative technical
solution.
On data, a hallmark, a necessary characteristic of a sound
cost estimate, is that it is based on historical program
performance from similar or related ongoing or past programs.
Historical data is variable. Not every aircraft costs the same
as every other aircraft.
And the measurement of this variability is accomplished
through statistical constructs, things like standard error of
the estimate, things like confidence intervals. We assume in
our community of cost estimating that the patterns of the past
will repeat in the future. But these patterns are almost always
statistically grounded patterns modeled with the powerful and
subtle techniques known collectively as regression. Further, we
know of no alternative approach to using the past as a guide to
the future if we want a scientific that is a reproducible and
auditable approach.
Not all estimating is done by government employees. There
really are three sub-communities. One is the government in-
house estimators. One is employees of the large vendors who
also design, develop and build what we buy--Boeing, Northrop,
Lockheed Martin, for example. And thirdly, there are support
contractors or consultants to the government. Those are the
three communities.
And surely, we need to increase the capacity and the
quality, the numbers and the training of the government
estimators, but so do we need to enhance the professionalism of
the other two communities.
There are currently no undergraduate curricula in cost
estimating, and there are only four educational institutions
that I am aware of that teach at least one course in cost
estimating, and those are the Naval Postgraduate School in
Monterey, where I am; the Air Force Institute of Technology at
Wright-Patterson in Ohio; Defense Acquisition University,
diverse locations with a capital campus at Fort Belvoir; and
the Massachusetts Institute of Technology, which offers one
elective course in cost estimating within its engineering
curriculum.
The recent separation of the business cost estimating and
financial management career field into two separate cost
estimating and financial management tracks is a very welcome
development and should be supported, but note that DAU support
is largely limited to military and DOD, not the other two
communities.
The Society of Cost Estimating and Analysis--we say SCEA--
whose membership includes approximately one-third of all cost
estimators supporting DOD, is a central and indispensable
player in the training, initial certification and periodic
recertification of cost estimators. I note with pleasure that
the executive director of SCEA, Mr. Elmer Clegg, is in the
room.
SCEA has collected a body of cost estimating knowledge, and
it provides the members of the cost estimating community,
provides training in cost estimating, has developed and offers
an examination and experience-based certification program.
Lockheed Martin, Boeing, Northrop Grumman, other vendors use
SCEA's training and certification as their standard.
I appreciate very much what this committee seeks to
accomplish, Mr. Chairman. This concludes my prepared statement,
and I would be pleased to answer your questions.
[The prepared statement of Dr. Nussbaum can be found in the
Appendix on page 66.]
Mr. Andrews. Well, thank you, Dr. Nussbaum, and I thank
each of the three witnesses for their testimony. As I say, we
have had the chance to review the written testimony. We are now
going to get to the questioning phase.
Mr. Patterson, I was intrigued by your reference in your
written testimony to two kinds of requirements costs, which you
express as customer requirements and derived requirements. What
is the difference between the two?
Mr. Patterson. Customer requirements are established in key
performance parameters, which are requirements that are
established for a particular weapons system that the weapons
system then must perform against. Derived requirements are
those requirements that the customer did not ask for but that,
as the name would suggest, derive as a consequence of the
design process. ``Oh, look, we could do this better if only
we''--and of course, that is taken to the program manager.
The program manager will say, ``Okay, we can do that. Just
bring money.'' And they do.
Mr. Andrews. Now, I notice on page seven of your testimony
that you say that a recent study prepared by the Monitor
Company Group was based on selected acquisition report data,
estimates that approximately 33 percent of the cost growth from
2000, 2007--I think I read this--is attributable to this second
category of requirements that you are talking about, right, the
derived?
Mr. Patterson. That is correct. Yes, sir.
Mr. Andrews. Now, how is that number reached? Where did
that 33 percent come from?
Mr. Patterson. That comes from the selected acquisition
reports (SARs). In fact, they lay that out in the SARs and
explain in each of the SARs where the----
Mr. Andrews. Well, one of the things that we want to do is
to make sure that we discriminate between derived changes that
are beneficial and those that may be superfluous. How would you
suggest that we might do that? In other words, I don't want to
leave the impression that we are saying, or the report is
saying, that ``Oh, that 33 percent was waste.''
Mr. Patterson. Oh, absolutely not.
Mr. Andrews. And how do we draw the line between
beneficial, cost-effective derived requirements and not-so-
beneficial derived requirements?
Mr. Patterson. Well, let me give you a real-life example.
In 1993, I came to the C-17 program. It was a troubled,
problem-plagued program.
The two program managers decided that much of the problem
was that requirements were growing in an airplane. They
established a set of rules. The set of rules was very simple:
If you have a requirement or an engineering change, which is
effectively a derived requirement, then it must go to an
engineering change board.
The engineering change board will evaluate the change for
its intrinsic merit. But if it doesn't meet the safety of
flight or other driven requirements, then it has to have a
three-to-one payback in savings and not perturbate the
schedule.
Mr. Andrews. Interesting.
Mr. Patterson. Real simple.
Mr. Andrews. Now, in terms of the scope of that C-17
program, how much were the derived requirements overrun? In
other words, how much was attributable to derived requirements?
Mr. Patterson. I would only be guessing. I can certainly
get that for the record.
[The information referred to can be found in the Appendix
on page 83.]
Mr. Andrews. And with this method that you just described,
in your judgment, what percent of the derived requirements were
beneficial, and which failed to make that three-to-one cut and
didn't happen?
Mr. Patterson. Again, let me give you an example.
At Edwards Air Force Base, they were testing the airplane.
One of the requirements was that one airplane needed to start
another airplane, but it started it with two hoses from the
pneumatic system. Somebody said, ``I wonder if it will start
with one hose.'' It didn't. ``Oh, my gosh. Well, now we have to
go and figure out why that is the case.''
And so, that took a considerable amount of test time and
money, and as it turned out, they really wanted to have it
start with one hose. I am sorry, that would have taken----
Mr. Andrews. Who was the someone who said that? Not the
person's name, but where were they in----
Mr. Patterson. That was the test community out on the ramp
at Edwards.
Mr. Andrews. Right.
I am going to go back for a second round, but Mr. Fitch
mentioned this probability of success metrics program. Has
there been built into that a litmus test for a probability, if
you would fall below it, things stop? In other words, does that
program have in it a built-in go or no-go line?
Mr. Fitch. It does not have a go or no-go line. It is
information that is updated and provided----
Mr. Andrews. Do you think it should?
Mr. Fitch. I think that there needs to be an informed
review and decisions that look at each of the things that occur
that is negative, because some of the things that are in the
probability of success metrics and reporting are positive.
So I think that it needs to have the program manager, when
that information is put together on a monthly basis, needs to
look at that. It needs to be reviewed by someone in his Program
Executive Office (PEO), okay, and potentially even at higher
authority when there are negative occurrences.
Could I add something about--reporting?
Mr. Andrews. Sure, and then we are going to go to Mr.
Coffman. Sure.
Mr. Fitch. Yes.
I just wanted to explain that the derived requirement
process is actually part of the systems engineering process so
that if, for instance, the warfighter says in an aircraft, ``I
need a display that has color, I need a display that has this
amount of resolution,'' et cetera, the first pass, even by the
warfighter or the cost estimators, may say, ``This amount of
processing may be sufficient for that.''
When you get into, ``Well, by the way, you are going to
have these other software capabilities.'' You put those
together, and they start to build on one another, you can find
all of a sudden that the processor that was planned may be a
commercial off-the-shelf item with X amount of memory,
throughput, et cetera, is insufficient.
At that point is the point where it would be very useful to
have the requirements community have a real dialogue with the
acquisition community to say do I now take away some of the
requirements so I can continue to use that commercial off-the-
shelf processor that is less expensive, or are those
requirements really important now that we have figured out what
they are. So the point is the derived requirements is a very
important process, and it is actually the process where you
allocate where features will go to hardware or software.
Mr. Andrews. Thank you very much.
Mr. Coffman.
Mr. Coffman. Thank you, Mr. Chairman.
This is to all of you. It seems like part of the problem is
sometimes we are dealing with immature technologies, and I
think we have had previous testimony to that effect, that
sometimes we are asking the contractors to develop something.
Should we bifurcate the process? In other words, that you
are contracting with one entity to develop the technology, to
develop the--I don't know if you would go as far as a
prototype, and then where you can go into fixed cost production
with another--maybe that entity would be allowed to bid on it
as well. But is it better to bifurcate the process?
Mr. Nussbaum. You know, I would say we would do that now.
If we have open competition, we certainly have a research and
development (R&D) contract followed by an acquisition contract
with--there is certainly no guarantee that the R&D contractor
will have a follow on.
So in some sense, we do that, but in another sense, the
Department has now mandated technology readiness levels, TRLs
of certain levels before a program can get beyond a milestone.
They have narrowed that cone of uncertainty by saying you have
to have a TRL level six. They would prefer seven, and GAO
prefers seven, but you have to pass a TRL level of six before
you go into the next milestone.
Mr. Fitch. As a part of the analysis of alternatives, that
process that immediately follows it, the technologies that
would be appropriate for that system, that capability, are
assessed for technology readiness, when you get to the point of
saying who should develop it, I prefer the concept of the
competitive prototyping, because you are going to get proposals
from industry.
They are not going to all have the same strengths. Some are
going to view that I want to do this more in a hardware
function. Others may be more software intensive. There are
different technologies involved.
Their proposal, and then watching them deliver on what they
promise and they say is possible in competitive prototyping, is
key to then being able to transition and make an award to that
contractor who has proven that they did what they say they
could do. So having a different contractor, if I understood
your premise correctly, to develop technology and then award it
to somebody differently to develop the system, I think that is
not the intent of the competitive prototyping system.
Mr. Patterson. I think, too, there is a problem that arises
when we talk about spiral development. Originally, spiral
development, evolutionary development, were designed to have a
weapons system that effectively was fielded and then have block
upgrades to improve the weapons system.
In many cases, that is not how that works, because you do
have a parallel R&D program that is working on improvements.
And before the weapon is actually fielded, you have the attempt
to integrate improvements.
Well, the consequence of that often is that you have a
stretched-out program. Costs escalate. And in the end,
performance is degraded, and it takes a while for the--much
longer than had been anticipated for that weapons system to be
fielded. So I think we need to have very specific and clear
understood standards by which we will integrate or put in
upgrades or technology advances into weapons system. It
certainly can't be during engineering, manufacturing,
development.
Mr. Coffman. I think it was mentioned, and it has been
mentioned repeatedly, that the changes drive a lot of the cost
as they go forward. Where did the changes emanate from? I mean,
are some of the changes sort of broader in scope where the
affected branch who will receive and utilize the weapons
system, is it changes in evolving doctrine, or is it mostly at
a very technical level with this engineering requirements?
Mr. Nussbaum. You know, I think that changes come from all
sorts of places, including changes in labor, overhead rates,
schedules, requirements, quantity, absolutely everything, just
like building a house. Everything that changes has the
contractor saying, ``Cement has gone up. Brick has gone up. My
subs have gone up''--or down, but they are always changing.
And so there is a great churn. Some of it is part of life,
and some of it we try to control by saying, ``Tell us what you
are building, and we will cost that program.'' But I think it
is just intrinsic in the cost estimating process.
Mr. Coffman. Is there any way to just have more--I mean,
what would be some of the methods for having more discipline
over the--and I take it some of it would--having a change
board?
Mr. Nussbaum. There is something called a CARD, a Cost
Analysis Requirement Document, which is, to the chagrin of all
program managers, we say three or six months before we are
going to a milestone, you tell us what you are building, how
many, what the technologies are, and we will cost that program.
It may be that things change after that, but that is the
program we are going to cost, because otherwise we are chasing
our tail, and we don't know until the very last moment what we
are costing.
And in fact, it takes time to do a cost estimate. It is a
part of the discipline at the Office of the Secretary of
Defense (OSD) is a Cost Analysis Requirements Document will be
prepared six months before the milestone, and the lack of
availability of that CARD results in a day-to-day slip in the
milestone. So that is real discipline, but it has its obvious
down side, too.
Mr. Patterson. I think, too, that in dealing with the
discipline and structure, that I must tell you, I submit that
it requires rules. And the rules simply have to say that, after
milestone B, you don't have any more requirements. And, I mean,
it is somewhat draconian, but nonetheless, if it--unless of
course it is a safety of flight or it is an obvious design
failure that needs to be corrected.
But those kinds of things are few and far between, quite
frankly. And then, this idea that a requirement or a change
would give you a three to one or four to one payback in savings
while not perturbing the schedule is not a bad idea, either.
And as technology moves forward, that is entirely possible over
the course of a program.
Mr. Fitch. There are times with the requirements--I recall
when I was developing a black box that was going to go in 17
different platforms. I would have this platform coming to me
and say, ``It would really be easier if you changed it this
way. It will make me less expensive for my integration, save my
costs and schedule.''
What I came to understand is the program manager has to be
prepared to say no during the development phase. Now, the
configuration steering boards we have in place anticipates that
there may be compelling reasons to change a requirement,
moderate a requirement during development, but that is the
process of the configuration steering board, is to raise that,
if you would, visibility to the pressures of changing
requirements to have a senior level decision made about the
requirement once the milestone has been approved.
Mr. Andrews. Thank you.
Mr. Cooper.
Mr. Cooper. Thank you, Mr. Chairman.
And I would especially like to welcome a fellow Tennessean,
David Patterson, and congratulate him on starting Defense
Business Institute at the great University of Tennessee in
Knoxville. Thank you not only for your past service, but what
you are doing right now.
Can anyone provide me with enough historical perspective to
help me understand how, during World War II, men like Henry
Kaiser were able to produce destroyers and--I forget exactly
who produced airplanes, but it was an amazingly productive
period. And I don't know how--was there less bureaucracy then?
How was so much able to be accomplished so quickly? Anybody
know?
Mr. Nussbaum. My reading of history is that there were a
lot of failures, and then we tend to remember the successes,
which were terrific. But in fact, there were false starts.
The P-51 was wonderfully successful, but because it had a
bigger gas tank so it could keep up with the other aircraft, it
gave us greater range, but we didn't build it for that reason.
Just we sort of lucked into that, if you will.
I think that things were much simpler. We were able to turn
technology generations around much faster and, therefore,
absorb the failures. Today's systems are very complex. They
take a long time, and we just won't accept failure.
Mr. Fitch. I think that what we remember a lot is those
ships going down the ways, okay, daily, okay? That was in the
production phase.
Even the P-51 that was just mentioned by Mr. Nussbaum, when
it was first fielded, it didn't have the engine that ended up
being what everybody remembers about the performance of that
aircraft. There were various modifications made after it was
fielded. It was in the field and found that it was not exactly
was needed.
Mr. Patterson. I think, too, you had a tremendous
industrial base that was able to accommodate to the level of
technology that it was asked to accommodate to. Today, I am not
sure that we would be able to do that again in that amount of
time.
Dr. Ron Sega did a study not long ago when he was at DDR&E,
Director of Defense Research and Engineering, in which he
looked at the industrial base and what it could do, and found
that 62 percent of all of the Ph.D candidates that are enrolled
in disciplines critical to national security have temporary
visas. And that tends to put us at a disadvantage, because they
generally don't stay around and work for Lockheed or Northrop
or Boeing, and Skunk Works, particularly, or Phantom Works.
And if you will look at the Aerospace Industries
Association data, you will find that in the 1990, 1991, you had
1.3 million touch labor workers, highly skilled workers. Today,
we have something less than 700,000. Those are statistics that
should give us pause to consider what we are going to do in the
future.
Mr. Cooper. Help me understand. I know that we have had
failures in acquisition throughout our history, and Mr. Fitch
mentioned a success when he posted the note that was on the
vehicle that saved lives.
But on the front page of the Washington Post recently was a
statement by Secretary of Defense Gates when he attended the
return of the remains of some of our troops. He asked how they
died, and he was told they were in an inadequate vehicle. And
he cursed because that was symbolic of the fact that we have
had difficulty fielding relatively simple platforms like Mine
Resistant Ambush Protected vehicles (MRAPs) or up-armored
Humvees.
And at least for some period of time there was only one
manufacturer of up-armored Humvees in America, and yet, at the
same time, we have automobile companies going bankrupt and
looking for vehicles to build. So there seems to be a mismatch
somehow between pretty basic demands for troops and our ability
to field and source those even when we have able and willing
automobile companies who are looking for ways to keep their
plants busy. Why the mismatch?
Mr. Patterson. Oh, I think it is a very fundamental problem
in that you have a Department of Defense that understands fully
that we are at war with terrorists, and you have a country that
doesn't.
Mr. Cooper. So Chrysler or General Motors (GM) or Ford
didn't want to bid on the up-armored Humvee, or we couldn't----
Mr. Patterson. Well, actually, the subsidiaries of those
folks did, but they really were not in the business of putting
out those kinds of vehicles. The people who were, International
Harvester, obviously, and--but the design, fortunately or
unfortunately for us, that was available were foreign designs
of up-armored vehicles that were designed to sustain the kinds
of things that an MRAP would have to sustain.
Mr. Fitch. And the number of companies that have as a core
competency in the technologies for armor, okay, is not the same
as we have for the auto industry. In fact, we are doing a lot
of investment in the Department of Defense today to try to find
better armor, cheaper armor, especially lighter armor, whether
that be for the vehicles because, when you put the armor on it,
it puts a demand on the engine. It puts a demand on the
drivetrain. If you double the weight of the vehicle, whatever
the percentage is, okay, it has additional tolls upon the
reliability of the vehicle.
So we don't have the same industrial base. In other words,
the auto industry isn't the industrial base for our Army,
either.
Mr. Nussbaum. In a sense, the Humvee is the wrong vehicle
to up-armor, but it is the only vehicle to up-armor. But it was
built as a replacement for the Jeep, which didn't go in harm's
way.
So it was designed to optimize all the functions that it
was going to perform, and we missed the fact that it was going
to go in harm's way. So it wasn't suitable, wasn't optimal, for
up-armoring, but it is what we have, so we are going to up-
armor it.
Mr. Cooper. I see that my time has expired. Mr. Chairman,
look forward to another round.
Mr. Andrews. Thank you.
With the consent of my colleagues, we are going to do
another round, if that is okay with the panel as well, if it
fits your schedule. Thank you.
Mr. Patterson just made a suggestion that we might have a
rule that says no new requirements after milestone B, and that
prompted this question. Again, this piece of data that you
cite, about 33 percent of the program cost growth being
attributable to these requirements changes, how many of those
requirements changes happened after milestone B? Do you know?
Mr. Patterson. That data, as I recall, is after milestone
B.
Mr. Andrews. It is all after milestone B.
Mr. Patterson. That is where the majority of the growth
generally takes place.
Mr. Andrews. A related question: in your oral testimony,
you talk about the fact that, if I read this correctly, the
average increase in unit cost of the 28 MDAP programs of less
than 5 years since development is only 1 percent.
Mr. Patterson. That is correct. That is----
Mr. Andrews. But it is 55 percent from years 5 to 9. How
many of that in the 55 percent category, the five to nine, got
a waiver through milestone B, didn't meet the requirements to
get to milestone B but got waived past it? Do you know?
Mr. Patterson. No, I don't. I don't know exactly how many.
Mr. Andrews. But, I mean, would it be accurate to say it is
probably most of the 55 percent cost overrun comes from that?
Mr. Patterson. I would say a significant portion of it,
sure.
Mr. Andrews. One of the things that is in the conference
report that we will be looking at in the WASTE TKO bill today
is what we call intensive care, where if a program is permitted
to go forward, even though it didn't achieve the milestone B
criteria, if it is waived past it, there is a whole set of
intensive requirements that are imposed upon that program to
try to get it back under control.
I wanted to come back to, again, this bifurcation that you
create between derived and customer requirements. Describe for
us the process that you think ought to be instituted to
determine whether a derived requirement is added to the package
or not.
Let's say we are at a point--assuming for a moment that we
accept your proposition that there are none after milestone B,
which I assume you say there should be some exceptions, now as
you said, for true safety or emergency purposes, but let's
assume we are living in a world where, except for those narrow
situations, there are going to be no changes in requirements
after we hit milestone B.
We are now in pre-milestone B, and an ``Oh, by the way''
comes up, as you said earlier, ``Oh, by the way, this can do
this.'' Who should make the decision as to whether that gets
added to the package, and by what criteria?
Mr. Patterson. Well, I think that the program manager
should have the initial cut at whether or not they are going to
include that into the program. But program managers generally
are colonels or, in very large programs, brigadier generals who
have significant oversight within the Department.
I think that if they have a set of rules that say, first of
all, if it is not a safety of flight or if it is not some sort
of safety issue, or if it doesn't give me a return on my
investment, then I am going to have a thumbs-down, initially.
Mr. Andrews. How do you measure the concept of return?
Mr. Patterson. Well, let's take, for example, you have--
again, I will turn to this C-17.
There were parts of that airplane that were originally
designed for aluminum lithium, for example. Well, aluminum
lithium is strong, but it is brittle.
So an engineering change was made to change that to a
different alloy. We were breaking the aluminum lithium cargo
floor guides at a regular pace, so changing that eliminated the
problem of having to constantly replace it. That was a savings.
And those are the kinds of things that I would suggest are
three-to-one. And even the suppliers were given the opportunity
to do that.
Mr. Andrews. Got you.
It strikes me that this is really the essence of the 20
percent the secretary talks about in his 80 percent solution,
that what he really is aiming to get at here is to give us an
adjudicatory mechanism that draws the line between the 80 and
the 20 when you get to this point, and it is your suggestion
the program manager should be the first person to weigh in on
this.
Who should evaluate his or her recommendation?
Mr. Patterson. Well, then you have an engineering change
board that would provide a corporate view of it. And if it is a
particularly expensive change, then you are going to go to your
service acquisition executive, or if it is an ACAT 1-D, an
acquisition category 1-D, then you have the Defense Department
acquisition executive who would have a cut at that.
Mr. Andrews. And my final question, are you confident that
we can quantify this concept of value sufficiently to hit the
three or four? In other words, are all of the values that we
want to promote--you gave a great example of saving, replacing
a piece of a plane that is going to go wrong. But do all of the
value concepts lend themselves to that kind of quantification
that would let us say, ``Well, this fails to meet three-to-one,
so out?'' Pretty hard?
Mr. Patterson. No, it is much more difficult than that. And
that is why it takes a lot of research and study and to set up
standards and conditions whereby you can evaluate these.
Mr. Andrews. Frankly, and I will just conclude with this,
it is one of the reasons why we are glad we have the three of
you and the institutions that you represent, because we really
do turn to institutions like yours to assemble those data,
analyze them and give us a factual basis to draw the lines that
my questions imply. Thank you.
Mr. Coffman, your turn.
Mr. Nussbaum. Mr. Chairman, is it----
Mr. Andrews. Sure, Doctor.
Mr. Nussbaum. Is it appropriate for me to make a remark,
or----
Mr. Andrews. Sure. It is okay.
Mr. Nussbaum. I think it is not hard to measure the value
of ideas that replace current capabilities. It is always harder
to measure the value of things that represent new capabilities.
They don't represent a savings for the operating and support
detail that you were going to incur.
But if you are replacing a current capability, then it is
pretty easy to do an estimate, but it is still an estimate, of
what does it cost to invest to make this thing happen, and what
do you save over time in the operating----
Mr. Andrews. I agree. The much more difficult proposition
is where you have a new function that could be added by
something that you discover. How do you measure that, and that
requires trade-off analysis. It requires opportunity cost
analysis. It requires a lot of broader inquiries.
Mr. Nussbaum. And if you have that long tail, the question
then is do you do any discounting on it, the technical issue of
net present value and at what rates. The Office of Management
and Budget (OMB) helps us there.
Mr. Andrews. Not to be hyper-technical, but one of our
problems is then matching up the federal credit scoring and net
present value rules with the real-world ones, that very often a
decision--a classic example is in energy. The Department has
guidance to hit 25 percent alternative fuels by 2020. And in
order for them to do that, they need to do multi-year
contracts. But to do multi-year contracts, the Congressional
Budget Office (CBO) scores that as putting the whole net
present value into one year, which makes it almost impossible
to do, which means we don't do much of it, which means we are
falling backwards. So marrying the CBO criteria with the real-
world criteria is a bit of a challenge, too.
Mr. Nussbaum. Yes, sir. And so the devil really is in the
details on this, and----
Mr. Andrews. Well, the devil is in the CBO in this case.
Now, don't tell Mr. Elmendorf I said that, okay?
Mr. Nussbaum. And one other comment, and that is, when you
go to the Configuration Control Board or your Service
Acquisition Executive (SAE), you are proposing to spend
investment dollars to make this thing happen, and the promise
is that you will return operation and support (O&S) dollars
later on. That is a nice conversation, but it doesn't accord
with the budgeting realities.
Mr. Andrews. And it doesn't score in our--nor should it.
Mr. Fitch, do you want to add one thing, then I am going to
go to my friend from Colorado?
Mr. Fitch. If I could. I just wanted to say, again, that I
think it is useful to talk about the operation requirements,
the derived requirements, but to say again that, when we get
the requirement from the user, it is stated usually in
operational terms, okay, the results they wanted to see.
Mr. Andrews. Yes.
Mr. Fitch. To deal with industry, and for industry to build
something, those need to be translated into technical terms. We
also use the term ``derived requirements'' for that process. So
the derived requirements are really--there are a couple types
we are talking about here, that which are a part of the normal
system that you have to do.
The other one I would just say is your question is how do
we know we get value. As a program manager, I had an
acquisition program baseline. That was my contract. The way I
viewed it, it was my contract with my--decision authority and
with the American taxpayer to produce a capability at such a
cost with certain milestones.
And I think that is what most of us take and go back to.
And as we do these questions about derived requirements and
everything else, I think we keep in mind that framework.
Mr. Andrews. Yes. I think the panel clearly understands
that some subset of derived requirements are quite legitimate,
necessary and desirable. And I think the three of you have
given us some interesting tools to discriminate between
undesirable derived requirements and desirable ones, which is
what we are about.
Mr. Coffman.
Mr. Coffman. Thank you, Mr. Chairman.
We talked today about having discipline in terms of change
requirements. But do we also have to have discipline when
programs get so far out of line that they become questionable?
And I am thinking about--and I am working if you could
reflect on the President's--I can't remember the nomenclature
of the follow-on helicopter. Littoral Combat Ship (LCS) is in
question. I wonder if you all could reflect on when a program
gets so far out of line.
Mr. Patterson. Well, it would be helpful, I think, in the
total scheme of things, if we never let them get out of line.
But nonetheless, you are exactly right. They do happen. It does
happen.
And one of the things that I think is difficult when you
take, for example, the VH-71, the President's helicopter, and
that is a perfect example of where you came in with one set of
criteria as requirements and, over the course of the time, it
changed dramatically.
I think that we don't have a set of standards or conditions
that raise a budget flag that say, ``Wait a second, I am sorry,
you are red here, and you have been red three reviews in a row,
and we are canceling your program.'' And what we do is we put
ourselves in a position where it is the only game in town.
You have the President's helicopter that is arguably old,
and we don't have an alternative. You chose a manufacturer and
a helicopter, and there is no off-ramp. There is no plan B. And
we do this rather consistently.
And I am almost of the opinion that we do it by design. And
I would offer the Marines' Expeditionary Fighting Vehicle
(EFV), for example. We don't have a way to walk the dog back
down the path to get an amphibious vehicle that would replace
the EFV, so we need to make that work.
Mr. Fitch. I don't have the particulars on either of the
programs you asked about. I was the systems engineer for the
VH-3D, the current--well, it is one of the two current
presidential helicopters. It is a unique mission. I think there
is a desire usually if the White House says, or whatever or
whoever it is, says, ``I need a capability,'' that you find a
way to do it.
Going back to it, most of the changes that I saw, and the
pressures and the surprises that I got, occur in that first--
traditionally, the first 12 to 15 months of a program when you
go towards a Preliminary Design Review (PDR), because the
contractor is off doing a lot of things. There isn't a lot of
object deliverables to figure out does he get it, he doesn't
get it.
I think that the competitive prototyping system to focus on
doing PDRs earlier, to have effective communications with the
contractor teams and oversight of the contractor teams during
that period of time will do much more to get to a stable
baseline at milestone B, which is the actual program
initiation.
Mr. Nussbaum. I am going to sound like a professor on one
hand. On the other hand, because there are some historical
examples of things which failed, failed, failed and then were
terrific. Aegis was one, and Tomahawk was another. They just
took a long time to bring aboard. And for some reason, we had
the fortitude to stick with them and not say, ``This is an A-12
or a Gama Goat,'' and get rid of it.
So now we come to the category of LCS. Is it a Tomahawk or
is it a Gama Goat? We don't know. But I am taken with my
colleague's remark that, once you get past milestone B, if you
have three reds in a row, you have got some serious explaining
to do, with the presumptive answer, ``You are out.'' It is a
rule.
But the problem is knowing the future, and that is always
the problem. I don't know whether LCS is a Tomahawk which is
going to be absolutely terrific after a long incubation period,
and the same for V-22. You just don't know, so we make
decisions as people.
Mr. Coffman. Thank you.
Thank you, Mr. Chairman.
Mr. Andrews. Thank you.
Well, gentlemen, thank you very much for your expertise.
Your reward for doing such a good job is we will have to call
on you again. As the committee goes forward, our intention is
to try to make legislative proposals for the fiscal year 2011
authorization bill that will deal with the area of the problem
that the WASTE TKO legislation that we are dealing with today
does not deal with.
And I think you have given us some very intriguing ways to
measure the gap between what we pay and what we get. It is also
heartening to hear what I have heard this morning, a consensus
that this panel's contribution to the WASTE TKO bill was
essentially two concepts.
The first was to add a whole series of reviews and a lot of
scrutiny pre-milestone B, with particular emphasis on the
requirements process. And that did make it into the conference
report, and that will become the law this week, we think.
And second, the panel was very interested in much more
rigorous review, what we call intensive care, of systems that
pass milestone B by waiver, that have not met the requirements,
or that fail Nunn-McCurdy standards and get exempted from the
penalties there and go forward anyway. And I think if you look
at the cost overruns, a huge majority of them fall into one of
those two categories.
So what we wanted to do was to take the best practices that
you very ably described this morning and engage them as
intensely as we can in the systems that, again, never met the
criteria to get past milestone B but get past it anyway, and
those that fail Nunn-McCurdy but continue to live on.
And as Mr. Patterson said a few minutes ago--I think it was
Mr. Patterson--our ultimate goal is not to have any of those
cases in the future by unraveling the requirements process and
looking at it more intensely to intensify that pre-milestone B
analysis of what is going on.
And the other point that I would make that is more on our
side of the table than yours, I think that the principal reason
that we get these cost overruns is that, once something passes
milestone B, an enormous political constituency develops around
it. Now, there are tens of thousands of people deriving their
paycheck from a project, hundreds or thousands of
subcontractors, dozens or hundreds of congressional districts.
And as Secretary Gates I think can attest, making changes
in those programs is very politically difficult. If you get to
these flawed programs earlier when their political
constituencies are smaller and weaker, the chance to do the
right thing is a lot higher.
So not just for analytical reasons, but we think, given the
dynamic of the way these decisions are made in the political
world, the more precise we are in our measurements and the more
focused we are in our evaluation, in the requirements phase and
a little bit beyond that, we think the better job that we will
do.
So I would say to each of the three of you we welcome your
continued participation and input. We are certainly going to
call upon you for your feedback as we go forward in our
drafting process. And thank you very, very much for your time
and attention this morning.
Members will have a period of time by contacting either
majority or minority staff to supplement the record with
written questions, and we would invite the witnesses to do the
same thing.
And with that, the hearing is adjourned.
[Whereupon, at 9:09 a.m., the panel was adjourned.]
=======================================================================
A P P E N D I X
May 19, 2009
=======================================================================
PREPARED STATEMENTS SUBMITTED FOR THE RECORD
May 19, 2009
=======================================================================
[GRAPHIC(S) NOT AVAILABLE IN TIFF FORMAT]
=======================================================================
WITNESS RESPONSES TO QUESTIONS ASKED DURING
THE HEARING
May 19, 2009
=======================================================================
RESPONSE TO QUESTION SUBMITTED BY MR. ANDREWS
Colonel Patterson. To appreciate the context in which early C-17
cost overruns occurred, some background and explanation of a
methodology for arriving at an estimate of how much of the cost
overruns could be attributed to derived requirements is in order. In
this case engineering changes were the predominant example of what I
have described as derived requirements and that appear in the public
record.
The first manufactured part for the first C-17 was milled in
November of 1987 and the first squadron was declared to have Initial
Operating Capability in January 1995. The period between these events
was one of significant turbulence for the C-17 program. The contract
for the C-17 program was a fixed-price, incentive fee development
contract, because the aircraft was to be designed using off-the-shelf
technology and was determined to be ``low'' risk.
At the time of contract award, both the Government and MDC
[McDonnell Douglas Corporation] envisioned a program based on
commercial practices, minimum Government involvement and
concurrent development/production effort. Consequently, a fixed
price incentive contract was used to match the perceived low
risk of the program.\1\
---------------------------------------------------------------------------
\1\ Report of the Defense Science Board Task Force on C-17 Review,
December 1993, p 3
As the program became more complex and engineering changes more
frequent, MDC had no recourse to recoup investments in solving the
engineering problems and meeting the cost of addressing requirements
that were being added during full scale engineering and development
than to make claims to the Government against the contract.\2\ Though
it is not feasible to determine precisely how much of the total cost
overrun was attributable to derived requirements or engineering
changes, it is possible to understand the relationship between the
unanticipated cascade of engineering changes and increased requirements
and cost overruns by virtue of the amount of the claim MDC believed
justifiable.
---------------------------------------------------------------------------
\2\ Ibid.
---------------------------------------------------------------------------
COST OVERRUNS AND ENGINEERING CHANGES
As early as 1985 engineering changes were impacting the C-17
program costs. The Selected Acquisition Report for December 1985
described cost increase of $214.5M for engineering changes ``needed for
a four-pallet ramp, a combat offload rail system, and DoD standard
avionics racks.'' \3\ Though these modifications were not in 1985
referred to as ``derived requirements,'' they do meet the definition
used in the study included with the oral statement provided for the
record on May 19, 2009.\4\ These were the typical derived requirements
not in the original contract, but identified by the program management
to be necessary during development and presented as engineering
changes. These changes and the corresponding cost increase occurred in
the very early stages of the program, In December of 1985 the Full
Scale Development contract had been signed.\5\ The Government became
alarmed following a 1992 Department of Defense Inspector General Report
that identified a cost overrun of $700M on a $6.6B contract ceiling.\6\
McDonnell Douglas was paying for costs overruns on the fixed-price
contract and submitted claims for approximately $300M for what MDC
asserted were changes in scope of its initial contract agreement which
included ``costs for engineering, development, testing and production
of six C-17's and three test planes.'' \7\ In the end, after
significant negotiations with the Department of Defense, McDonnell
Douglas agreed to spend $456M in process improvements and the Defense
Department agreed to ``provide an additional $438M for the program.''
---------------------------------------------------------------------------
\3\ Selected Acquisition Report As of December 1985, p 10
\4\ Derived requirements, on the other hand, are requirements that
the customer has not specified directly as a requirement but that
emerge or derive from the design decisions that are made. Derived
requirements are not capabilities that the customer specifically has
identified. Particularly troublesome is a subset of derived
requirements that fall into the category of engineering changes--those
changes that improve on ``good enough'' and that have a combined effect
of driving up costs and missing schedule milestones.
\5\ Aeroflight, Aircraft of the World, Boeing C-17 Globemaster III,
retrieved May 27, 2009, from http://www.aeroflight.co.uk/types/usa/
boeing/c-17/c-17.htm
\6\ Battershell, A. L., The DoD C-17 versus the Boeing 777, A
Comparison of Acquisition and Development, National Defense University,
Washington, D.C. 1999, p 87-88
\7\ Adelson, A., Company News; McDonnell May Submit Big C-17 Bill,
New York Times, January 14, 1993, retrieved May 27, 2009, from http://
www.nytimes.com/1993/01/14/business/company-news-mcdonnell-may-submit-
big-c-17-bill.html
---------------------------------------------------------------------------
A significant portion of the cost problems centered on a cascade of
engineering changes that typified the years and immediate several
months leading up to the negotiated settlement between MDC and the
Department of Defense. The magnitude of the engineering changes
occurring during this period was described clearly in the December 1993
Defense Science Board's report as it stated, ``In May 1993 alone, over
1,000 changes were issued directly impacting production. The backlog of
engineering changes at the end of May 1993 reveals 5,800 open work
authorizations.'' \8\ Additionally, costly changes to the program were
government-imposed. A typical example is the C-17 test program which
was increased from an 80 aircraft month test program to a ``rebaselined
test program of approximately 152 aircraft months'' at the
recommendation of the Defense Science Board study, with cost split
evenly between the Government and MDC.\9\
---------------------------------------------------------------------------
\8\ Report of the Defense Science Board Task Force on C-17 Review,
December 1993, p 3
\9\ Ibid. p 12
---------------------------------------------------------------------------
SOME EDUCATED CONCLUSIONS
With the data that is immediately available from the historical
record it is reasonable to conclude that the engineering changes were
significant and made an impact on cost, schedule and performance on the
C-17 program. Referring again to Battershell, of the $7.3B for the
development costs and the first six aircraft in production Lots I and
II, MDC invested approximately $1.7B of its own funds.\10\ This amounts
to about 26 percent of the original contract amount of $6.6B. MDC
proposed a claim against the government of $1.2B in addition to the
$438M it had received in the settlement \11\ intended to recover what
it believed to be legitimate costs associated with program changes a
significant number of which were engineering changes. Though the term
``significant'' used above is not a precise accounting regarding ``how
much'' of the program was ``attributable'' to derived requirements, it
does provide a qualitative data point that is important in evaluating
opportunities to improve acquisition programs in the future. [See page
11.]
---------------------------------------------------------------------------
\10\ Battershell, p 90
\11\ Government Accountability Office Report, (GAO/NDIAD-94-141),
Military Airlift C-17 Settlement Is Not a Good Deal, April, 1994, p 6
---------------------------------------------------------------------------
NEWSLETTER
|
Join the GlobalSecurity.org mailing list
|
|