SlideShare a Scribd company logo
1 of 8
Download to read offline
A New Approach to
Generic Attributes*
Curt Wells,1, †
Linda Ibrahim,2
and Larry LaBruyere3
1
I-metrics LLC, 90 S. Camino Real, Uhland, TX 78640
2
Federal Aviation Administration, 800 Independence Avenue SW, Washington, DC 20591
3
Northrop Grumman Mission Systems, 475 School Street SW, Washington, DC 20024
A NEW APPROACH TO GENERIC ATTRIBUTES
Received 10 January 2003; Accepted 20 June 2003
DOI 10.1002/sys.10050
ABSTRACT
Generic Attributes (GAs) are measures of process performance introduced by the systems
engineering capability model, EIA/IS 731. The systems model defines two GAs—Effectiveness
and Value. While the concept of GAs is generally accepted as valid, they have been used very
little, due to difficulties in interpretation and appraisal. Revised definitions and a new
approach for appraising and using GAs have been developed for the Federal Aviation
Administration Integrated Capability Maturity Model®
v2.0 (FAA-iCMM®‡
). The improved
approach to Generic Attributes provides the needed clarity and measurement objectivity.
Definitions of the iCMM®
GAs and their relationship to process model concepts are described
in this paper, along with a practical approach to appraising GAs. © 2003 Wiley Periodicals, Inc.§
Syst Eng 6: 301–308, 2003
Key words: process appraisal effectiveness, generic attributes, Integrated Capability Maturity
Model®
1. INTRODUCTION
The Need for Generic Attributes.Although capability
and maturity levels indicate the level of process disci-
pline that organizational units may achieve in one or
more process areas, the levels do not necessarily trans-
late into the quality of work products or quality/value
of results in meeting objectives. It is possible to have
mature processes (e.g., capability level 3, 4, or 5) that
produce unsatisfactory work products (e.g., of less than
desirable quality, or that do not meet needs, for greater
Correspondence
*Measurement Category Best Paper Award, 2002 INCOSE Sympo-
sium (updated for this article).
†
Author to whom all correspondence should be addressed (e-mail:
cwells@i-metrics.net).
‡
CMM and Capability Maturity Model are registered in the U.S.
Patent and Trademark Office.
Systems Engineering, Vol. 6, No. 4, 2003
© 2003WileyPeriodicals,Inc.§
This article is a US Government work
and, as such, is in the public domain in the United States of America.
301
than expected expense). In such cases, enhanced proc-
ess capability may not result in the actual performance
improvements that organizations expect from pursuing
process improvement. Measurements of the quality or
value of process results (i.e., GA-like measurements)
have not been used heretofore, to any significant extent,
largely because of the difficulty and lack of objective
methods to do so.
1.1. GAs Introduced by EIA/IS 731
These concerns led the authors of EIA/IS 731 [EIA,
1999] to define GAs and incorporate them into the
appraisal process. While the need for GAs (as defined
by EIA/IS 731) was widely acknowledged by the proc-
ess improvement and appraisal community, operational
difficulties have preventedtheirwidespreaduse.EIA/IS
731 defined a Value GA and an Effectiveness GA and
provided high-level definitions for 5 levels of each GA.
The terms used by EIA/IS 731 to characterize the Value
GA included: “desirability,” “value,” “benefit,” “use-
fulness,” and “utility.” Terms and concepts used to
characterize the Effectiveness GA include: “benefits
received for the effort invested,” “...benefits received
worth the cost of the effort….” “…activities provide a
reasonable benefit….,” “...obviously beneficial...,”
“…maximum benefit for the amount of effort….” The
5 level rating scale was designed for use in ratings and
comprised the levels of Marginal, Adequate, Signifi-
cant, Measurably Significant, and Optimal”.
1.2. Improvements Needed for GAs
The following issues were raised (by both GA advo-
cates and detractors):
• The Effectiveness GA is difficult to understand;
some of its levels read like “return on investment”
while other levels seem to be almost the same as
the Value GA.
• Respondents to GA questionnaires neither under-
stand the GAs nor have a conceptual basis for an
answer.
• There are no guidelines or objective evidence
examples that would provide a basis for a respon-
dent’s answer; thus answers are highly subjective
(leading to appraisal results that are not repeat-
able).
• GAs should not be used to determine process
capability or maturity ratings; they represent a
different dimension.
1.3. Broad Expression of Interest in GAs
In spite of the practical problems, GAs (like an idea
whose time has come) “refuse to go away.” Continued
interest in GA-like concepts continues to be expressed,
as indicated in the following work:
• ISO 9004 [ISO, 2000] suggests consideration of
self-assessment output together with the potential
benefits to be gained for each process.
• ISO/IEC TR 15504-7:1998(E) [ISO, 1998] de-
scribes effectiveness measures as: “the extent to
which the software process achieves goals de-
rived directly from an analysis of specific circum-
stances, needs and business goals of the
organization.” p 23.
• “An Empirical Methodology for Introducing
Software Processes” [Shull, Carver, and Travas-
sos, 2001].
In Section 3.1, “Did the Process Provide Useable
Results”, the authors discuss feasibility studies to test
the effectiveness (did the process provide useable re-
sults?) of a process. In the following section, “Was the
Time Well Spent,” the authors discuss the determination
of whether the return on investment is reasonable.
The above sources suggest augmentation of process
appraisals with measurement of targeted process output
attributes to meet overall goals for organizational or
business improvement. ISO 15504 describes the use of
Effectiveness measures (process outputs) to support the
choice of processes to be improved and the monitoring
of improvement results.
The paper by Shull et al. [2001] outlines a multifac-
eted method for iteratively validating and/or improving
processes. An experiment, providing evidence of the
usefulness of the method, is described that includes
experimental results of investigating the usefulness and
cost effectiveness of a process.
1.4. Achieving a Workable Approach to GAs
FAA iCMM V2 [FAA, 2001] builds on, and extends,
EIA/IS 731 and the above work to describe a practical
approach to the use of Generic Attributes. The iCMM
V2 approach recognizes that GAs are not a factor in
process capability/maturity, but are indeed a different
dimension that can be used to confirm the effectiveness
of process capability/maturity measurements and pri-
oritize improvement areas. Objectivity in GA measure-
ments is obtained by providing clear definitions of the
GAs and by establishing an appraisal method that:
• Identifies and focuses on key results of process
areas for GA measurements.
• Identifies and distinguishes between “develop-
ers” and “users” of the key results.
• Asks appropriate questions of developers (e.g,
cost of developing the key result).
302 WELLS, IBRAHIM, AND LABRUYERE
• Asks appropriate questions of users (e.g., on use-
fulness of results and needs satisfaction).
• Identifies and examines objective evidence to
corroborate developer and user inputs.
2. DESCRIPTION OF FAA-iCMM V2
GENERIC ATTRIBUTES
Generic Attributes are a way of measuring two charac-
teristics of process performance results: Usefulness and
Cost Effectiveness. The Usefulness Generic Attribute
indicates the traditional sense of usefulness or value of
the products or results of performing processes relating
to a process area. The Cost Effectiveness Generic At-
tribute indicates the extent to which the benefit received
is worth the resources invested in producing the prod-
ucts or results.
These two characteristics of process performance
were chosen because they represent key issues in organ-
izational improvement. They help an organization ad-
dress the bottom-line questions such as: Are our
processes providing useful results in relation to our
business objectives? Are we spending our resources
wisely? Are we actually improving? Are our processes
more effective? Are we being efficient? Do we need
further improvement? Generic Attributes are measured
on a different scale from process capability and matur-
ity and are not considered in the determination of proc-
ess capability or maturity level. They are valuable as an
independent check on the alignment of process capabil-
ity and process performance. They do, however, pre-
sume that a specific described process is being followed
in order to come up with the results that are measured
by Generic Attributes. Then, if GAs have low scores for
a process area that is relatively high in process capabil-
ity, the improvement practices being performed at that
capability level may not actually be effective, and this
would suggest an important improvement path for the
organization to investigate.
Alternatively, if GA scores are acceptable from a
business perspective for a process area, further im-
provement in the process area may not be a high prior-
ity. The trend of GA scores for process areas of constant
capability level may also yield useful information about
process improvement methods and priorities. Other
useful enquiries based on GAs are: Is there a difference
in the perceptions of Usefulness when viewed across an
organization? Between perceptions of Cost Effective-
ness? Why? The collection of GA data on a given
process area begins with determination of the key serv-
ices or products that result from the process. This is
called the Key Result(s)for a GA appraisal.Key Results
can be any output of the process, relating to a Process
Area, that an organization chooses to focus on. These
outputs may be tangible outputs (e.g., work product
such as Trade Studies), or intangible outputs (e.g., a
service such as Customer Issue Response Time, or an
output such as “absence of risk consequences”). The
Key Result is the focus of GA appraisal data collection.
Interviews and objective evidence review focus on the
Key Result(s).
2.1. The Usefulness GA
Defining Usefulness. Usefulness is defined as: the ex-
tent to which work products or services provide the
needed benefits in actual use.
Usefulness is determined for the Key Result(s) as
indicated in Table I. In the table, interpret “meet the
need” based on the process area and Key Result(s).
Process area purpose, goals, and typical work products
can be used as starting points to identify the Key Re-
sult(s).
Sources of Usefulness Data. Usefulness data is
obtained primarily from interviewing users of the Key
Results and from review of related objective evidence.
Objective evidence may come from verification or vali-
Table I. Usefulness Generic Attribute
A NEW APPROACH TO GENERIC ATTRIBUTES 303
dation evaluations of Key Result(s). GA review of
evaluation objective evidence would consider charac-
teristics of the results (e.g., number of defects); in
contrast with objective evidence review for process
appraisals, which considers only the existence of re-
sults. The concepts of Key Result(s), identification of
users as the “authority” on their usefulness, and identi-
fication of categories of Key Result(s)/user objective
evidence represent “breakthroughs” for practical and
objective GA appraisals.
Scoring the Usefulness GA. A Usefulness score of
“U” indicates that there were no deficiencies in the Key
Result(s) with respect to meeting the need. U- indicates
that there were some deficiencies, and U-- indicates
significant deficiencies. U+ indicates that the stated
need was met and some additional value was provided
that fulfilled an unrecognized need. U++ indicates that
the stated need was met and significant additional value
was provided that fulfilled an unrecognized need. Pro-
jects should usually pursue Usefulness with caution, as
exceeding expectations may result in unnecessary ex-
pense (lowered Cost Effectiveness). The Usefulness
Generic Attribute is closely related to validation of
incremental work products described in the iCMM V2
Evaluation Process Area (PA 08). The process area
describes evaluations, including validation of the work
products or services of a process, to determine whether
they satisfy their need. The measurement of the Useful-
ness GA is very similar to validation of a product or
service component but it additionally measures the
extent to which the need is satisfied. A process (ap-
praisal method) for objectively measuring Usefulness
is further described in thefollowing sectionof thepaper.
2.2. The Cost Effectiveness GA
Defining Cost Effectiveness. Cost effectiveness is de-
fined as: the extent to which the benefits received are
worth the resources invested. Cost effectiveness is de-
termined through the use of an intermediate parame-
ter—“Resource Efficiency.”
Resource Efficiency. Resource efficiency is the de-
viation of the actual resources required to produce the
Key Result(s) from the benchmarked (expected) value
of needed resources.
The reference or target for resource investment for a
given work product or result is the benchmarked stand-
ard. If actual benchmark values are not available, other
credible reference values may be used—e.g., a value
supported by a planning estimate. In order to determine
Cost Effectiveness, the Resource Efficiency is first de-
termined as indicated in Table II.
The key to improving GA appraisal objectivity for
Cost Effectiveness, in addition to focusing on Key
Result(s), is the recognition that Key Result(s) develop-
ers and their supporting objective evidence are the
“authority” on Resource Efficiency. The personnel as-
signing values in the above table are knowledgeable of
the resource planning and expenditures relating to the
identified work products or results, in contrast to the
personnel making inputs on the Usefulness, who were
the users of the products or services. Types of objective
evidence for Resource Efficiency would include bench-
mark cost (or other cost basis) and actual cost related to
development of the Key Result(s).
An organization or enterprise can tailor the GAs
based on the type and availability of data or to meet its
unique needs. For example, it could set its own percent-
ages or other criteria for measuring the Resource Effi-
ciency or use only three levels (e.g., E, E+, and E–).
Tailoring does not preclude the use of GAs for bench-
marking; it just restricts benchmarking to the organiza-
tion or enterprise that defines the tailoring. A resource
efficiency measure may be pursued only if the product
is at least marginally useful (to avoid casesofestimating
resource efficiency for useless products).
Table II. Resource Efficiency
* The exact percentages can be tailored by the organization.
304 WELLS, IBRAHIM, AND LABRUYERE
Determining Cost Effectiveness. Once the re-
source efficiency is determined, the Cost Effectiveness
is represented according to the graph in Figure 1. The
figure illustrates GA plots for example process areas
(PAs). Key Results in the upper right quadrant, having
both high Usefulness and high Resource Efficiency, are
clearly highly Cost Effective. The lower left quadrant
represents low Cost Effectiveness. The other two quad-
rants represent intermediate Cost Effectiveness; any
degree of Cost Effectiveness in which the Usefulness is
below some threshold could be considered unaccept-
able.
3. GA APPRAISAL PROCESS
The GA appraisal process involves the following steps:
• Select a process area for GA appraisal.
• Determine the primary output/result(s) (Key Re-
sults) of the selected process area.
• Identify the process that produces the Key Re-
sults and the lead and staff for the process.
• Identify the process that uses the Key Results and
the lead and staff for the process.
• Collect data on the Usefulness GA from key
result users.
• Collect data on the Resource Efficiency parame-
ter from key result producers
• Score the GAs based on the collected data and the
GA definitions.
Identifying Process Results for GA Appraisals.
Generic Attributes measure characteristics of the out-
puts or results from performing a process. The FAA-
iCMM v2.0 model description helps the organization
identify what results might be measured during a GA
appraisal. For example, FAA-iCMM process area de-
scriptions include a purpose, goals, essential base prac-
tices for attaining the purpose or achieving the goals,
and example practice outputs or results that might be
expected. Such work products [possible Key Result(s)]
may include documents, services, or processes such as
a training process; example outcomes might be ex-
pressed in the PA purpose or goals. Understanding the
purpose, goals, and practices of a process helps in
determining the Key Result(s) (e.g., what results might
be expected and appraised by means of GAs). Examples
indicating work products (Key Results), potential users
and anticipated needs are shown in Table III. A GA
appraisal might ask users if these needs were being met
as a result of using the outputs of the process.
The appraisal data collection task, then, is to deter-
mine how well the key result(s) met the need (Useful-
ness GA) and the extent to which the cost of producing
the key result correlates to what it “should cost” (Re-
source Efficiency). Ideally the “should cost” comes
from benchmarking. When benchmarking data are not
available, historically based planning data can be used.
Determination of the Usefulness GA and the Resource
Efficiency parameter allow the Cost Effectiveness GA
to be determined.
One aspect of improving objectivity in GA measure-
ments is the recognition that Usefulness and Resource
Efficiency data typically come from two different
groups of project staff. Users of the key results tend to
be the authoritative source of data on Usefulness, while
producers of key results tend to be the authoritative
source of data on Resource Efficiency.
Key Result Producers. Interviews of key result
producers focus on the following questions:
• Please describe the benchmark data used to plan
development of the Key Result(s).
• How did the actual cost of development compare
with benchmark data?
• Please describe the results of any reviews of the
key results and any rework required or issues.
Key Result Users. Questions asked of key results
users are:
• Please describe how the Key Result was used.
• Were there any issues associate with use of the
Key Results? Any unexpected problems? Any
unexpected benefits?
Figure 1. Cost effectiveness.
A NEW APPROACH TO GENERIC ATTRIBUTES 305
Improving Objectivity. Another aspect of improv-
ing objectivity is obtaining quantitative data to support
the GA determinations. Objective evidence is requested
to support each interview question. Benchmarked cost
and the amount of rework (required to meet the need)
can be quantitatively tied to Resource Efficiency and
operational deficiencies or benefits can be quantita-
tively tied to the Usefulness GA.
4. AUGMENTING PROCESS APPRAISALS
Current Methods. Current process appraisal methods,
including the FAA-iCMM Appraisal Method (FAM)
version 1.0 [FAA, 1999], focus on assessing the capa-
bility and maturity of the processes. These methods
focus on reasonableness of the process and do not
evaluate the Usefulness and Cost Effectiveness of the
results of a process. The FAM defines a formal full
robust process appraisal methodology along with five
variations that all follow this focus. It is planned that
the GA appraisal method will be added to the FAM as
an additional variation to provide an optional way of
evaluating Usefulness and Cost Effectiveness of the
process results.
The FAM allows for combining, tailoring, and aug-
menting the different appraisal types. This means that
the GA appraisal may be performed by itself or as part
of a process appraisal depending on the sponsor’s goals
for an appraisal.
Opportunities to Use GAs. Although GA apprais-
als can be done independently of process appraisals, it
is often simple to integrate the process appraisal and the
GA appraisal. Generic Attributes can augment process
improvement in the following ways:
• If an organization has recently performed a proc-
ess appraisal, or would like to better understand
a product/service result that is not related to an
FAA-iCMM process area they have been cur-
Table III. Sample Processes, Outputs, Users, and Needs
306 WELLS, IBRAHIM, AND LABRUYERE
rently pursuing, a standalone GA appraisal may
be required.
• If an organization is just starting out with process
improvement, a variation of the FAM may be
used to focus on initial, high-level, issues that the
organization needs to focus on. In addition, a GA
appraisal may be used to develop baseline meas-
urements.
• If an organization has been doing process im-
provement for some time and would like to have
a formal FAM to verify that they have achieved a
capability or maturity level, a GA appraisal may
be used to additionally determine the efficiency
and effectiveness of a process. If other GA ap-
praisals have been done for an organization in the
past, the trend and correlation between the proc-
ess improvement and its effects on the prod-
uct/service can be observed. (Note: The real
power of the GA appraisal is realized when its
results are compared with process appraisal re-
sults and trends are established.)
Integrating GAs into Appraisals. It is simple to
integrate the process appraisal and the GA appraisal.
Both types of appraisals interview people and review
documents. In most cases they are talking to the same
people and looking at the same documents, just from
different perspectives. Depending on the process and
GA appraisal scope, one, or at most two, people can be
added to the process appraisal team whose role is to
focus on the GA appraisal. The existing appraisal Proc-
ess Area focused mini-teams can be used to help with
document review and providing the correlation of the
process to GA attributes. Process appraisals normally
take 1–2 weeks of on-site time, depending on scope,
and this time would normally not need to be extended.
Piloting Generic Attributes. A program is under-
way to evaluate the appraisal method for Generic At-
tributes through a series of pilots that began in the forth
quarter of 2002. The first pilot involved appraisal of the
Usefulness GA for Risk Management. In meetings with
the appraisal sponsors, several Risk Management out-
comes were considered as possible Key Results that
could be used as a focus for the appraisal. These were:
• the extent to which unanticipated events affected
the projects performance (relates to the overall
usefulness of the risk management program)
• the fraction of adverse project events that were
not identified as risks during risk identification
activities (relates to the usefulness of the risk
identification activities)
• the fraction of “High” risk items whose level of
risk was reduced to “Low” within the planned
time frame (relates to usefulness of the risk miti-
gation activities).
The fraction of risk items reduced from high to low
risk was selected as the Usefulness Key Result to be
measured in the first pilot because of the availability of
objective data and the value of the appraisal data to the
sponsor and projects. Initial results from the survey-
based appraisal of the Usefulness GA for Risk Manage-
ment underscore the dependence on measurement for
high confidence (quantitative) results. That is, to
achieve high confidence in the GA appraisal results, an
organization’s measurement program would yield
measurement data that relate to the GA Key results.
5. SUMMARY AND CONCLUSIONS
The approach presented in this paper, along with pre-
liminary GA appraisal results, show that the Generic
Attributes of Usefulness and Cost Effectiveness can
indeed be appraised with objectivity comparable to that
of process appraisals. While the GAs are not totally
independent, this does not distract from their value to
the appraised organization. Further, it is not necessary
to perform a GA appraisal using both GAs. By aug-
menting process appraisals with the measurement of
GAs, organizations can now establish a relationship
between process capability and process results.
The FAA, with the iCMM V2, is continuing to
evaluate and refine approaches for appraising process
performance via Generic Attributes. The FAA invites,
and strongly encourages, other organizations experi-
menting with these concepts to contact the authors
regarding their ideas and experiences. E-mail addresses
for the authors are: cwells@i-metrics.net, Linda.ibra-
him@faa.gov, Larry.labruyere@auatac.com.
REFERENCES
EIA, Systems Engineering Capability (SECM), EIA/IS 731,
Electronic Industries Alliance (EIA) Government Elec-
tronics and Information Technology Association, Ar-
lington, VA, http:/www.geia.org/eoc/G47 (March 31,
1999).
FAA, The Federal Aviation Administration Integrated Capa-
bility Maturity Model (FAA-iCMM), Version 2.0, Federal
Aviation Administration, Washington, DC,
http:/www.faa.gov/aio (September 2001).
FAA, The FAA-iCMM Appraisal Method (FAM), Version
1.0, Federal Aviation Administration, Washington, DC,
http:/www.faa.gov/aio (April 1999).
ISO, ISO/IEC TR 15504, Informationtechnology—Software
process assessment, Part 7: Guidelines for software proc-
ess improvement, International Organization for Stand-
A NEW APPROACH TO GENERIC ATTRIBUTES 307
ardization and International Electrotechnical Commis-
sion, Geneva, Switzerland, 1998.
ISO, ISO 9004:2000, Quality management systems—Guide-
lines for performance improvements, ISO 2000-12-15,
International Organization for Standardization, Geneva,
Switzerland, 2000.
F. Schull, J. Carver, and G. Travassos, An empirical method-
ology for introducing software processes, 8th Eur Soft-
ware Eng Conf and 9th ACM SIGSOFT Symp
Foundations of Software Engineering, Vienna, Austria,
September 2001, pp. 288–296.
Curt Wells provides process improvement and training services through his company, I-metrics LLC, and
is an authorized lead appraiser for the FAA iCMM®
. He is a co-author of the SE-CMM®
, EIA/IS 731,
CMMI V1, and FAA iCMM®
V2. Curt retired from Lockheed Martin after 30 years service in various
systems engineering and management roles. He holds a Masters Degree in Physics from Sam Houston
State University.
Linda Ibrahim joined FAA in 1996. She is FAA Chief Engineer for Process Improvement, providing
technical leadership for FAA-wide process improvement. She led development of FAA-iCMM®
v1.0 and
v2.0, and its appraisal method. Linda has been working in software engineering for over 30 years, as
practitioner, educator, and researcher; in the United States, Europe, and Middle East. She worked at the
Software Engineering Institute for several years, and is a member of the CMMI Steering Group. Linda
holds a B.A. in Mathematics, an M.S. in Information Science, and a Ph.D. in Electrical Engineering. She
is a member of INCOSE, ACM, and IEEE.
Larry LaBruyere has been with Northrop Grumman Mission Systems for over 20 years and supports the
Federal Aviation Administration’s Chief Information Officer on a technical assistance contract supporting
process improvement based on the FAA-iCMM®
. He is the appraisal coordinator and an FAM appraisal
lead for the FAA and participated in the authoring of the FAA-iCMM®
and FAM. Mr. LaBruyere is an
“Authorized CBA IPI Lead Assessor” for the SW-CMM®
and authorized Standard CMMI Appraisal
Method for Process Improvement (SCAMPI) Lead Appraiser for the CMMI. He isanauthorizedinstructor
for Practical Software and Systems Measurement. He obtained an M.S. from American University and a
B.A. from Anderson University.
308 WELLS, IBRAHIM, AND LABRUYERE

More Related Content

What's hot

Quality management manual template
Quality management manual templateQuality management manual template
Quality management manual template
selinasimpson331
 
Nandyal_20150513_1145_1230.original.1431384030
Nandyal_20150513_1145_1230.original.1431384030Nandyal_20150513_1145_1230.original.1431384030
Nandyal_20150513_1145_1230.original.1431384030
Raghav Nandyal
 
Quality management procedures
Quality management proceduresQuality management procedures
Quality management procedures
selinasimpson1201
 

What's hot (20)

Quality management manual template
Quality management manual templateQuality management manual template
Quality management manual template
 
Response Surface Regression
Response Surface RegressionResponse Surface Regression
Response Surface Regression
 
A lean model based outlook on cost & quality optimization in software projects
A lean model based outlook on cost & quality optimization in software projectsA lean model based outlook on cost & quality optimization in software projects
A lean model based outlook on cost & quality optimization in software projects
 
Nandyal_20150513_1145_1230.original.1431384030
Nandyal_20150513_1145_1230.original.1431384030Nandyal_20150513_1145_1230.original.1431384030
Nandyal_20150513_1145_1230.original.1431384030
 
2010-36-0387I_Paper
2010-36-0387I_Paper2010-36-0387I_Paper
2010-36-0387I_Paper
 
Cmmi model – capabilities maturity model integration
Cmmi model – capabilities maturity model integrationCmmi model – capabilities maturity model integration
Cmmi model – capabilities maturity model integration
 
An Application of Green Quality Function Deployment to Designing an Air Condi...
An Application of Green Quality Function Deployment to Designing an Air Condi...An Application of Green Quality Function Deployment to Designing an Air Condi...
An Application of Green Quality Function Deployment to Designing an Air Condi...
 
Process Control
Process ControlProcess Control
Process Control
 
8 quality PMBOK
8 quality PMBOK8 quality PMBOK
8 quality PMBOK
 
Process capability
Process capabilityProcess capability
Process capability
 
Ch 7(spi)intro tocm-mi2013
Ch 7(spi)intro tocm-mi2013Ch 7(spi)intro tocm-mi2013
Ch 7(spi)intro tocm-mi2013
 
Introduction to CMMI-DEV v1.3 - Day 4
Introduction to CMMI-DEV v1.3  - Day 4Introduction to CMMI-DEV v1.3  - Day 4
Introduction to CMMI-DEV v1.3 - Day 4
 
Quality management procedures
Quality management proceduresQuality management procedures
Quality management procedures
 
2nd Generation Six Sigma the Breakdown (at Samsung)
2nd Generation Six Sigma the Breakdown (at Samsung)2nd Generation Six Sigma the Breakdown (at Samsung)
2nd Generation Six Sigma the Breakdown (at Samsung)
 
S3 p2 lee_kong_hui
S3 p2 lee_kong_huiS3 p2 lee_kong_hui
S3 p2 lee_kong_hui
 
A041130105
A041130105A041130105
A041130105
 
Tools and Techniques of Quality Planning
Tools and Techniques of Quality PlanningTools and Techniques of Quality Planning
Tools and Techniques of Quality Planning
 
Introduction to CMMI-DEV v1.3 - Day 2
Introduction to CMMI-DEV v1.3  - Day 2Introduction to CMMI-DEV v1.3  - Day 2
Introduction to CMMI-DEV v1.3 - Day 2
 
Benchmarking : An essential Quality tool
Benchmarking : An essential Quality toolBenchmarking : An essential Quality tool
Benchmarking : An essential Quality tool
 
Continuous improvement methods summary by the sig rev052914
Continuous improvement methods summary by the sig rev052914Continuous improvement methods summary by the sig rev052914
Continuous improvement methods summary by the sig rev052914
 

Similar to Ga article

Performance management
Performance managementPerformance management
Performance management
vasudhapaul
 
Project on quality management
Project on quality managementProject on quality management
Project on quality management
selinasimpson2901
 
Performance measurement system design
Performance measurement system designPerformance measurement system design
Performance measurement system design
Fernanda Araujo
 
Measurement-Process-Effectiveness_paper_updated210
Measurement-Process-Effectiveness_paper_updated210Measurement-Process-Effectiveness_paper_updated210
Measurement-Process-Effectiveness_paper_updated210
pbaxter
 
Quality management interview questions
Quality management interview questionsQuality management interview questions
Quality management interview questions
selinasimpson2301
 

Similar to Ga article (20)

Performance management
Performance managementPerformance management
Performance management
 
Improving the roi of software quality assurance activities
Improving the roi of software quality assurance activitiesImproving the roi of software quality assurance activities
Improving the roi of software quality assurance activities
 
Project on quality management
Project on quality managementProject on quality management
Project on quality management
 
Performance measurement system design
Performance measurement system designPerformance measurement system design
Performance measurement system design
 
Crafting a CMMI V2 Compliant Process for Governance Practice Area: An Experie...
Crafting a CMMI V2 Compliant Process for Governance Practice Area: An Experie...Crafting a CMMI V2 Compliant Process for Governance Practice Area: An Experie...
Crafting a CMMI V2 Compliant Process for Governance Practice Area: An Experie...
 
PROCESS IMPROVEMENT ‎CYCLE
PROCESS IMPROVEMENT ‎CYCLE PROCESS IMPROVEMENT ‎CYCLE
PROCESS IMPROVEMENT ‎CYCLE
 
Test performance indicators
Test performance indicatorsTest performance indicators
Test performance indicators
 
Maintenance Leading and Lagging Key Performance Indicators (KPIs)
Maintenance Leading and Lagging Key Performance Indicators (KPIs)Maintenance Leading and Lagging Key Performance Indicators (KPIs)
Maintenance Leading and Lagging Key Performance Indicators (KPIs)
 
Gm assessing performance
Gm assessing performanceGm assessing performance
Gm assessing performance
 
Chapter 4 - Performance Testing Tasks
Chapter 4 - Performance Testing TasksChapter 4 - Performance Testing Tasks
Chapter 4 - Performance Testing Tasks
 
Quality management topics
Quality management topicsQuality management topics
Quality management topics
 
Process improvement for General Counsel and Law Firms
Process improvement for General Counsel and Law FirmsProcess improvement for General Counsel and Law Firms
Process improvement for General Counsel and Law Firms
 
Using Benchmarking to Quantify the Benefits of Software Process Improvement
Using Benchmarking to Quantify the Benefits of Software Process ImprovementUsing Benchmarking to Quantify the Benefits of Software Process Improvement
Using Benchmarking to Quantify the Benefits of Software Process Improvement
 
Quality - An Introduction-170715
Quality - An Introduction-170715Quality - An Introduction-170715
Quality - An Introduction-170715
 
Process of Operation management
Process of Operation management Process of Operation management
Process of Operation management
 
Measurement-Process-Effectiveness_paper_updated210
Measurement-Process-Effectiveness_paper_updated210Measurement-Process-Effectiveness_paper_updated210
Measurement-Process-Effectiveness_paper_updated210
 
Beanchmarketing
BeanchmarketingBeanchmarketing
Beanchmarketing
 
Benchmarking
BenchmarkingBenchmarking
Benchmarking
 
Operations Management: Six sigma benchmarking of process capability analysis...
Operations Management:  Six sigma benchmarking of process capability analysis...Operations Management:  Six sigma benchmarking of process capability analysis...
Operations Management: Six sigma benchmarking of process capability analysis...
 
Quality management interview questions
Quality management interview questionsQuality management interview questions
Quality management interview questions
 

Ga article

  • 1. A New Approach to Generic Attributes* Curt Wells,1, † Linda Ibrahim,2 and Larry LaBruyere3 1 I-metrics LLC, 90 S. Camino Real, Uhland, TX 78640 2 Federal Aviation Administration, 800 Independence Avenue SW, Washington, DC 20591 3 Northrop Grumman Mission Systems, 475 School Street SW, Washington, DC 20024 A NEW APPROACH TO GENERIC ATTRIBUTES Received 10 January 2003; Accepted 20 June 2003 DOI 10.1002/sys.10050 ABSTRACT Generic Attributes (GAs) are measures of process performance introduced by the systems engineering capability model, EIA/IS 731. The systems model defines two GAs—Effectiveness and Value. While the concept of GAs is generally accepted as valid, they have been used very little, due to difficulties in interpretation and appraisal. Revised definitions and a new approach for appraising and using GAs have been developed for the Federal Aviation Administration Integrated Capability Maturity Model® v2.0 (FAA-iCMM®‡ ). The improved approach to Generic Attributes provides the needed clarity and measurement objectivity. Definitions of the iCMM® GAs and their relationship to process model concepts are described in this paper, along with a practical approach to appraising GAs. © 2003 Wiley Periodicals, Inc.§ Syst Eng 6: 301–308, 2003 Key words: process appraisal effectiveness, generic attributes, Integrated Capability Maturity Model® 1. INTRODUCTION The Need for Generic Attributes.Although capability and maturity levels indicate the level of process disci- pline that organizational units may achieve in one or more process areas, the levels do not necessarily trans- late into the quality of work products or quality/value of results in meeting objectives. It is possible to have mature processes (e.g., capability level 3, 4, or 5) that produce unsatisfactory work products (e.g., of less than desirable quality, or that do not meet needs, for greater Correspondence *Measurement Category Best Paper Award, 2002 INCOSE Sympo- sium (updated for this article). † Author to whom all correspondence should be addressed (e-mail: cwells@i-metrics.net). ‡ CMM and Capability Maturity Model are registered in the U.S. Patent and Trademark Office. Systems Engineering, Vol. 6, No. 4, 2003 © 2003WileyPeriodicals,Inc.§ This article is a US Government work and, as such, is in the public domain in the United States of America. 301
  • 2. than expected expense). In such cases, enhanced proc- ess capability may not result in the actual performance improvements that organizations expect from pursuing process improvement. Measurements of the quality or value of process results (i.e., GA-like measurements) have not been used heretofore, to any significant extent, largely because of the difficulty and lack of objective methods to do so. 1.1. GAs Introduced by EIA/IS 731 These concerns led the authors of EIA/IS 731 [EIA, 1999] to define GAs and incorporate them into the appraisal process. While the need for GAs (as defined by EIA/IS 731) was widely acknowledged by the proc- ess improvement and appraisal community, operational difficulties have preventedtheirwidespreaduse.EIA/IS 731 defined a Value GA and an Effectiveness GA and provided high-level definitions for 5 levels of each GA. The terms used by EIA/IS 731 to characterize the Value GA included: “desirability,” “value,” “benefit,” “use- fulness,” and “utility.” Terms and concepts used to characterize the Effectiveness GA include: “benefits received for the effort invested,” “...benefits received worth the cost of the effort….” “…activities provide a reasonable benefit….,” “...obviously beneficial...,” “…maximum benefit for the amount of effort….” The 5 level rating scale was designed for use in ratings and comprised the levels of Marginal, Adequate, Signifi- cant, Measurably Significant, and Optimal”. 1.2. Improvements Needed for GAs The following issues were raised (by both GA advo- cates and detractors): • The Effectiveness GA is difficult to understand; some of its levels read like “return on investment” while other levels seem to be almost the same as the Value GA. • Respondents to GA questionnaires neither under- stand the GAs nor have a conceptual basis for an answer. • There are no guidelines or objective evidence examples that would provide a basis for a respon- dent’s answer; thus answers are highly subjective (leading to appraisal results that are not repeat- able). • GAs should not be used to determine process capability or maturity ratings; they represent a different dimension. 1.3. Broad Expression of Interest in GAs In spite of the practical problems, GAs (like an idea whose time has come) “refuse to go away.” Continued interest in GA-like concepts continues to be expressed, as indicated in the following work: • ISO 9004 [ISO, 2000] suggests consideration of self-assessment output together with the potential benefits to be gained for each process. • ISO/IEC TR 15504-7:1998(E) [ISO, 1998] de- scribes effectiveness measures as: “the extent to which the software process achieves goals de- rived directly from an analysis of specific circum- stances, needs and business goals of the organization.” p 23. • “An Empirical Methodology for Introducing Software Processes” [Shull, Carver, and Travas- sos, 2001]. In Section 3.1, “Did the Process Provide Useable Results”, the authors discuss feasibility studies to test the effectiveness (did the process provide useable re- sults?) of a process. In the following section, “Was the Time Well Spent,” the authors discuss the determination of whether the return on investment is reasonable. The above sources suggest augmentation of process appraisals with measurement of targeted process output attributes to meet overall goals for organizational or business improvement. ISO 15504 describes the use of Effectiveness measures (process outputs) to support the choice of processes to be improved and the monitoring of improvement results. The paper by Shull et al. [2001] outlines a multifac- eted method for iteratively validating and/or improving processes. An experiment, providing evidence of the usefulness of the method, is described that includes experimental results of investigating the usefulness and cost effectiveness of a process. 1.4. Achieving a Workable Approach to GAs FAA iCMM V2 [FAA, 2001] builds on, and extends, EIA/IS 731 and the above work to describe a practical approach to the use of Generic Attributes. The iCMM V2 approach recognizes that GAs are not a factor in process capability/maturity, but are indeed a different dimension that can be used to confirm the effectiveness of process capability/maturity measurements and pri- oritize improvement areas. Objectivity in GA measure- ments is obtained by providing clear definitions of the GAs and by establishing an appraisal method that: • Identifies and focuses on key results of process areas for GA measurements. • Identifies and distinguishes between “develop- ers” and “users” of the key results. • Asks appropriate questions of developers (e.g, cost of developing the key result). 302 WELLS, IBRAHIM, AND LABRUYERE
  • 3. • Asks appropriate questions of users (e.g., on use- fulness of results and needs satisfaction). • Identifies and examines objective evidence to corroborate developer and user inputs. 2. DESCRIPTION OF FAA-iCMM V2 GENERIC ATTRIBUTES Generic Attributes are a way of measuring two charac- teristics of process performance results: Usefulness and Cost Effectiveness. The Usefulness Generic Attribute indicates the traditional sense of usefulness or value of the products or results of performing processes relating to a process area. The Cost Effectiveness Generic At- tribute indicates the extent to which the benefit received is worth the resources invested in producing the prod- ucts or results. These two characteristics of process performance were chosen because they represent key issues in organ- izational improvement. They help an organization ad- dress the bottom-line questions such as: Are our processes providing useful results in relation to our business objectives? Are we spending our resources wisely? Are we actually improving? Are our processes more effective? Are we being efficient? Do we need further improvement? Generic Attributes are measured on a different scale from process capability and matur- ity and are not considered in the determination of proc- ess capability or maturity level. They are valuable as an independent check on the alignment of process capabil- ity and process performance. They do, however, pre- sume that a specific described process is being followed in order to come up with the results that are measured by Generic Attributes. Then, if GAs have low scores for a process area that is relatively high in process capabil- ity, the improvement practices being performed at that capability level may not actually be effective, and this would suggest an important improvement path for the organization to investigate. Alternatively, if GA scores are acceptable from a business perspective for a process area, further im- provement in the process area may not be a high prior- ity. The trend of GA scores for process areas of constant capability level may also yield useful information about process improvement methods and priorities. Other useful enquiries based on GAs are: Is there a difference in the perceptions of Usefulness when viewed across an organization? Between perceptions of Cost Effective- ness? Why? The collection of GA data on a given process area begins with determination of the key serv- ices or products that result from the process. This is called the Key Result(s)for a GA appraisal.Key Results can be any output of the process, relating to a Process Area, that an organization chooses to focus on. These outputs may be tangible outputs (e.g., work product such as Trade Studies), or intangible outputs (e.g., a service such as Customer Issue Response Time, or an output such as “absence of risk consequences”). The Key Result is the focus of GA appraisal data collection. Interviews and objective evidence review focus on the Key Result(s). 2.1. The Usefulness GA Defining Usefulness. Usefulness is defined as: the ex- tent to which work products or services provide the needed benefits in actual use. Usefulness is determined for the Key Result(s) as indicated in Table I. In the table, interpret “meet the need” based on the process area and Key Result(s). Process area purpose, goals, and typical work products can be used as starting points to identify the Key Re- sult(s). Sources of Usefulness Data. Usefulness data is obtained primarily from interviewing users of the Key Results and from review of related objective evidence. Objective evidence may come from verification or vali- Table I. Usefulness Generic Attribute A NEW APPROACH TO GENERIC ATTRIBUTES 303
  • 4. dation evaluations of Key Result(s). GA review of evaluation objective evidence would consider charac- teristics of the results (e.g., number of defects); in contrast with objective evidence review for process appraisals, which considers only the existence of re- sults. The concepts of Key Result(s), identification of users as the “authority” on their usefulness, and identi- fication of categories of Key Result(s)/user objective evidence represent “breakthroughs” for practical and objective GA appraisals. Scoring the Usefulness GA. A Usefulness score of “U” indicates that there were no deficiencies in the Key Result(s) with respect to meeting the need. U- indicates that there were some deficiencies, and U-- indicates significant deficiencies. U+ indicates that the stated need was met and some additional value was provided that fulfilled an unrecognized need. U++ indicates that the stated need was met and significant additional value was provided that fulfilled an unrecognized need. Pro- jects should usually pursue Usefulness with caution, as exceeding expectations may result in unnecessary ex- pense (lowered Cost Effectiveness). The Usefulness Generic Attribute is closely related to validation of incremental work products described in the iCMM V2 Evaluation Process Area (PA 08). The process area describes evaluations, including validation of the work products or services of a process, to determine whether they satisfy their need. The measurement of the Useful- ness GA is very similar to validation of a product or service component but it additionally measures the extent to which the need is satisfied. A process (ap- praisal method) for objectively measuring Usefulness is further described in thefollowing sectionof thepaper. 2.2. The Cost Effectiveness GA Defining Cost Effectiveness. Cost effectiveness is de- fined as: the extent to which the benefits received are worth the resources invested. Cost effectiveness is de- termined through the use of an intermediate parame- ter—“Resource Efficiency.” Resource Efficiency. Resource efficiency is the de- viation of the actual resources required to produce the Key Result(s) from the benchmarked (expected) value of needed resources. The reference or target for resource investment for a given work product or result is the benchmarked stand- ard. If actual benchmark values are not available, other credible reference values may be used—e.g., a value supported by a planning estimate. In order to determine Cost Effectiveness, the Resource Efficiency is first de- termined as indicated in Table II. The key to improving GA appraisal objectivity for Cost Effectiveness, in addition to focusing on Key Result(s), is the recognition that Key Result(s) develop- ers and their supporting objective evidence are the “authority” on Resource Efficiency. The personnel as- signing values in the above table are knowledgeable of the resource planning and expenditures relating to the identified work products or results, in contrast to the personnel making inputs on the Usefulness, who were the users of the products or services. Types of objective evidence for Resource Efficiency would include bench- mark cost (or other cost basis) and actual cost related to development of the Key Result(s). An organization or enterprise can tailor the GAs based on the type and availability of data or to meet its unique needs. For example, it could set its own percent- ages or other criteria for measuring the Resource Effi- ciency or use only three levels (e.g., E, E+, and E–). Tailoring does not preclude the use of GAs for bench- marking; it just restricts benchmarking to the organiza- tion or enterprise that defines the tailoring. A resource efficiency measure may be pursued only if the product is at least marginally useful (to avoid casesofestimating resource efficiency for useless products). Table II. Resource Efficiency * The exact percentages can be tailored by the organization. 304 WELLS, IBRAHIM, AND LABRUYERE
  • 5. Determining Cost Effectiveness. Once the re- source efficiency is determined, the Cost Effectiveness is represented according to the graph in Figure 1. The figure illustrates GA plots for example process areas (PAs). Key Results in the upper right quadrant, having both high Usefulness and high Resource Efficiency, are clearly highly Cost Effective. The lower left quadrant represents low Cost Effectiveness. The other two quad- rants represent intermediate Cost Effectiveness; any degree of Cost Effectiveness in which the Usefulness is below some threshold could be considered unaccept- able. 3. GA APPRAISAL PROCESS The GA appraisal process involves the following steps: • Select a process area for GA appraisal. • Determine the primary output/result(s) (Key Re- sults) of the selected process area. • Identify the process that produces the Key Re- sults and the lead and staff for the process. • Identify the process that uses the Key Results and the lead and staff for the process. • Collect data on the Usefulness GA from key result users. • Collect data on the Resource Efficiency parame- ter from key result producers • Score the GAs based on the collected data and the GA definitions. Identifying Process Results for GA Appraisals. Generic Attributes measure characteristics of the out- puts or results from performing a process. The FAA- iCMM v2.0 model description helps the organization identify what results might be measured during a GA appraisal. For example, FAA-iCMM process area de- scriptions include a purpose, goals, essential base prac- tices for attaining the purpose or achieving the goals, and example practice outputs or results that might be expected. Such work products [possible Key Result(s)] may include documents, services, or processes such as a training process; example outcomes might be ex- pressed in the PA purpose or goals. Understanding the purpose, goals, and practices of a process helps in determining the Key Result(s) (e.g., what results might be expected and appraised by means of GAs). Examples indicating work products (Key Results), potential users and anticipated needs are shown in Table III. A GA appraisal might ask users if these needs were being met as a result of using the outputs of the process. The appraisal data collection task, then, is to deter- mine how well the key result(s) met the need (Useful- ness GA) and the extent to which the cost of producing the key result correlates to what it “should cost” (Re- source Efficiency). Ideally the “should cost” comes from benchmarking. When benchmarking data are not available, historically based planning data can be used. Determination of the Usefulness GA and the Resource Efficiency parameter allow the Cost Effectiveness GA to be determined. One aspect of improving objectivity in GA measure- ments is the recognition that Usefulness and Resource Efficiency data typically come from two different groups of project staff. Users of the key results tend to be the authoritative source of data on Usefulness, while producers of key results tend to be the authoritative source of data on Resource Efficiency. Key Result Producers. Interviews of key result producers focus on the following questions: • Please describe the benchmark data used to plan development of the Key Result(s). • How did the actual cost of development compare with benchmark data? • Please describe the results of any reviews of the key results and any rework required or issues. Key Result Users. Questions asked of key results users are: • Please describe how the Key Result was used. • Were there any issues associate with use of the Key Results? Any unexpected problems? Any unexpected benefits? Figure 1. Cost effectiveness. A NEW APPROACH TO GENERIC ATTRIBUTES 305
  • 6. Improving Objectivity. Another aspect of improv- ing objectivity is obtaining quantitative data to support the GA determinations. Objective evidence is requested to support each interview question. Benchmarked cost and the amount of rework (required to meet the need) can be quantitatively tied to Resource Efficiency and operational deficiencies or benefits can be quantita- tively tied to the Usefulness GA. 4. AUGMENTING PROCESS APPRAISALS Current Methods. Current process appraisal methods, including the FAA-iCMM Appraisal Method (FAM) version 1.0 [FAA, 1999], focus on assessing the capa- bility and maturity of the processes. These methods focus on reasonableness of the process and do not evaluate the Usefulness and Cost Effectiveness of the results of a process. The FAM defines a formal full robust process appraisal methodology along with five variations that all follow this focus. It is planned that the GA appraisal method will be added to the FAM as an additional variation to provide an optional way of evaluating Usefulness and Cost Effectiveness of the process results. The FAM allows for combining, tailoring, and aug- menting the different appraisal types. This means that the GA appraisal may be performed by itself or as part of a process appraisal depending on the sponsor’s goals for an appraisal. Opportunities to Use GAs. Although GA apprais- als can be done independently of process appraisals, it is often simple to integrate the process appraisal and the GA appraisal. Generic Attributes can augment process improvement in the following ways: • If an organization has recently performed a proc- ess appraisal, or would like to better understand a product/service result that is not related to an FAA-iCMM process area they have been cur- Table III. Sample Processes, Outputs, Users, and Needs 306 WELLS, IBRAHIM, AND LABRUYERE
  • 7. rently pursuing, a standalone GA appraisal may be required. • If an organization is just starting out with process improvement, a variation of the FAM may be used to focus on initial, high-level, issues that the organization needs to focus on. In addition, a GA appraisal may be used to develop baseline meas- urements. • If an organization has been doing process im- provement for some time and would like to have a formal FAM to verify that they have achieved a capability or maturity level, a GA appraisal may be used to additionally determine the efficiency and effectiveness of a process. If other GA ap- praisals have been done for an organization in the past, the trend and correlation between the proc- ess improvement and its effects on the prod- uct/service can be observed. (Note: The real power of the GA appraisal is realized when its results are compared with process appraisal re- sults and trends are established.) Integrating GAs into Appraisals. It is simple to integrate the process appraisal and the GA appraisal. Both types of appraisals interview people and review documents. In most cases they are talking to the same people and looking at the same documents, just from different perspectives. Depending on the process and GA appraisal scope, one, or at most two, people can be added to the process appraisal team whose role is to focus on the GA appraisal. The existing appraisal Proc- ess Area focused mini-teams can be used to help with document review and providing the correlation of the process to GA attributes. Process appraisals normally take 1–2 weeks of on-site time, depending on scope, and this time would normally not need to be extended. Piloting Generic Attributes. A program is under- way to evaluate the appraisal method for Generic At- tributes through a series of pilots that began in the forth quarter of 2002. The first pilot involved appraisal of the Usefulness GA for Risk Management. In meetings with the appraisal sponsors, several Risk Management out- comes were considered as possible Key Results that could be used as a focus for the appraisal. These were: • the extent to which unanticipated events affected the projects performance (relates to the overall usefulness of the risk management program) • the fraction of adverse project events that were not identified as risks during risk identification activities (relates to the usefulness of the risk identification activities) • the fraction of “High” risk items whose level of risk was reduced to “Low” within the planned time frame (relates to usefulness of the risk miti- gation activities). The fraction of risk items reduced from high to low risk was selected as the Usefulness Key Result to be measured in the first pilot because of the availability of objective data and the value of the appraisal data to the sponsor and projects. Initial results from the survey- based appraisal of the Usefulness GA for Risk Manage- ment underscore the dependence on measurement for high confidence (quantitative) results. That is, to achieve high confidence in the GA appraisal results, an organization’s measurement program would yield measurement data that relate to the GA Key results. 5. SUMMARY AND CONCLUSIONS The approach presented in this paper, along with pre- liminary GA appraisal results, show that the Generic Attributes of Usefulness and Cost Effectiveness can indeed be appraised with objectivity comparable to that of process appraisals. While the GAs are not totally independent, this does not distract from their value to the appraised organization. Further, it is not necessary to perform a GA appraisal using both GAs. By aug- menting process appraisals with the measurement of GAs, organizations can now establish a relationship between process capability and process results. The FAA, with the iCMM V2, is continuing to evaluate and refine approaches for appraising process performance via Generic Attributes. The FAA invites, and strongly encourages, other organizations experi- menting with these concepts to contact the authors regarding their ideas and experiences. E-mail addresses for the authors are: cwells@i-metrics.net, Linda.ibra- him@faa.gov, Larry.labruyere@auatac.com. REFERENCES EIA, Systems Engineering Capability (SECM), EIA/IS 731, Electronic Industries Alliance (EIA) Government Elec- tronics and Information Technology Association, Ar- lington, VA, http:/www.geia.org/eoc/G47 (March 31, 1999). FAA, The Federal Aviation Administration Integrated Capa- bility Maturity Model (FAA-iCMM), Version 2.0, Federal Aviation Administration, Washington, DC, http:/www.faa.gov/aio (September 2001). FAA, The FAA-iCMM Appraisal Method (FAM), Version 1.0, Federal Aviation Administration, Washington, DC, http:/www.faa.gov/aio (April 1999). ISO, ISO/IEC TR 15504, Informationtechnology—Software process assessment, Part 7: Guidelines for software proc- ess improvement, International Organization for Stand- A NEW APPROACH TO GENERIC ATTRIBUTES 307
  • 8. ardization and International Electrotechnical Commis- sion, Geneva, Switzerland, 1998. ISO, ISO 9004:2000, Quality management systems—Guide- lines for performance improvements, ISO 2000-12-15, International Organization for Standardization, Geneva, Switzerland, 2000. F. Schull, J. Carver, and G. Travassos, An empirical method- ology for introducing software processes, 8th Eur Soft- ware Eng Conf and 9th ACM SIGSOFT Symp Foundations of Software Engineering, Vienna, Austria, September 2001, pp. 288–296. Curt Wells provides process improvement and training services through his company, I-metrics LLC, and is an authorized lead appraiser for the FAA iCMM® . He is a co-author of the SE-CMM® , EIA/IS 731, CMMI V1, and FAA iCMM® V2. Curt retired from Lockheed Martin after 30 years service in various systems engineering and management roles. He holds a Masters Degree in Physics from Sam Houston State University. Linda Ibrahim joined FAA in 1996. She is FAA Chief Engineer for Process Improvement, providing technical leadership for FAA-wide process improvement. She led development of FAA-iCMM® v1.0 and v2.0, and its appraisal method. Linda has been working in software engineering for over 30 years, as practitioner, educator, and researcher; in the United States, Europe, and Middle East. She worked at the Software Engineering Institute for several years, and is a member of the CMMI Steering Group. Linda holds a B.A. in Mathematics, an M.S. in Information Science, and a Ph.D. in Electrical Engineering. She is a member of INCOSE, ACM, and IEEE. Larry LaBruyere has been with Northrop Grumman Mission Systems for over 20 years and supports the Federal Aviation Administration’s Chief Information Officer on a technical assistance contract supporting process improvement based on the FAA-iCMM® . He is the appraisal coordinator and an FAM appraisal lead for the FAA and participated in the authoring of the FAA-iCMM® and FAM. Mr. LaBruyere is an “Authorized CBA IPI Lead Assessor” for the SW-CMM® and authorized Standard CMMI Appraisal Method for Process Improvement (SCAMPI) Lead Appraiser for the CMMI. He isanauthorizedinstructor for Practical Software and Systems Measurement. He obtained an M.S. from American University and a B.A. from Anderson University. 308 WELLS, IBRAHIM, AND LABRUYERE