Upload
others
View
2
Download
0
Embed Size (px)
Citation preview
Quantifying the Effectiveness of Systems Engineering
Presenters: Joseph P. ElmDr. Dennis Goldenson
Software Engineering Institute
Copyright 2013Carnegie Mellon University and IEEEThis material is based upon work funded and supported by the Department of Defense under Contract No. FA8721-05-C-0003 with Carnegie Mellon University for the operation of the Software Engineering Institute, a federally funded research and development
tcenter.NO WARRANTY. THIS CARNEGIE MELLON UNIVERSITY AND SOFTWARE ENGINEERING INSTITUTE MATERIAL IS FURNISHED ON AN “AS-IS” BASIS. CARNEGIE MELLON UNIVERSITY MAKES NO WARRANTIES OF ANY KIND, EITHER EXPRESSED OR IMPLIED, AS TO ANY MATTER INCLUDING, BUT NOT LIMITED TO, WARRANTY OF FITNESS FOR PURPOSE OR MERCHANTABILITY, EXCLUSIVITY, OR RESULTS OBTAINED FROM USE OF THE MATERIAL. CARNEGIE MELLON UNIVERSITY DOES NOT MAKE ANY WARRANTY OF ANY KIND WITH RESPECT TO FREEDOM FROM PATENT, TRADEMARK, OR COPYRIGHT INFRINGEMENT.This material has been approved for public release and unlimited distribution.The Government of the United States has a royalty-free government-purpose license to use, duplicate, or disclose the work, in whole or in part and in any manner, and to have or permit others to do so, for government purposes pursuant to the copyright license under the clause at 252.227-7013 and 252.227-7013 Alternate I.This material may be reproduced in its entirety without modification and freely distributed in written or electronic form withoutThis material may be reproduced in its entirety, without modification, and freely distributed in written or electronic form without requesting formal permission. Permission is required for any other use. Requests for permission should be directed to the Software Engineering Institute at [email protected]® is registered in the U.S. Patent and Trademark Office by Carnegie Mellon University.DM-0000243
2Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
The Value of System EngineeringGAO-09-362T - Actions Needed to Overcome Long-standing Challenges with Weapon Systems Acquisition and Service Contract Management • “costs … of major defense acquisition programs increased 26 percent and
development costs increased by 40 percent from first estimates”• “programs … failed to deliver capabilities when promised—often forcing p g p p g
warfighters to spend additional funds on maintaining legacy systems” • “current programs experienced, on average, a 21-month delay in delivering
initial capabilities to the warfighter”
Why?
“… managers rely heavily on assumptions about system i t t h l d d i t it hi hrequirements, technology, and design maturity, which are
consistently too optimistic. These gaps are largely the result of a lack of a disciplined systems engineering analysis prior
to beginning system development
3Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
to beginning system development …
Showing the Value of SE:The 2012 SE Effectiveness StudyThe 2012 SE Effectiveness StudyPerformed by NDIA, IEEE-AESS, and SEI
MethodMethod• Contact development programs using the
resources of NDIA, AESS, and INCOSE• Survey programs to assess their:Survey programs to assess their:
– SE activities– Program performance (cost, schedule, technical)– Degree of challenge
• Analyze responses for statistical relationships between assessed data
Survey Tenets• All data submitted anonymously• All data submitted anonymously
– necessary to collect proprietary data and promote truthful responses • All data handled confidentially by the SEI• Only aggregated data is released (no ID of person program or organization)
4Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
• Only aggregated data is released (no ID of person, program, or organization)
Artifact-based assessment of SE Practices
• 14 Process Areas• 31 Goals• 87 Practices• 199 Work Products
SystemsCMMI-SE/SW/IPPD v1.1
• 25 Process Areas• 179 Goals• 614 Practices
SystemsEngineering-related Filter
614 Practices• 476 Work Products
• 13 Process Areas• 23 Goals
4
Size Constraint Filter
Considered significant • 45 Practices• 71 Work Products
to Systems Engineering
Survey content is based on a recognized standard (CMMI)
5Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Survey content is based on a recognized standard (CMMI)
Assessment of Program Performance
Assess TOTAL Program Performance• Program Cost Program Schedule Technical PerformanceProgram Cost, Program Schedule, Technical Performance• Focus on commonly used measurements
– EVMS, baseline management– requirements satisfactionq– budget re-baselining and growth– milestone and delivery satisfaction
Assessment of Other Factors• Program Challenge – some programs are more complex than others
• Prior Experience – some acquirers are more capable than others
6Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Study Participants
Participant Solicitation• Contacted key members of major defense
t t t t t d ti i ticontractors to promote study participation• Contacted the memberships of NDIA SE Division,
IEEE AESS, and INCOSE
Collected 148 valid responsesCollected 148 valid responses
116120140
Which of these best describes your industry or service?
130140
Please enter the country in which most of the design and development engineering will be/was
performed.
7 2 0 2 9 0 1 100
20406080
100
406080
100120
Ind.
Mfg
& S
vc:
defe
nse
Ind.
Mfg
& S
vc:
Ele
ctro
nic…
nd. M
fg a
nd S
vc:
Oth
er
Tran
spor
tatio
n
Ene
rgy
Com
mun
icat
ions
nsum
er G
oods
&
Svc
Hea
lth C
are
Oth
er
6 3 2 2 1 1 1 10
2040
US
A
UK
Sou
th
Afri
ca
Aus
tralia
Can
ada
Indi
a
The
Net
herla
n…
Sw
eden
Finl
and
7Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
I
Co N
Study Results
60
Total program contract valueStd. Dev.
Mean
Median
488 M$2.22 B$50.5 M$ 60
Program Performance(Perf)
3.033.583.98
Median (2nd quartile)1st quartile
3rd quartile
10
20
30
40
50
10
20
30
40
50
0
10
100 K$ 1 M$ 10 M$ 100 M$ 1 B$ 10 B$ 100 B$0
10
1 1.5 2 2.5 3 3.5 4 4.5 5
P Ch ll 2 221 t tilT t l SE D l d 1 t til
506070
Program Challenge 2.222.502.68
Median (2nd quartile)1st quartile
3rd quartile
40
50
60
Total SE Deployed onProgram (SEC_Total)
2.783.033.41
Median (2nd quartile)1st quartile
3rd quartile
010203040
1 1 5 2 2 5 3 3 5 40
10
20
30
1 1 5 2 2 5 3 3 5 4
8Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
1 1.5 2 2.5 3 3.5 41 1.5 2 2.5 3 3.5 4
The Bottom Line
Across ALL programs, 1/3 are at each performance level15% performance level
For Lower SECprograms, only 15%deliver higher
33%
47%
15% 24%
56%
performance
For Middle SECprograms, 24% deliver higher performance
52%29% 20%
24%
higher performance
For Higher SEC programs, 57% deliver higher performanceg p
Gamma = 0.49 represents a VERY STRONG relationship
Gamma = 0.49 p-value < 0.001
9Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
The Effect of Program Challenge
100%
Perf vs. SEC_Total (Low PC)
%
Perf vs. SEC_Total (High PC)
23% 23%52%
60%70%80%90%
100%
23%
%
8%26%
62%60%70%80%90%
100%
32%
45%58%
36%20%30%40%50%60%
69%39%
35%
12%20%30%40%50%60%
32%19% 12%
0%10%
Lower SEC (n=22)
Middle SEC (n=26)
Higher SEC (n=25)
39% 27%0%
10%
Lower SEC (n=26)
Middle SEC (n=23)
Higher SEC (n=26)
Gamma = 0.34 p-value = 0.029 Gamma = 0.62 p-value = 0.000
A STRONG relationship between Total SE and Program Performance for
A VERY STRONG relationship between Total SE and Program Performance for
10Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
LOWER CHALLENGE programs HIGHER CHALLENGE programs
A Deeper Look at SE Activities
Our survey questions addressed 11 areas of SE Activities• Program Planning• Requirements Development and Management• Product Architecture• Trade Studies• Product IntegrationProduct Integration• Verification• Validation• Risk Management• Configuration Management• Integrated Product Teams• Program Monitoring and Control
This enabled us to assess a program’s deployment of SE in each of these areas
11Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Program Planning vs. Performance
13%100%
Perf vs. SEC-PP
Higher 19%10%
30%80%
100%Perf vs. SEC-PP (High PC)
33%
42%
34%50%
40%
60%
80%Higher Perf
Middle Perf
71%
30% 24%
19%
39%
10%
30%67%
20%
40%
60%
80%
54%
24% 22%
28%
0%
20%
40%
Lower SEC (n 48) Middle SEC Higher SEC All
Lower Perf
Perf vs SEC PP (Low PC)
30% 24%0%
Lower SEC (n=31)
Middle SEC (n=23)
Higher SEC (n=21)
Gamma = 0.65 p-value = 0.000
Lower SEC (n=48) Middle SEC (n=50)
Higher SEC (n=50)
Gamma = 0.46 p-value = 0.000
All
59%44% 41%
18%37% 38%
40%
60%
80%
100%Perf vs. SEC-PP (Low PC)
24% 19% 21%
44% 41%
0%
20%
40%
Lower SEC (n=17)
Middle SEC (n=27)
Higher SEC (n=29)
G 0 16 l 0 313
The relationship:for the set of all programs 0.46 = Very Strongfor the set of High Challenge programs 0.65 = Very Strongfor the set of Low Challenge programs 0 16 = Weak
12Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.16 p-value = 0.313for the set of Low Challenge programs 0.16 = Weak
Requirements Dev’t & Mg’t vs. Performance
21% 18%100%
Perf vs. SEC-REQ
Higher 18% 21%100%
Perf vs. SEC-REQ (High PC)
29%52%
21% %
58%
40%
60%
80%Higher Perf
Middle Perf 61%
46%26%
21%33%
13%
61%
20%
40%
60%
80%
50%30% 20%
22%
0%
20%
40%
All
Lower Perf
26%0%
Lower SEC (n=28)
Middle SEC (n=24)
Higher SEC (n=23)
Gamma = 0.5 p-value = 0.001
Lower SEC (n=48) Middle SEC (n=50)
Higher SEC (n=50)
Gamma = 0.44 p-value = 0.000
All
40% 69%
25% 15%56%
60%
80%
100%Perf vs. SEC-REQ (Low PC)
35%15% 15%
69%
30%
0%
20%
40%
Lower SEC (n=20)
Middle SEC (n=26)
Higher SEC (n=27)
G 0 36 0 01
The relationship:for the set of all programs 0.44 = Very Strongfor the set of High Challenge programs 0.50 = Very Strongfor the set of Low Challenge programs 0 36 = Strong
13Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.36 p-value = 0.017for the set of Low Challenge programs 0.36 = Strong
A Deeper Look at SE Activities
Our survey questions addressed 11 areas of SE Activities• Program Planning• Requirements Development and Management• Product Architecture• Trade Studies• Product Integration
“Early” SE
Product Integration• Verification• Validation• Risk Management• Configuration Management• Integrated Product Teams• Program Monitoring and Control
This enabled us to assess a program’s deployment of SE in each of these areas
14Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Early SE is the MOST Important
26%6%
35%80%
100%
Perf vs. EarlySE (High PC)
12%32%
90%100%
Perf vs. EarlySE
Higher
68%43%
26%
22%
19%
35%67%
20%
40%
60%
80%
34%
40%
32%56%
40%50%60%70%80% Perf
Middle Perf
14%0%
Lower Early SE (n=31)
Middle Early SE (n=23)
Higher Early SE (n=21)
Gamma = 0.69
Perf vs EarlySE (Low PC)
54%28%
16%
29%
0%10%20%30%40%
Lower Early SE Middle Early SE Higher Early SE All
Lower Perf
47%53%
21% 30%46%
60%
80%
100%
Perf vs. EarlySE (Low PC)Lower Early SE (n=50)
Middle Early SE (n=53)
Higher Early SE (n=45)
Gamma = 0.53
All
32% 17% 17%
53%38%
0%
20%
40%
Lower Early SE (n=19)
Middle Early SE (n=30)
Higher Early SE (n=24)
The relationship:for the set of all programs 0.53 = Very Strongfor the set of High Challenge programs 0.69 = Very Strongfor the set of Low Challenge programs 0 25 = Moderate G 0 25
15Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
for the set of Low Challenge programs 0.25 = Moderate Gamma = 0.25
Summary of Relationships
-0.3 -0.2 -0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7
Total SE
E l SE
Performance vs. SE Capability - All Projects
Early SE
Project Planning
Req'ts Dev't & Mg't
Verification
Product Architecture
Configuration Mg't
Trade Studies
Monitor & Control
Validation
Product Integration
Risk Management
I t P d t TInteg. Product Teams
Project Challenge
Prior Experience
16Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
StrongModerate Very StrongWeakModerate
Summary of Relationships
-0.3 -0.2 -0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7
Total SE
E l SE
Performance vs. SE Capability - High Challenge
Early SE
Project Planning
Verification
Configuration Mg't
Monitor & Control
Req'ts Dev't & Mg't
Product Architecture
Validation
Trade Studies
Product Integration
Integ. Product Teams
Ri k M tRisk Management
Project Challenge
Prior Experience
17Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
StrongModerate Very StrongWeakModerate
Using the Findings of This StudySystem Developers can use this report to:
• plan SE capability improvement efforts focusing on those SE activities most strongly associated with improved program performance
• serve as an industry benchmark for their organization’s SE performance.– Assess programs within the organization and compare with the study results to leverage strengths, and improve
weaknesses
• justify and defend SE activities applied to programs.
System Acquirers may use this report to:• incorporate SE requirements into RFPs and source selection activities
– Ensure that SE activities are included in schedules and budgets– Demand SE deliverables (e.g. SE Management Plan) during program execution
Require SE evaluations of contractors during source selection and during program execution– Require SE evaluations of contractors during source selection and during program execution
• employ this survey or similar methods to collect data from during program execution as a means of identifying supplier SE deficiencies contributing to program risks.
SE Educators may use this report to:• Focus curricula on key aspects of SE• Convey to students the value of SE
All may use this report to:
18Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
• identify critical SE capabilities to guide Workforce Development
Call to Action
Download the 2012 report at http://www.sei.cmu.edu/library/abstracts/reports/12sr009.cfm
• Search for ways to apply the findings within your own work and your own organization
Help with the continuing effort of showing the value of SE
• Join the INCOSE SE Effectiveness Working Group– Go to http://www.incose.org/practice/techactivities/wg/seewg/
– Or contact Joseph Elm ([email protected])p (j @ g)
• Join the NDIA SE Effectiveness Committee– Go to
http://www.ndia.org/Divisions/Divisions/SystemsEngineering/Pages/Systems Engineering Effecp g y g g g y _ g g_tiveness_Committee.aspx
– Or contact Al Brown ([email protected])
19Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
For more information, contact:William F. LyonsIEEE-AESS Board of Governors
Alan R. BrownNDIA SE Effectiveness Committee Chair
[email protected] [email protected]
Joseph P. ElmSoftware Engineering Institutej l @ i d
Steve HenryNDIA SE Division Chair
h h @[email protected] [email protected]
Geoff DraperNDIA SE Division Vice Chairgdraper@harris com
Robert C. RassaNDIA SE Division Chair (emeritus)RCRassa@raytheon com
20Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
BACK UP
References
Elm, J.; Goldenson, D.; El Emam, K.; Donatelli, N.; Neisa, A. “A Survey of Systems Engineering Effectiveness – Initial Results”. Carnegie Mellon University; Pittsburgh PA 2007University; Pittsburgh, PA. 2007
Elm, J.; Goldenson, D. “The Business Case for Systems Engineering Study: Results of the Systems Engineering Effectiveness Survey”. Carnegie y g g y gMellon University; Pittsburgh, PA 2012 (available at http://www.sei.cmu.edu/library/abstracts/reports/12sr009.cfm)
Gruhl W “Lessons Learned Cost/Schedule Assessment Guide” InternalGruhl, W. Lessons Learned, Cost/Schedule Assessment Guide Internal Presentation (unpublished), NASA Comptrollers office. 1992
Honour, E., “Systems Engineering Return on Investment” PhD thesis, Defence and Systems Institute, University of South Australia. 2013 http://www.hcode.com/secoe
22Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Verification vs. Performance
16% 24%100%
Perf vs. SEC-VER
Higher 19%10%
26%100%
Perf vs. SEC-VER (High PC)
39%38%
24%
54%
40%
60%
80%Higher Perf
Middle Perf
71%48%
22%
19%
26%
22%
26%56%
20%
40%
60%
80%
45% 38%19%
28%
0%
20%
40%
L SEC ( 44) Middl SEC Hi h SEC All
Lower Perf
22%0%
Lower SEC (n=21)
Middle SEC (n=27)
Higher SEC (n=27)
Gamma = 0.6 p-value = 0.000
Lower SEC (n=44) Middle SEC (n=50)
Higher SEC (n=54)
Gamma = 0.43 p-value = 0.000
All
57% 52%
22% 22%52%
60%
80%
100%Perf vs. SEC-VER (Low PC)
22% 26% 15%
57%33%
0%
20%
40%
Lower SEC (n=23)
Middle SEC (n=23)
Higher SEC (n=27)
G 0 2 0 084
The relationship:for the set of all programs 0.43 = Very Strongfor the set of High Challenge programs 0.60 = Very Strongfor the set of Low Challenge programs 0 27 = Moderate
23Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.27 p-value = 0.084for the set of Low Challenge programs 0.27 = Moderate
Architecture vs. Performance
16%100%
Perf vs. SEC-ARCH
Higher8%
35%100%
Perf vs. SEC-ARCH (High PC)
36%
35%
31%49%
40%
60%
80%Higher Perf
Middle Perf 63%
46%28%
29%
19%20%
35%52%
20%
40%
60%
80%
49%33%
18%
33%
0%
20%
40%
L SEC ( 45) Middl SEC Hi h SEC All
Lower Perf
28%0%
Lower SEC (n=24)
Middle SEC (n=26)
Higher SEC (n=25)
Gamma = 0.49 p-value = 0.001
Lower SEC (n=45) Middle SEC (n=54)
Higher SEC (n=49)
Gamma = 0.41 p-value = 0.000
All
43% 50%
24% 29%46%
60%
80%
100%Perf vs. SEC-ARCH (Low PC)
33% 21% 8%
50%46%
0%
20%
40%
Lower SEC (n=21)
Middle SEC (n=28)
Higher SEC (n=24)
G 0 31 l 0 051
The relationship:for the set of all programs 0.41 = Very Strongfor the set of High Challenge programs 0.49 = Very Strongfor the set of Low Challenge programs 0 31 = Strong
24Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.31 p-value = 0.051for the set of Low Challenge programs 0.31 = Strong
Integrated Product Teams vs. Performance
22%100%
Perf vs. SEC-IPT
Higher14% 31%80%
100%
Perf vs. SEC-IPT (High PC)
41%35% 27%
22%35% 42%
40%
60%
80%Higher
Perf
Middle Perf 57%
38% 38%
29%31%
5%
31%57%
20%
40%
60%
80%
37% 31% 31%0%
20%
40%
Lower SEC(n=51) Middle SEC (n=52) Higher SEC (n=45) All
Lower Perf
Perf vs SEC-IPT (Low PC)
0%Lower SEC
(n=28)Middle SEC
(n=26)Higher SEC
(n=21)Gamma = 0.4 p-value = 0.007
Lower SEC(n=51) Middle SEC (n=52) Higher SEC (n=45)
Gamma = 0.18 p-value = 0.101
All
57% 38% 46%
30% 38% 29%
40%
60%
80%
100%
Perf vs. SEC-IPT (Low PC)
13% 23% 25%
57% 38% 6%
0%
20%
40%
Lower SEC (n=23)
Middle SEC (n=26)
Higher SEC (n=24)
Gamma = 0 12 p value = 0 436
The relationship:for the set of all programs 0.18 = Weakfor the set of High Challenge programs 0.40 = Strongfor the set of Low Challenge programs -0 12 = Weak Neg
25Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = -0.12 p-value = 0.436for the set of Low Challenge programs -0.12 = Weak Neg.
Risk Management vs. Performance
24% 29%
100%
Perf vs. SEC-RSKM
Higher 28% 22%%80%
100%
Perf vs. SEC-RSKM (High PC)
38% 36%30%
24% 29%43%
40%
60%
80% Perf
Middle Perf 55%
39% 39%
17% 39%18%
28%43%
20%
40%
60%
80%
38% 36% 26%
30%
0%
20%
40%
Lower SEC (n=50) Middle SEC (n=45) Higher SEC (n=53) All
Lower Perf
Perf vs SEC-RSKM (Low PC)
0%Lower SEC
(n=29)Middle SEC
(n=18)Higher SEC
(n=28)Gamma = 0.24 p-value = 0.124
Lower SEC (n=50) Middle SEC (n=45) Higher SEC (n=53)
Gamma = 0.21 p-value = 0.05
All
67% 33%
19% 33% 44%
40%
60%
80%
100%
Perf vs. SEC-RSKM (Low PC)
14%33%
12%
44%
0%
20%
40%
Lower SEC (n=21)
Middle SEC (n=27)
Higher SEC (n=25)
G 0 18 l 0 256
The relationship:for the set of all programs 0.21 = Moderatefor the set of High Challenge programs 0.24 = Moderatefor the set of Low Challenge programs 0 18 = Weak
26Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.18 p-value = 0.256for the set of Low Challenge programs 0.18 = Weak
Trade Studies vs. Performance
13%100%
Perf vs. SEC-TRD
Higher 15%32%80%
100%Perf vs. SEC-TRD (High PC)
43%
34%
33%52%
40%
60%
80%
gPerf
Middle Perf 56% 50%
25%
30%18%
20%
32%55%
20%
40%
60%
80%
43% 33% 23%
25%
0%
20%
40%
Lower SEC (n=46) Middle SEC (n=58) Higher SEC (n=44) All
Lower Perf
Perf vs SEC-TRD (Low PC)
25%0%
Lower SEC (n=27)
Middle SEC (n=28)
Higher SEC (n=20)
Gamma = 0.43 p-value = 0.004
Lower SEC (n=46) Middle SEC (n=58) Higher SEC (n=44)
Gamma = 0.38 p-value = 0
All
63%50%
11%33%
50%
40%
60%
80%
100%Perf vs. SEC-TRD (Low PC)
26% 17% 21%
50% 29%
0%
20%
40%
Lower SEC (n=19)
Middle SEC (n=30)
Higher SEC (n=24)
G 0 29 l 0 062
The relationship:for the set of all programs 0.38 = Strongfor the set of High Challenge programs 0.43 = Very Strongfor the set of Low Challenge programs 0 29 = Moderate
27Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.29 p-value = 0.062for the set of Low Challenge programs 0.29 = Moderate
Validation vs. Performance
17% 27%
100%
Perf vs. SEC-VAL
Higher 36%0%
31%80%
100%Perf vs. SEC-VAL (High PC)
47% 34%
27%56%
40%
60%
80%gPerf
Middle Perf
64%49%
27%
36%
21%
18%
31%55%
20%
40%
60%
80%
36% 38%21%
23%
0%
20%
40%
Lower SEC (n=36) Middle SEC (n=73) Higher SEC (n=39) All
Lower Perf
Perf vs SEC VAL (Low PC)
27%0%
Lower SEC (n=14)
Middle SEC (n=39)
Higher SEC (n=22)
Gamma = 0.48 p-value = 0.002
Lower SEC (n=36) Middle SEC (n=73) Higher SEC (n=39)
Gamma = 0.33 p-value = 0.003
All
55% 50%
27% 24%59%
60%
80%
100%Perf vs. SEC-VAL (Low PC)
18% 26% 12%
55%29%
0%
20%
40%
Lower SEC (n=22)
Middle SEC (n=34)
Higher SEC (n=17)
G 0 23 l 0 127
The relationship:for the set of all programs 0.33 = Strongfor the set of High Challenge programs 0.48 = Very Strongfor the set of Low Challenge programs 0 23 = Moderate
28Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.23 p-value = 0.127for the set of Low Challenge programs 0.23 = Moderate
Product Integration vs. Performance
16%100%
Perf vs. SEC-PI
Higher 12%
33%80%
100%Perf vs. SEC-PI (High PC)
34%
38%26%
32%49%
40%
60%
80%
gPerf
Middle Perf
65%43% 33%
24%
25%17%
33%50%
20%
40%
60%
80%
50%30% 26%
26%
0%
20%
40%
Lower SEC (n=32) Middle SEC (n=81) Higher SEC (n=35) All
Lower Perf
Perf vs SEC PI (Low PC)
0%Lower SEC
(n=17)Middle SEC
(n=40)Higher SEC
(n=18)
Gamma = 0.42 p-value = 0.01
Lower SEC (n=32) Middle SEC (n=81) Higher SEC (n=35)
Gamma = 0.33 p-value = 0.003
All
47%51%
20% 32% 47%
40%
60%
80%
100%Perf vs. SEC-PI (Low PC)
33%17% 18%
51% 35%
0%
20%
40%
Lower SEC (n=15)
Middle SEC (n=41)
Higher SEC (n=17)
G 0 23 l 0 153
The relationship:for the set of all programs 0.33 = Strongfor the set of High Challenge programs 0.42 = Very Strongfor the set of Low Challenge programs 0 23 = Moderate
29Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.23 p-value = 0.153for the set of Low Challenge programs 0.23 = Moderate
Configuration Management vs. Performance
17%100%
Perf vs. SEC-CM
Higher 14%
35%80%
100%Perf vs. SEC-CM (High PC)
37%
33%
39% 47%
40%
60%
80%gPerf
Middle Perf
66%38% 25%
21%
27%20%
35%55%
20%
40%
60%
80%
46%27% 21%
32%
0%
20%
40%
Lo er SEC (n 59) Middle SEC (n 51) Higher SEC (n 38) All
Lower Perf
Perf vs SEC CM (Low PC)
25%0%
Lower SEC (n=29)
Middle SEC (n=26)
Higher SEC (n=20)
Gamma = 0.53 p-value = 0
Lower SEC (n=59) Middle SEC (n=51) Higher SEC (n=38)
Gamma = 0.38 p-value = 0.001
All
53%
20%44% 39%
60%
80%
100%Perf vs. SEC-CM (Low PC)
27% 16% 17%
40% 44%
0%
20%
40%
Lower SEC (n=30)
Middle SEC (n=25)
Higher SEC (n=18)
The relationship:for the set of all programs 0.38 = Strongfor the set of High Challenge programs 0.53 = Very Strongfor the set of Low Challenge programs 0 22 = Moderate
30Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.22 p-value = 0.203for the set of Low Challenge programs 0.22 = Moderate
Program Monitoring & Control vs. Performance
17%31%
100%
Perf vs. SEC-PMC
Higher 14% 26%80%
100%
Perf vs. SEC-PMC (High PC)
38%37%
31%50%
40%
60%
80%gPerf
Middle Perf
68%44%
27%
18%30%
19%
54%
20%
40%
60%
80%
46%33%
21%
29%
0%
20%
40%
Lower SEC (n=48) Middle SEC (n=52) Higher SEC (n=48) All
Lower Perf
Perf vs SEC-PMC (Low PC)
27%0%
Lower SEC (n=22)
Middle SEC (n=27)
Higher SEC (n=26)
Gamma = 0.53 p-value = 0
Lower SEC (n=48) Middle SEC (n=52) Higher SEC (n=48)
Gamma = 0.38 p-value = 0
All
54%44%
19%36% 45%
60%
80%
100%
Perf vs. SEC-PMC (Low PC)
27% 20% 14%
44%41%
0%
20%
40%
Lower SEC (n=26)
Middle SEC (n=25)
Higher SEC (n=22)
The relationship:for the set of all programs 0.38 = Strongfor the set of High Challenge programs 0.53 = Very Strongfor the set of Low Challenge programs 0 27 = Moderate
31Quantifying the Effectiveness of SE14-May-2013© 2013 Carnegie Mellon University
Gamma = 0.27 p-value = 0.092for the set of Low Challenge programs 0.27 = Moderate