37
8/21/2006 Version 10 Code of Best Practice for Joint Experimentation 1 Technical Committee on Information Technical Committee on Information and Command & Control Systems and Command & Control Systems Code of Best Practice for Joint Experimentation

Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation1

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Code of Best Practice forJoint Experimentation

Page 2: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation2

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Agenda

• Overview• Specifics• Piggybacking• Duck Bites• The Pay Off• Summary

Page 3: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation3

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Background

• AIAA Task Force formed May 1998 to develop a Code of Best Practice (COBP) for Joint Experimentation with focus on Information Superiority

• Activities to date– Quarterly meetings focused on experimentation sites and issues– ISX 1.1: Assessment of process and lessons learned within

EFX 98– JEFX 99 (Assessment methodology)

JEFX 00 (Concept development)– Modeling and Simulation in Joint Experimentation

Workshop (December 98)– Support for Military Operations Research Society (MORS) Mini-

Symposium & Workshop (March 99) on Joint Experimentation

Page 4: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation4

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Definitions

• Experiment - To determine the efficacy of something previously untried (hypotheses generation), to examine the validity of a hypothesis (hypotheses testing), or to demonstrate a known truth (demonstration). Experiments are always empirical (involve systematic observation and measurement). Experiments involve three phases: Pre- , Conduct and Post-

• Experimental Campaign - A series of related activities that explore and mature knowledge about a concept of interest (voyage of discovery).

• Experimental Venue - An integrated construct designed to support individual initiatives.

Page 5: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation5

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Experimental Space

Imm

a tur

eM

atur

e

Maturity of Knowledge Domain

Demonstration

Refined Hypothesis

Preliminary Hypothesis

Discovery

Complexity(Multi-dimensional)

Interaction

Echelon

Functional

Simple Complex

Low Fide

lity

High Fide

lity

Brainstorming

Expert Elicitation

Wargaming

M&S (constructive,virtual)

Live Exercises

Real World Lessons Learned

Tools

DomainStructure

Campaign Vector

InformationFlow

Stability ofEnvironment

Number of ActorsUncertainty

Page 6: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation6

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Experimental Space (con’t)

Imm

a tur

eM

atur

e

Low Fide

lity

High Fide

lity

Maturity of Knowledge Domain

Complexity

Tools

Simple Complex

Experimental Activities Withina Campaign Plan

Page 7: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation7

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Maturity of Knowledge Domain

• Demonstration Experiments– Predictive models– Estimates values of some factors given values of others

• Preliminary and Refined Hypothesis Experiments– Explanatory models– Models cause and effect– Explains how different factors interact

• Discovery Experiments– Conceptual models– Identify important factors– Enables process classification Immaturity

Maturity

Page 8: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation8

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Types of Tools

• “Soft” Operational Research Tools– Brainstorming– Expert elicitation– Wargaming

• “Solving” Tools– Linear programming– Heuristic searching

• Modeling and Simulation– Constructive, virtual tools– Live events

• Real World Lessons Learned• Supporting Tools

– Data analysis– Visualization– Data mining

Page 9: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation9

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Dimensions of Complexity

FFaaccttoorr

•• NNuummbbeerr ooffAAccttoorrss

FFeeww((HHoommooggeenneeoouuss))

MMaannyy((HHeetteerrooggeenneeoouuss))

•• SSttaabbiilliittyy ooffEEnnvviirroonnmmeenntt

SSttaattiicc DDyynnaammiicc

•• EEcchheelloonn OOnnee MMaannyy

•• FFuunnccttiioonn SSiinnggllee MMuullttiippllee

•• IInnffoorrmmaattiioonnFFllooww

SSttoovveeppiippeedd NNeettwwoorrkkeedd

•• IInntteerraaccttiioonn NNoonn--iinntteerraaccttiivvee FFuullllyy IInntteerraaccttiivvee

•• DDoommaaiinnSSttrruuccttuurree

PPoooorrllyy DDeeffiinneedd WWeellll DDeeffiinneedd

•• UUnncceerrttaaiinnttyy LLooww HHiigghh

Simple Complex

Page 10: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation10

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Campaign Plan

• Purpose: Explore and mature a concept• Goal: Develop requirements for Mission Capability

Packages (MCPs) {e.g. DOTMLP plus other considerations}• Concept: Several linked activities designed to transit

the experimental space– Some concept explorations will share experimental events– Thematic experimental activities may generate

concepts for specific exploration

• Best Practice and Guidelines– Must parse the key concept into meaningful parts– Develop structure that insures capabilities for valid generalization

• Balance of control and live play• Baselines and control cases• Minimize intrusions

Information

Mission CapabilityMission CapabilityPackagePackage

DDoctrineoctrineOOrganizationrganizationTTraining & Educationraining & EducationMaterialMaterialLLeadershipeadershipPPeopleeople+ plus other considerations+ plus other considerations

Page 11: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation11

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Creation of a Mission Capability Package

CO-EVOLUTION

Organization

Command ArrangementsCONOPS/Doctrine

C4ISR SystemsTraining/Education

Perso

nnel

Weapons Systems

Logistics

FieldedMission

CapabilityPackage

E X P E R I M E N T A T I O NDoctrine Development

Command Reorganization

Education

Training Wargaming

Analysis

Systems Development

Simulations

Models

ATDsCommand and ControlForce Structure

Doctrine

Technology Requirements

Mission

Technology

EmbryonicMCPIdeas

Page 12: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation12

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Developed andRefined Military

Capability

Campaign Plan Progression

Concept(High Risk)

DemonstratedCapability(Low Risk)

ConceptMaturityMaturity ProcessProcess ProductsProducts

Preliminary Hypothesis

Discovery/Experiment

Refined Hypothesis

Demonstration

•Research &DevelopmentPriorities

•JROC Acquisition

•Military CapabilityPackageRequirements

Page 13: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation13

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Balanced Experimentation

Pre

Pre --

Expe

rimen

tEx

perim

ent

Experimen t

Exp erimen t

PostPost --Experim

entExperim

ent

Page 14: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation14

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

PrePre--Experiment 1Experiment 1

PostPost--Experiment 3Experiment 3

Campaign Plan Concept:Linked Experiments

•Pre- and Post-Experimental activities largely depend on analytic tools to include constructive, virtual, and M&S)

•Knowledge gained is often greatest in Post- phase but quality of that knowledge depends heavily on the Pre-Phase preparation and planning

Exp

erim

ent

1E

xper

imen

t 1

PrePre--Experiment 2Experiment 2

Exp

erim

e nt

2E

x pe r

ime n

t 2

Exp

erim

ent

3E

xper

imen

t 3

PostPost--Experiment 1Experiment 1

Learning&

Knowledge

PrePre--Experiment 3Experiment 3

PostPost--Experiment 2Experiment 2

KnowledgeBase

Page 15: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation15

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

COBP Milestone Timeline

PrePre--ExperimentExperimentPhasePhase

ExperimentExperimentPhasePhase

PostPost--ExperimentExperimentPhasePhase

Time in % of Overall Experimental Activity

Plan the experiment

StabilizeExp. Infrastructure

Initiate trainingRehearse and pre-test

Conduct the experiment

Conduct supporting M&S

Ensure quality control

Provide interim assessments & hotwashes

Analyze data

Generate products

Review products

0 5025 10075

Define objectives & concepts,establish experimental task force

Disseminate products

Good Idea Cut-off Date

Page 16: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation16

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Balanced Experimentation Phases of Activity

• Pre-Experiment: Determines Potential for Success– Sharpens focus– Assesses feasibility– Maximizes success in the “Real World”

• Experiment: Enables Success– Emphasizes credibility– Increases visibility– Generates data

• Post-Experiment: Explores Success, Consolidates Gains

– Extends data and assesses sensitivity– Translates data to information/knowledge– Prepares and focuses next experiment– Feeds other communities

(i.e. Acquisition, Doctrine, Training)

MOREMORE

Page 17: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation17

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Pre-Experiment Phase (1 of 4)

• Define objective of the experiment• Develop and refine the concept(s)• Identify one or more hypotheses to be tested/characterize variables• Establish context of the hypotheses

– Formulate scenarios and interesting parts of the scenario space– Identify ranges of the variables

• Assess feasibility and expected utility of experiment by estimating:– Costs– Other resources required (e.g., personnel, facilities)– Expected outcomes– Risks

• Execute appropriate models to:– Establish parameters– Explore the scenario space– Test assumptions

16a

Page 18: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation18

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

• Create Plan for Experimentation– Specify null hypotheses– Develop operational definitions & measures– Identify training requirements (for participants, data collectors, etc.)– Define learning objectives– Develop data collection and data analysis plans– Specify required control and adaptation during the experiment

• Specify the Experimental Structure– Dependent variables (objective functions)– Independent variables

• Experimental focus• Context and conditions

• Identify infrastructure needs– Assure adequacy of models, simulations and tools– Define facility requirements

The Pre-Experiment Phase (2 of 4)

16b

Page 19: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation19

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Pre-Experiment Phase (3 of 4)

• Develop configuration management plans• Specify appropriate Measures of Merit

– Measures of Performance (MOPs)– Measures of C2 Effectiveness (MOEs)– Measures of Force Effectiveness (MOFEs)– Measures of Policy Effectiveness (MOPEs)

• Specify other observables of interest– Meaningful scale points

• Refine data collection and data analysis plans

16c

Page 20: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation20

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

• Create experimental environment– Provide infrastructure– Create scenario– Install data driver– Install models, C3I– Integrate and federate models– Instrument for data collection– VIP interface

• Select and train participants– Identify contingencies that training should address– Tailor selection, training and assessment proficiency for:

• Friendly and coalition forces(blue)• OPFOR (red) and neutral (gray)• Control(white team)• Data collection and analysis team• Quality control team• VIP interface

• Rehearse / Pretest

The Pre-Experiment Phase (4 of 4)

16d

Page 21: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation21

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Experiment Phase (1 of 3)

• Maintain the experimental infrastructure• Run the scenario

– Simulate non-live effects• Insert sensor inputs, enemy responses, system degradation, operating

environment • Stimulate, represent, predict, and generate effects

– Maintain consistent time and granularity– Ensure replicability

• Ensure valid live effects• Conduct supporting modeling or simulation activities

– Conduct sanity check/quality control– Run additional base-lining with same inputs– Play roles– Step through control variables– Iterate as part of the plan

16e

Page 22: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation22

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Experiment Phase (2 of 3)

• Execute data collection plan– Collect data in accordance with plan– Identify insights

• Ensure quality control– Sample representative observations – Ensure continuous collection

• Maintain discipline– Maintain the integrity of the experiment– Control the scenario– Apply exit & restart criteria– Manage anomalies – Document changes/anomalies– Manage experiment– Manage configuration

16f

Page 23: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation23

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Experiment Phase (3 of 3)

• Hold interim assessments– Qualitative– Quantitative– Evaluate utility of process– Identify anomalies – Use the “hotwash” to inform the iterative campaign process and enhance

quality– Visualize data to determine adjustments to experiment– Perform intra-experiment adjustment– Capture insights

• Prompt “hotwash”– Identify experimental process insights & lessons recorded– Develop major findings based on available data and preliminary

observations– Identify post-experimental issues– Organize for experimental report

16g

Page 24: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation24

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Post-Experiment Phase (1 of 2)

• Execute data analysis plan• Conduct additional data analyses

– Sensitivity analysis– Extrapolation beyond observed ranges & conditions– Explore anomalies/insights– Data mining (as appropriate)

• Identify critical issues• Generate products

– Use M&S to clarify ambiguous results– Develop comprehensive reports and briefings– Conduct peer review– Promulgate results

16h

Page 25: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation25

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

The Post-Experiment Phase (2 of 2)

• Refine/improve M&S– Conceptual – Executable

• Exploit experimental results– Build knowledge base– Ensure that “lessons recorded” become “lessons learned”

16i

Page 26: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation26

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Code of Best Practice Insights

To maximize probability of success and knowledge gains: • Establish experimental concept credibility early• Bound the problem efficiently• Gain high level visibility and leverage• Achieve a formal agreement to collaborate• Ensure sufficient influence on scenario to tailor it• Obtain early involvement in the planning process (e.g., Stakeholders)• Capture sufficient resources for training, data collection, and analysis• Create a robust experimentation environment, including adequate:

– Number and variety of experimental events– Instrumentation– Free play (opportunities to fail)

17

Page 27: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation27

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Piggybacking Experiments May Seem Like a Good Idea...

ServiceServiceExperimentExperiment

JointJointExperimentExperiment

18

Page 28: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation28

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Risks of Piggybacking

• Loss of Control– Early involvement in planning process necessary to influence scenario

building & free play (opportunity to fail)– Very detailed backup planning necessary to cope with unplanned events,

such as:• Adverse weather impacts• Experimental infrastructure problems

• Lack of Visibility– Piggy back event needs 1everage on main experiment planning and

conduct • Lack of Credibility

– Hypothesis may not be tested adequately– All participants need to collaborate (formally & informally) to ensure

• Sufficient data collection opportunities• Main experiments and piggy backing events do not confound or contradict one

another

19

Page 29: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation29

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

…But it Does Have Some Risks

Service ExperimentService Experiment

JointJointExperimentExperiment

20

Page 30: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation30

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

A Win-Win SituationKeith Kirkland:This is based on

Dr. Starr’s final outbrief on behalf

of the MORS Synthesis

committee

Keith Kirkland:This is based on This is based on

Dr. Dr. StarrStarr’’s s final final outbrief outbrief on behalf on behalf

of the MORS of the MORS Synthesis Synthesis

committeecommittee

Proper Utilization of anExperimental CampaignPlan Results in aWin-Win Situation

Proper Utilization of anProper Utilization of anExperimental CampaignExperimental CampaignPlan Results in aPlan Results in aWinWin--Win SituationWin Situation

ExperimentalExperimentalCampaign PlanCampaign Plan

ServiceServiceExpExp..

JointJointExpExp..

21

Page 31: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation31

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Duck Bites Can Kill

Planning

PreparationExecution Data

Collection

Administrative Actions

MOREMORE

22

Page 32: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation32

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

ISX Lessons Learned(1 of 4)

• Planning – Limit the number of initiatives– Maintain tighter control of the Master Scenario Event List (MSEL)– Institute configuration control on hardware, software for all

systems in an experiment– Select a range with sufficient air, ground, space– Schedule ISR assets with sufficient slack time– Do not conduct offensive information operations if systems are

fragile

22a

Page 33: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation33

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

ISX Lessons Learned (2 of 4)

• Preparation– Indoctrinate the players about the objective of the experiment– Schedule adequate training on experimental systems, CONOPS– Check out communication completely prior to start of live fly

activities– Establish and maintain improved time synchronization

• Execution – Synchronize live and simulated events– Maintain an equipment status board– Establish a white cell to make timely decisions

Page 34: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation34

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

ISX Lessons Learned (3 of 4)

• Data Collection– Provide sufficient data collection tools (e.g., collaborative virtual

workstations (CVWs))– Install, check-out all data collection equipment at least 3 days prior

to live fly– Train data collectors/assessors on the experimental process,

simulation tools employed – Debrief air crews using assessors on the experimental process,

simulation tools employed– Debrief air crews using assessors (focusing on experiment

objectives)– Select data collectors/Assessors with the proper backgrounds,

familiarity with key systems– Develop rapport between the players and data collectors/assessors

Page 35: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation35

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

ISX Lessons Learned (4 of 4)

• Administration– Schedule Live Fly Events in times when weather is traditionally

good– Don’t schedule critical targeting events during lunch hour– Provide assistance with classification issues, procedures– Enforce discipline in issuing badges– Impose greater control on visitiors – Publish a telephone directory for experimental participants

Page 36: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation36

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Impact of the COBP

Current Practice• Pre-

– Flawed due to:• Lack of qualified/experience

personnel• Lessons recorded are not lessons

learned– Rushed– May required additional resources

• Experiment– Receives all the attention & the

bulk of resources– Demonstrations mentality

overshadows knowledge collection– Inadequate/primitive data capture

capability– Overemphasis on hotwash - not on

sound analysis• Post-

– Short changed; data not exploited or archived

– Full promise not realized– Some lessons remain unrecorded

COBP• Pre-

– Appropriate time& resource allocation

– Correct expertise

• Experiment– Put “The Show” in proper

perspective

– Emphasizes knowledge gained

• Post-– Provides more resources– Provides more time

Results

23

Page 37: Code of Best Practice for Joint Experimentation · 4 8/21/2006 Code of Best Practice for Joint Experimentation Version 10 Technical Committee on Information and Command & Control

8/21/2006 Version 10Code of Best Practice for Joint Experimentation37

Technical Committee on InformationTechnical Committee on Informationand Command & Control Systemsand Command & Control Systems

Conclusions

• Best Practice for Joint Experimentation requires:– An experimental campaign plan that will enable us to explore and mature

concepts from idea to Mission Capability Packages (MCPs)– Proper emphasis on Pre-Experiment Phase to enhance potential for

success– Disciplined conduct of actual experiment to enhance knowledge gain and

feed subsequent Experimental Events– Proper Post-Experiment exploitation to enhance knowledge gained and

feed follow-on Pre-Experiment phase

Proper use of Code of Best Practice can yield greatly enhanced Proper use of Code of Best Practice can yield greatly enhanced learning.learning.