auditing absval 9 july 2008 prepared for workshop iii: determining a validation methodology for...

27
Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis, Inc. 2001 North Beauregard Street Alexandria, Virginia 22311-1739 SPA Serial 50028-08

Upload: mae-robinson

Post on 21-Jan-2016

214 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

Auditing ABSVAL

9 July 2008

Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation

bySystems Planning and Analysis, Inc.

2001 North Beauregard Street Alexandria, Virginia 22311-1739

SPA Serial 50028-08

Page 2: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

SPA Role

• Review Measures of Validity

• Audit VV&A Applications Focus for Today

• Participate in Test Case Workshops

Page 3: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Approach to Auditing the Applications

• Play role of Accreditation Agent

• Focus on reports as a reflection of the process.– More a starting point than a constraint on scope.

• Formulate, more or less independently of the V&V process, a series of questions we would want answered about the model in support of an accreditation decision.

– We introduced these at workshop II, but will skip a couple based on feedback received at that time.

• For each of these questions, ask how clearly and convincingly the V&V report answers it, and how readily the answer is extracted from the report.

• Use the V&V reports to answer the second set of questions.

• Comment on both the reports and the process.

Page 4: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Scope of Review

• Pythagoras COIN: Two reports covering separate phases of V&V, as an artifact of the validation team structure.– “Assumption Testing Report” (Verification of Conceptual Model)– “Validation Report” (Validation of Instantiated Model)– These reports treated as a single V&V decision package.

• Pythagoras Obstacle Clearance Model– Validity Assessment Report

Page 5: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Scope of Validation Effort

P-COIN P-OCM

Assumption Testing

Model Verification

Results Validation

Determination

Assumption Testing

Model Verification

Results Validation

Determination

P-COIN Assumption Testing Report P-OCM Validity Assessment Report

Validation Report for the Pythagoras COIN Model

Page 6: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Summary

• For both P-COIN and P-OCM, we were able to use the information reported by the validation teams to identify a recommended accreditation decision and a convincing rationale underlying it.

• The accreditation recommendation was tentatively adverse for both applications, and neither test case exercised all parts of the ABSVAL process.

• Sufficient information to support a favorable accreditation recommendation for either application may not have been available to the V&V team

• ABSVAL response to challenges introduced by ABS, emergent behavior in particular, not really demonstrated by the test cases.

Page 7: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Review Questions

Page 8: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Do the reports clearly identify the application (set of study questions) for which the model is being validated, and the model’s role in addressing those questions?

Yes. Report states that the application is course-of-action analysis (afloat vs. ashore) and that we don’t necessarily need accurate point estimates of outcomes, but a reliable ranking of those point estimates under the alternatives.

Review Questions: P-COIN

Page 9: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Do the reports clearly describe the tests that were performed on the model, the possible outcomes for each test, and the criteria for passing?

It was reasonably easy to understand from the reports what tests (or questions) P-COIN was subjected to, once we found them. They were more clearly enumerated in the Assumption Testing report. In the Validation Report, we were able to infer the tests (which related to methodology) from observations presented in a section titled “Data Considerations.”

Review Questions: P-COIN

Page 10: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

For each test performed, is the result clearly presented in a way that relates directly to the specified acceptance criteria?

This question would be more applicable to results validation tests, of which none were available for our review. The tests performed were generally qualititative in nature.

When problems were found, the reports sometimes stopped short of assessing their implications for acceptance.

Review Questions: P-COIN

Page 11: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Do the reports provide a recommended decision for the accreditation authority?

Yes, two different ones!

– Assumption Testing: “…we have to advise against reporting out any results from the model as actionable in any sense” (sect. 6.5.4)

• Unambiguous, not repeated in recommendations summary

– Validation: “Subject to the caveats, limitations, and cautions listed above, P-COIN can answer (the analysis) question…” (sect. 8.3)

Also: “…there is little risk in using the results…since the analysis does not advocate a change in current…procedure…” (sect. 8.3)

• This “after the fact” risk assessment leaves open the question of whether the model is more useful than a magic 8-ball, and the remainder of the quoted sentence seems to acknowledge this.

• Can the model increase our confidence that the default COA is right?

We used Boolean logic along with our own assessment of the information provided to infer that yes and no means no, don’t use it in its current state to support the afloat vs. ashore decision.

Review Questions: P-COIN

Page 12: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Do the reports make a convincing argument that the tests conducted collectively provide a sufficient basis for the recommended accreditation decision?

Yes. The tests and results presented were sufficient to cause us to doubt the model’s ability to reliably support the afloat-vs-ashore decision at this moment, so the practical mandate (regarding the Accreditation decision) is fulfilled.

Supporting a positive accreditation recommendation, had that been “the right answer,” would probably have required much more work on the validation side.

Where problems were found in assumption testing, these were assessed not to invalidate the conceptual model. Although we believe in the value of assumption testing, we have to wonder how squarely ABSVAL is actually founded on challenging a null hypothesis of validity, given the appearance that effort was expended on zero-power tests.

– Is it at least possible in theory that a assumption could be shown to be bad enough to invalidate the CM?

Review Questions: P-COIN

Page 13: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Do the reports make clear what elements (parameters) were treated as part of the model and what elements were treated as data subject to variation?

For those parameters identified as data that might vary within the scope of the application for which accreditation is recommended (if it is recommended), has the V&V team identified a general approach to validating the data for applications other than the test case?

Do the reports appear to reflect an efficient plan of attack?

Review Questions: P-COIN

Page 14: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Are recommendations provided that are actionable by a model improvement program? (This question was added in response to feedback received in Workshop II. It is understood that these are not a required ABSVAL output)

Yes, there is one in the Validation Report but the report asserted elsewhere* that options for pursuing it within the Pythagoras framework may be limited.

Recommendations for the ABSVAL framework were also provided (Assumption Testing Report).

* “Pythagoras does not have the capability to implement a Markov Chain directly.” (Sect. 4.2.1)

Review Questions: P-COIN

Page 15: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Does the report clearly identify the application (set of study questions) for which the model is being validated, and the model’s role in addressing those questions?

Yes. Impact of:– Accuracy (intel and targeting)– Timing/sequencing of detonations– Bomb type– Number of bombs used

On AAV survivability against MIW threat in the surf zone

Review Questions: P-OCM

Page 16: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Does the report clearly describe the tests that were performed on the model, the possible outcomes for each test, and the criteria for passing?

Yes. Again, the tests generally did not have quantitative thresholds or other features that called for any elaborate description of the test and its purpose prior to giving the result.

Review Questions: P-OCM

Page 17: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

For each test performed, is the result clearly presented in a way that relates directly to the specified acceptance criteria?

Yes, except that the acceptance criteria were implicit.

Counterintuitive results validation findings without the ability to drill down on them.

Review Questions: P-OCM

Page 18: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Does the report provide a recommended decision for the accreditation authority?

Yes. “P-OCM has to be considered invalid for analytic applications due to the near-impossibility of some of its results.” (Section 6.6.1)

….given that it was not possible to drill down on the causes of those results. Doing so might have produced:– recommendations for model improvement– recommendation to increase numbers of replications– an attribution of the counterintutive result to something other than a problem with the model.– a recommendation for changing the set-up to avoid the observed problem

Review Questions: P-OCM

Page 19: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Does the report make a convincing argument that the tests conducted collectively provide a sufficient basis for the recommended accreditation decision?

Yes. The tests and results presented were sufficient to cause us to doubt the model’s ability to reliably answer the study questions.

Again, supporting a positive accreditation recommendation, had that been “the right answer,” would probably have required much more work on the validation side.

Where problems were found in assumption testing, these were again assessed not to invalidate the conceptual model.

Review Questions: P-OCM

Page 20: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Are recommendations provided that are actionable by a model improvement program? (This question was added in response to feedback received in Workshop II. It is understood that these are not a required ABSVAL output)

No, but the model was unavailable for diagnosis of counter-intuitive results, the activity that would produce most such recommendations as its by-product.

Recommendations for the ABSVAL framework were provided.

Review Questions: P-OCM

Page 21: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Audit Summary

Validation Report P-COIN P-OCM

Application clearly identified

Tests clearly explained

Some a little under-the-radar

Test results clearly presented

Validity implications of adverse results

Validity implications of adverse results

Recommendation on Accreditation

Some post-processing required

Support for Accreditation Recommendation

Validity implications of adverse results

Model Improvement Recommendations

N/A

Page 22: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

What Was Tested

ResultsReferent

Ideal Sim

ConceptualModel

InstantiatedModel

CM validation thru assumption testing (both cases)

Referent Validation (P-OCM)

P-OCM results could not be validated

against a referent, but against analyst

intuition.

No accessible referent for P-COIN

Simplified Cloud Chart

Instantiated Model verification done for

P-COIN

Page 23: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Other Comments on the Test Cases (1/2)

• The Assumption Testing Report included material describing the foundations of ABSVAL and some components of the process. However, it was difficult to infer from the reports much about the overall ABSVAL “recipe” (or whether there is one).

– Probably in part because material supporting an adverse Accreditation recommendation was found before much ground needed to be covered.

• Demonstrations suggested that subjective elements (e.g., assumption testing) are difficult to eliminate from V&V altogether without material degradation to the process outcome.

• In Assumption Testing, there is no way to guarantee a complete enumeration of the key assumptions, but decomposition of the CM and assumption types is a good way to ensure a reasonably diligent search.

– Fundamentally, it’s brainstorming.

Page 24: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Other Comments on the Test Cases (2/2)

• Reports reflect significant effort to develop and demonstrate broadly applicable V&V techniques and principles, but…

• … test cases did not appear to showcase or exercise any features of ABSVAL specifically responsive to the particular challenges of ABS, especially emergent behavior.

– Limited selection of real world ABS applications available on which to test the process.

• ABSVAL initiative too proactive for its own good?– Casting the process to emphasize model improvement and

building credibility (vice trying to invalidate) might help draw more test cases out of the woodwork

• Process as defined is inherently adversarial, and sponsor reluctance to expose completed work to scrutiny can limit opportunities to apply it.

• How would results validation be done for an ABS that had actual emergent behavior? Test cases did not force this question.

Page 25: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

General Recommendations for ABSVAL (1/2)

• We saw a collection of validation techniques described and applied, but only vague indications of any rules governing their application, or guidelines for identifying what steps are necessary or sufficient in a given situation.

– Necessity/sufficiency criteria; ordering of steps; if you find this, ABSVAL says you should do that.

• It would be helpful to clarify the minimum body of information/ material to which the V&V team must have access in order to reach a favorable accreditation recommendataion.

– When accreditation support (vice model improvement) is the goal, why start if no chance for up-check?

• Need to adjudicate “potential problems,” caveats, etc. as, e.g., “show stopper” “limits scope of valid uses” or “not really a problem after all.”

– Could guide a portion of the results testing.– Provide immediate feedback to development team– When IM does not conform to CM, test the assumptions underlying the

“offending” portion of the IM for impact.

Page 26: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

General Recommendations for ABSVAL (2/2)

• Third party, after-the-fact validation is fundamentally a game of catch-up. We believe that much of the Validation effort (the part not specific to a given use of the model) could be accomplished in conjunction with model development and would be largely reusable in support of multiple Accreditation decisions.

– Example: identification of the assumptions is re-usable, determination that they are OK for a specific application is not.

– Example: Assessment of the fidelity of an IM* to the CM is reusable, results validation generally is not, although an adverse finding may be apply broadly.

• Suggest a future test of ABSVAL be structured with a reusable component and an application-specific component, applied in sequence.

– Reusable component could be developed as a concurrent task within model development, if an appropriate test case were available.

* Including hardware and OS used

Page 27: Auditing ABSVAL 9 July 2008 Prepared for Workshop III: Determining a Validation Methodology for Agent-Based Simulation by Systems Planning and Analysis,

9 July 2008

Audit Wrap-Up

• Found that ABSVAL process was demonstrated successfully on the test cases – Leaving some questions about ABSVAL unanswered.– Observed areas that would benefit from refinement.

• Completed to date:– Reviewed Measures of Validity– Audited VV&A Applications – Participated in Test Case Workshops

• SPA Final Step: Audit Report – Summary of today’s presentation and key insights gained from final

workshop.– Includes SPA recommendations for improvement of ABSVAL process,

and related practical issues.