Applying Modeling and Simulation Verification, Validation and Accreditation (VV&A) Techniques to Test and Laboratory Facilities Dr. James Elele, Jeremy.

Slides:



Advertisements
Similar presentations
System Integration Verification and Validation
Advertisements

Software Quality Assurance Plan
© 2005 Prentice Hall6-1 Stumpf and Teague Object-Oriented Systems Analysis and Design with UML.
DoD Information Technology Security Certification and Accreditation Process (DITSCAP) Phase III – Validation Thomas Howard Chris Pierce.
VV&A in Human, Social, Cultural Behavior (HSCB) Simulation Dr. Jimmie McEver Dr. David T. Signori Dr. Mike Smeltzer Evidence Based Research, Inc.
Knowledge Acquisitioning. Definition The transfer and transformation of potential problem solving expertise from some knowledge source to a program.
Herman Aguinis, University of Colorado at Denver Prentice Hall, Inc. © 2006 Measuring Results and Behaviors: Overview  Measuring Results  Measuring Behaviors.
The Lumina Center Grantseeking Workshop Series Presents Outcomes & Evaluations April 20, 2006.
1 Introduction to System Engineering G. Nacouzi ME 155B.
Chapter 5: Project Scope Management
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
Introduction to Software Testing
What is Business Analysis Planning & Monitoring?
1 Lecture 5.3: SEF Ch 4 Requirements Analysis Dr. John MacCarthy UMBC CMSC 615 Fall, 2006.
S/W Project Management
What is Software Engineering? the application of a systematic, disciplined, quantifiable approach to the development, operation, and maintenance of software”
1 Validation & Verification Chapter VALIDATION & VERIFICATION Very Difficult Very Important Conceptually distinct, but performed simultaneously.
Software System Engineering: A tutorial
NIST Special Publication Revision 1
Outcome Based Evaluation for Digital Library Projects and Services
Elizabeth Godfrey 1.  Periodic assessment of results Appropriateness, effectiveness, efficiency, impact, sustainability  Identifies intended and unintended.
Certification and Accreditation CS Phase-1: Definition Atif Sultanuddin Raja Chawat Raja Chawat.
What is a Business Analyst? A Business Analyst is someone who works as a liaison among stakeholders in order to elicit, analyze, communicate and validate.
Intent Specification Intent Specification is used in SpecTRM
ISM 5316 Week 3 Learning Objectives You should be able to: u Define and list issues and steps in Project Integration u List and describe the components.
REVISIONS TO GENERAL EDUCATION STUDENT LEARNING OUTCOMES Auburn University Senate Information Item, August 2014.
Lecture 7: Requirements Engineering
Integrated Risk Management Charles Yoe, PhD Institute for Water Resources 2009.
P1516.4: VV&A Overlay to the FEDEP 20 September 2007 Briefing for the VV&A Summit Simone Youngblood Simone Youngblood M&S CO VV&A Proponency Leader
Process Improvement. It is not necessary to change. Survival is not mandatory. »W. Edwards Deming Both change and stability are fundamental to process.
Leadership Chapter 7 – Path-Goal Theory.  Path-Goal Theory Perspective  Conditions of Leadership Motivation  Leader Behaviors & Subordinate Characteristics.
Search Engine Optimization © HiTech Institute. All rights reserved. Slide 1 What is Solution Assessment & Validation?
NCHRP Project Development of Verification and Validation Procedures for Computer Simulation use in Roadside Safety Applications SURVEY OF PRACTITIONERS.
Chap. 5 Building Valid, Credible, and Appropriately Detailed Simulation Models.
IV&V Facility 26SEP071 Validation Workshop Dr. Butch Caffall Director, NASA IV&V Facility 26SEP07.
MODES-650 Advanced System Simulation Presented by Olgun Karademirci VERIFICATION AND VALIDATION OF SIMULATION MODELS.
SAAMF Roadshow Durban CSIR NML Eddie Tarnow Metrologist: Torque & Automotive 14 June 2006 ISO/TS 16949:2002 certification – Meeting the requirements of.
Chapter 10 Verification and Validation of Simulation Models
Job Analysis - Competency Modeling MANA 5322 Dr. Jeanne Michalski
Building Simulation Model In this lecture, we are interested in whether a simulation model is accurate representation of the real system. We are interested.
SOFTWARE PROJECT MANAGEMENT
Verification and Validation — An OSD Perspective — Fred Myers Deputy Director, Test Infrastructure Test Resource Management Center November 4, 2009.
Software Architecture Evaluation Methodologies Presented By: Anthony Register.
A Metrics Program. Advantages of Collecting Software Quality Metrics Objective assessments as to whether quality requirements are being met can be made.
Business Analysis. Business Analysis Concepts Enterprise Analysis ► Identify business opportunities ► Understand the business strategy ► Identify Business.
Irregular Warfare Modeling/Data Validation Best Practices 11 th Annual MOVES Research Education Summit 12 July 2011.
Software Quality Assurance SOFTWARE DEFECT. Defect Repair Defect Repair is a process of repairing the defective part or replacing it, as needed. For example,
Search Engine Optimization © HiTech Institute. All rights reserved. Slide 1 Click to edit Master title style What is Business Analysis Body of Knowledge?
VV&A-1 The Basics of the M&S VV&A Process Simone Youngblood The Johns Hopkins University Applied Physics Lab.
SCOPE DEFINITION,VERIFICATION AND CONTROL Ashima Wadhwa.
Ms. Lisa Jean Moya WernerAnderson, Inc. 01 May 2007 Validation Methodology for Agent-Based Simulations Workshop DoD Validation Baseline.
5 September 2002AIAA STC Meeting, Santa Fe, NM1 Verification and Validation for Computational Solid Mechanics Presentation to AIAA Structures Technical.
Company LOGO. Company LOGO PE, PMP, PgMP, PME, MCT, PRINCE2 Practitioner.
1 Lecture 2.3: SE Process (SEF Ch 3) Dr. John MacCarthy UMBC CMSC 615 Fall, 2006.
Building Valid, Credible & Appropriately Detailed Simulation Models
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
 System Requirement Specification and System Planning.
1 The XMSF Profile Overlay to the FEDEP Dr. Katherine L. Morse, SAIC Mr. Robert Lutz, JHU APL
Air Force Institute of Technology
Technology Readiness Assessment (TRA)
2012 Spring Simulation Interoperability Workshop
INCOSE Usability Working Group
(Additional materials)
Software Quality Engineering
Chapter 10 Verification and Validation of Simulation Models
Introduction to Software Testing
Assessment Workshop Title of the Project (date)
MECH 3550 : Simulation & Visualization
Building Valid, Credible, and Appropriately Detailed Simulation Models
M. Kezunovic (P.I.) S. S. Luo D. Ristanovic Texas A&M University
Presentation transcript:

Applying Modeling and Simulation Verification, Validation and Accreditation (VV&A) Techniques to Test and Laboratory Facilities Dr. James Elele, Jeremy Smith NAWCAD BSMVV Branch James.Elele@navy.mil David Hall, Charles Pedriani SURVICE Engineering Company Dave.Hall@survice.com ASME V&V Conference 3 May 2012

Introduction Tasking to support accreditation of Test and Evaluation Facilities in support of IFF Program Develop an accreditation case for T&E facilities for operational testing Applied risk-based M&S VV&A approach to facilities Approach applied successfully to M&S for over 20 years Test case for future T&E facility accreditation efforts Successful application can support efforts to institutionalize process for T&E as well as M&S IFF = Identification Friend or Foe T&E = Test and Evaluation M&S = Model and Simulation

M&S VV&A Definitions Verification: The process of determining that a model implementation and its associated data accurately represent the developer's conceptual description and specifications. Does the model do what the originator intended, and is it relatively error free? Validation: The process of determining the degree to which a model and its associated data are an accurate representation of the real world from the perspective of the intended uses of the model. How well do model results match real world data, in the context of your needs? Accreditation: The official certification [determination] that a model, simulation, or federation of models and simulations and its associated data are acceptable for use for a specific purpose Does the accreditation authority have adequate evidence to be confident that a model is fit for purpose? Did you build the model right? Did you build the right model? Did your customer accept it? Definitions from DODI 5000.61 dated 13 May 2003

Underlying Principles The ultimate goal of VV&A efforts is to form a foundation for making good decisions Nature and extent of information required to support accreditation decision is at the discretion of the accreditation authority and is generally based on assessment of risk Role of M&S results in decision making process Importance of decision that M&S is supporting Severity of the consequences of making incorrect decisions because M&S were wrong Probability that analysis results based upon M&S will be challenged Best of Show Good Enough “Better is the Enemy of Good Enough!”

Steps to an Accreditation Decision Analyze Intended Use Intended Use Statement Develop M&S Reqts and Accreditation Info Reqts Accreditation Plan Develop/Execute Accreditation Plan V&V Develop Accreditation Case A Review Accreditation Case Fit for Intended Use Make Accreditation Decision

How Much Credibility Is “Enough”? It Depends on Risk M&S User A Makeshift Bridge is Good Enough If You Need To Cross a Meandering Shallow Stream M&S In attempting to make an accreditation assessment, most assessors face the common question of “how much of this V&V information is necessary to demonstrate credibility for a particular application?”. The answer to this questions depends in large part on risk. If the consequences of using an inaccurate simulation are not severe, the risk is relatively low and one would need much less evidence of simulation credibility. Just as in this picture, a makeshift structure of evidence may be satisfactory for low risk applications. BUT

Evidence of Greater Credibility Greater Risks... Indicate the Need for Evidence of Greater Credibility M&S Supporting Evidence However, in those cases where inaccurate simulation predictions could have disastrous consequences, one needs much more evidence to guarantee that the simulation is credible enough to reduce the risks. For example, the simulations that are being used to prove that the ballistic missile defense system will work under all circumstances must have a lot of credibility to justify the expenditure of funds on so large and important a system. The risks associated with erroneous performance predictions and a potentially inadequate missile umbrella would be extremely high. In these cases, the User should demand a comprehensive accreditation assessment that is supported by a strong foundation of extensive V&V results. In these situations, the case supporting the accreditation needs to be fully documented and reviewed. The guiding principle is that quantity and quality of the evidence is directly related to the level of risk of the intended use. PROBLEM CREDIBLE SOLUTION

V&V: The Central Pillars of Simulation Credibility S/W Accuracy Data Accuracy Output Accuracy Simulation meets design requirements, operates as designed and is free of errors in software Simulation input data, validation data and data manipulations are appropriate and accurate Simulation outputs match the real world “well enough” to be of use in a particular problem Software (S/W) Accuracy Data Accuracy Output Accuracy Verification Validation Approaching the question of what makes a simulation credible from an abstract perspective, one can think of a simulation as the bridge. To build confidence that the bridge is adequate, one would assess the capabilities and characteristics of the bridge in light of the requirements for the bridge. Typically, one would review and evaluate some combination of design information, engineering certifications, test reports, builder qualifications, maintenance records, etc. to determine if the bridge was adequate for the type and level of traffic that was expected. In an analogous way, we can assess the credibility of a simulation for a specific application based on knowledge of its design, the accuracy of its software (verification results), the accuracy of its outputs (validation or test results), the accuracy of its input data, its usage history, etc.. Most people think of simulation credibility in terms of three types of accuracy: software accuracy, data accuracy and output accuracy. Software accuracy refers to the error-freeness of software design and coding; data accuracy refers to the accuracy and appropriateness of input data to the simulation; output accuracy refers to the degree to which simulation outputs match the real world. These aspects of simulation credibility are easily recognized as the objects of verification and validation activities. These elements give one an understanding of simulation accuracy. However, these elements, alone, do not comprise a comprehensive assessment of simulation credibility? "V" & "V" But, V&V are Just the Middle of the Bridge!

The “Other Pillars” of Simulation Credibility Capability Usability Simulation possesses all required functionality and fidelity for the problem being solved Anchors the M&S to the Problem Simulation has adequate user support to facilitate correct operation and interpretation of its outputs Ties the M&S to a Useful Solution M&S Capability Accuracy Usability M&S Requirements User Capabilities Our experience suggests that simulation accuracy (as defined in the previous slide) is a necessary but not a sufficient condition for a robust determination of simulation credibility. The software may be demonstrably free of errors, the data may be demonstrably accurate and correct, and simulation outputs may correlate well with real world data. But does the simulation represent all the real world phenomena that affect the problem you are trying to solve? Does it simulate these functions at the appropriate level of detail? Is the simulation (software, documentation and data) managed well? Is it accompanied by enough support structure to allow you to use it credibly, given the user’s level of experience with the simulation? Questions like these provide the approaches to our simulation “bridge” and provide the linkage between the bridge and the anchor points on the banks of the river that must be crossed. Any robust assessment of simulation credibility must consider not only “accuracy”, but also “capability” and “usability” in determining whether it is suitable for use in a particular application. Any accreditation assessment that is evaluating simulation credibility for a given application must consider all five of these credibility elements. Each of these five elements, capability, software accuracy, data accuracy, output accuracy, and usability, will be discussed individually in the following sections of the tutorial. Credible Solution Problem Accuracy of: Software Data Outputs Can I Be Sure I’m Not Mis-Using the M&S? Does the M&S Do What I Need It To Do?

The Essence of Accreditation M&S REQUIREMENTS IDENTIFY M&S DEFICIENCIES M&S INFORMATION Capability Accuracy Usability Data Quality M&S Documentation Design Documentation Configuration Mgt V&V Results Etc. IDENTIFY WORK-AROUNDS, USAGE CONSTRAINTS, REQUIRED IMPROVEMENTS AND RISKS Provided by the Model Developer or Model Proponent Defined by the User (Formally or Implied) ACCREDITATION DECISION Accreditation and the associated accreditation assessment have been discussed. However, it is important to understand just what accreditation entails. This diagram shows a simplified view of what accreditation really is and how it is typically done. Within the DoD, accreditation is defined as the official determination that a model is acceptable for a specific purpose. In practical terms, this definition implies that accreditation depends on a comparison of a model’s capabilities, limitations, and attributes with the simulation requirements that are generated from the specific problem for which the model will be used. To make such a judgment about model suitability, one must have a complete set of simulation requirements. These requirements are derived from the unique problem which the User is trying to solve. In addition to the simulation requirements, one must also have documented evidence of the simulation’s capabilities, limitations, and attributes. Typically, this information is found in the sources indicated. A key part of this information is the evidence of data quality. This evidence is quite frequently obtained from sources completely separated from the Model Manager. In making the assessment, one usually cannot be satisfied with a simple “no” answer if the capability, accuracy, or usability of the simulation does not meet requirements. In these cases one must analyze the impacts of any deficiencies, determine if any work-arounds exist, and evaluate these against the requirements. Such an analysis should yield a list of tasks that must be done to make the simulation acceptable. PROBLEM CONTEXT TO PROVE THE M&S IS SUITABLE FOR THE NEED REQUIRES AN OBJECTIVE COMPARISON OF M&S INFORMATION WITH M&S REQUIREMENTS WITHIN THE CONTEXT OF THE PROBLEM

How Much V&V is Enough? It Depends on Risk RISK = PROBABILITY x IMPACT Risk means something “bad” might happen because you believed an incorrect simulation result Decisions based on M&S results are at risk VV&A reduces that risk RISK HIGH LOW MODERATE “How much credibility is enough” can only be answered in the context of the application. That is, the answer to that question depends on what decisions you’re going to make using model outputs, and what the risks are if you make the wrong decision. If you have a very high risk application, such as decisions that might involve life or death situations (such as embedded flight-critical software in an aircraft), then the potential risk of a wrong decision is very high. In that case, the model must have very high credibility, which will require a high level of capability, accuracy and usability, and a high level of verification and validation activity to demonstrate that credibility. If the risk is low, then less is required to demonstrate that the model is suitable for your application. JASA has developed an Accreditation Information Requirements Guide (AIRGuide), which can be used to help assess the risks inherent in the application of models and simulations, and which guides the development of a V&V plan to mitigate those risks. The AIRGuide process is based on assessing both the impact of wrong decisions made with incorrect model outputs and the probability of a wrong decision being made. Risk is defined as the product of impact and probability. More information on the AIRGuide is available in the handouts for the tutorial. PROBABILITY IMPACT

Quantifying Risk Level Probability of Incorrect Simulation Result Level of Impact on the Program Negligible Marginal Serious Critical Catastrophic Frequent Low Moderate High Probable Occasional Remote Impossible Once the User has quantified the probabilities and impact levels, these can be combined to arrive at an overall level of risk (for each individual risk factor that is identified). This table is one of several contained in the MIL-STD. It again is an example that can be tailored to the needs of the individual application. This type of table allows the User to determine the level of risk based on a level of impact and probability of each risk factor. If there are multiple risk factors associated with a particular application, each risk factor should be analyzed individually. (For example, if a simulation is being used to to determine the footprint of the missile, one risk factor is the chance of hitting an observer on the range if the predicted footprint is too small. Another is the likelihood of wasting resources to clear too large a test area because the predicted footprint was too large.) The highest level risk, considering the risk of each factor, will determine the level of credibility that is needed in the simulation being used. RISK LEVEL VALUES ARE: Subjective Consistent with MIL-STD-882 Tailorable to each application HIGHER RISK MEANS MORE CREDIBILITY EVIDENCE IS NEEDED TO ACHIEVE ACCREDITATION

Risk Reduction Strategies 5 4 Likelihood Improved M&S Credibility 3 2 1 1 2 3 4 5 Reduced reliance on M&S results Impact Low Moderate High Risk = Likelihood x Impact

Application to Test and Laboratory Facilities This was a trial application of the M&S VV&A approach to test facilities Identification Friend-or-Foe (IFF) system testing to evaluate new IFF system performance Accreditation of test facilities required by Commander, Operational Test and Evaluation Force (COMOPTEVFOR) Facilities used for system assessment include: All-up ship radar and related system representations Simple stimulators Engineering Test Equipment (ETE) facility specific to IFF system testing

ETE Facility Generates waveform signals to stimulate a production IFF transponder in the laboratory To evaluate system requirement for resistance to signals from transmitters other than the desired transmitter Critical technical parameter “susceptibility of the IFF system to false interrogations” Metric: Probability of resistance to false signals = Pr

Initial ETE Risk Assessment M&S Characteristic Criterion Risk CAPABILITY Intended Use The specific intended use(s) of the facility, model or simulation is/are clearly stated. LOW Design The facility and analysis process (framework, algorithms, data sources, and assumptions) produces credible results. MODERATE ACCURACY Input Data For each facility, model or simulation, input data are credible and subject to review and revision. System Verification The facility, model or simulation has been formally tested or reviewed and has been demonstrated to accurately represent the specific intended use(s) and requirements. Results Validation The facility’s, model’s or simulation’s responses have been compared with known or expected behavior from the subject it represents and has been demonstrated to be sufficiently accurate for the specific intended use(s). USABILITY Configuration Management For each facility, model or simulation, modeled components are supported by a sound written Configuration Management (CM) Plan. User Community For each facility, model or simulation, the capability is designed and developed for the level of competency for its intended purpose. The capability is supported by documents such as user’s manual, technical manual, and/or reference guide.

Initial ETE Assessment and Recommendations* CAPABILITY: The intended use is clearly stated: to evaluate the probability of responding to a false signal (Pr) No formal design documents exist: recommend laboratory design be adequately documented ACCURACY: Input data are provided by actual hardware Recommend documenting a complete set of test cases and results and any previous verification activities Recommend independent subject matter expert (SME) review of laboratory results USABILITY: Recommend facility develop and implement an overall configuration management plan The test approach appears to have been successfully used over a span of many years to support a variety of identification programs for DOD and the FAA: recommend the facility provide documented results of previous uses * Incorporated into ETE Facility Accreditation Plan

Observations Application to test facility similar to M&S Differences: Biggest issue for both: getting good documentation Lack of configuration management plans Poor documentation of prior V&V results They did V&V, they just didn’t write it down Accreditation Support Package (ASP) document format works equally well for both Differences: Developing intended use statement more natural for test facilities than for M&S Use of T&E facilities seems to have more focused objectives initially than use of M&S

Accreditation Support Package ACCURACY Software Accuracy Software Verification Results Design Verification Implementation Verification Software Development and Management Environment Software Development Environment Configuration Management Software Quality Assessment Implications for M&S Use Data Accuracy Input Data Data Transformations Output Accuracy Sensitivity Analysis Benchmarking Face Validation Results Validation EXECUTIVE SUMMARY ASP OVERVIEW CAPABILITY M&S Description Functional Capabilities Development History Assumptions and Limitations Implications for M&S Use USABILITY Documentation Assessment User Support Usage History

Thoughts on Broader Application Risk-based approach seems as applicable to test facilities as to M&S Risk assessment can also help prioritize which facilities justify spending more VV&A resources Suggest standardizing and institutionalizing risk-based VV&A process for both M&S and T&E No consistent application across DOD for either Risk-based VV&A promotes cost-effective VV&A for both M&S and T&E facilities