Review of Cognitive Metrics for C2 Mandy Natter Jennifer Ockerman, PhD Leigh Baumgart.

Slides:



Advertisements
Similar presentations
Agency for Healthcare Research and Quality (AHRQ)
Advertisements

Ch 8: Experimental Design Ch 9: Conducting Experiments
Donald T. Simeon Caribbean Health Research Council
Dynamic Decision Making Lab Social and Decision Sciences Department Carnegie Mellon University 1 MODELING AND MEASURING SITUATION AWARENESS.
1 C2 Maturity Model Experimental Validation Statistical Analyses of ELICIT Experimentation Data Dr. David S. Alberts.
Item Writing Techniques KNR 279. TYPES OF QUESTIONS Closed ended  Checking yes/no, multiple choice, etc.  Puts answers in categories  Easy to score.
Power to the Edge “Net Work” Network Science Issues of Interest Dr. David S Alberts Director, Research OASD(NII) October 2008.
International Journal of Industrial Ergonomics 35 (2005) 939–953 Situation awareness and workload in driving while using adaptive cruise control and a.
Methods and Techniques of investigating user behavior Introduction - why M & T? Gerrit C. van der Veer aims theory methods planning presentation.
Evaluation Methodologies
Creating Architectural Descriptions. Outline Standardizing architectural descriptions: The IEEE has published, “Recommended Practice for Architectural.
3 Chapter Needs Assessment.
Exhibit 5.1: Many Ways to Create Internal Structure
Building Knowledge-Driven DSS and Mining Data
PPA 502 – Program Evaluation Lecture 4a – Qualitative Data Collection.
Software Process and Product Metrics
DEFINING JOB PERFORMANCE AND ITS RELATIONSHIP TO ASSESSMENTS.
Architecture Tradeoff Analysis Method Based on presentations by Kim and Kazman
Choosing Your Primary Research Method What do you need to find out that your literature did not provide?
What is Business Analysis Planning & Monitoring?
Experimental Research
Chapter 1: Introduction to Statistics
Chapter 7: Memory and Training Slide Template. WORKING MEMORY.
September 19-20, 2005 Rio de Janeiro, Brazil Internationally Comparable General Disability Measures Barbara M. Altman National Center for Health Statistics.
Authentic Assessment Lynne E. Houtz, Ph.D. Associate Professor of Education Creighton University.
Chapter 8 Architecture Analysis. 8 – Architecture Analysis 8.1 Analysis Techniques 8.2 Quantitative Analysis  Performance Views  Performance.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
VTT-STUK assessment method for safety evaluation of safety-critical computer based systems - application in BE-SECBS project.
Robert W. Lingard California State University, Northridge EEET July 11, 2009.
Creating a Shared Vision Model. What is a Shared Vision Model? A “Shared Vision” model is a collective view of a water resources system developed by managers.
A COMPETENCY APPROACH TO HUMAN RESOURCE MANAGEMENT
Role of Statistics in Geography
The Analysis of the quality of learning achievement of the students enrolled in Introduction to Programming with Visual Basic 2010 Present By Thitima Chuangchai.
Assessment Professional Learning Module 5: Making Consistent Judgements.
September 2007 Survey Development Rita O'Sullivan Evaluation, Assessment, & Policy Connections (EvAP) School of Education, University of North Carolina-Chapel.
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Baker ONR/NETC July 03 v.4  2003 Regents of the University of California ONR/NETC Planning Meeting 18 July, 2003 UCLA/CRESST, Los Angeles, CA ONR Advanced.
1 Assessment Professional Learning Module 5: Making Consistent Judgements.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
SURVEY RESEARCH.  Purposes and general principles Survey research as a general approach for collecting descriptive data Surveys as data collection methods.
Copyright Catherine M. Burns
Advanced Research Methods Unit 3 Reliability and Validity.
Copyright  2003 by Dr. Gallimore, Wright State University Department of Biomedical, Industrial Engineering & Human Factors Engineering Human Factors Research.
Search Engine Optimization © HiTech Institute. All rights reserved. Slide 1 What is Solution Assessment & Validation?
Decision Making Chapter 7. Definition of Decision Making Characteristics of decision making: a. Selecting a choice from a number of options b. Some information.
Chapter 3 System Performance and Models Introduction A system is the part of the real world under study. Composed of a set of entities interacting.
Human Computer Interaction
U SER I NTERFACE L ABORATORY Situation Awareness a state of knowledge, from the processes used to achieve that state (situation assessment) not encompass.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Petter Nielsen Information Systems/IFI/UiO 1 Systems development Methodologies IN364.
Descriptive Research & Questionnaire Design. Descriptive Research Survey versus Observation  Survey Primary data collection method based on communication.
Unit 6 Understanding and Implementing Crew Resource Management.
MGT-555 PERFORMANCE AND CAREER MANAGEMENT LECTURE NO
3/3/2016Copyright Ed Lipinski and Mesa Community College, All rights reserved. 1 Research Methods Summer 2009 Making Systematic Observations.
Chapter Two Copyright © 2006 McGraw-Hill/Irwin The Marketing Research Process.
Network Centric Planning ---- Campaign of Experimentation Program of Research IAMWG Dr. David S. Alberts September 2005.
Behavioral Attitude Measurement Theory Don W. Stacks, Ph.D. School of Communication University of Miami Coral Gables, FL and The Institute for Public.
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Michelle Hester Advanced Human Factors (PSYC 562)
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Introduction Social ecological approach to behavior change
OPERATING SYSTEMS CS 3502 Fall 2017
Information for marketing management
Robert W. Lingard California State University, Northridge
Chapter 7: Memory and Training
Analyzing Reliability and Validity in Outcomes Assessment Part 1
Robert W. Lingard California State University, Northridge
The MR process.
Analyzing Reliability and Validity in Outcomes Assessment
Week 14 More Data Collection Techniques Chapter 5
Presentation transcript:

Review of Cognitive Metrics for C2 Mandy Natter Jennifer Ockerman, PhD Leigh Baumgart

Agenda  Measures Overview: What to consider  Summary of Cognitive C2 Metrics  Workload  Situation Awareness  Decision Making  Collaboration  Future Work

Measures Overview: What to consider Measurement types: qualitative, quantitative, subjective, objective Measurement scales: nominal, ordinal, interval, ratio Number of participants: power analysis from pilot study How measures are analyzed: descriptive statistics or inferential statistics Raters: self reports, experimenter, subject matter experts Timing: real-time, post hoc Measurement Criteria

Measures Overview: What to consider (cont’d) CriteriaDescription ValidityMeasuring the mental construct of interest 1 Repeatability / reliability Same results of the mental construct when tests are administered more than once 1 SensitivityDetect changes in the level of the mental construct imposed by task difficulty or resource demand DiagnosticityDiscriminate the amount of the mental construct imposed on different operator resources (e.g., perceptual versus processing versus motor resources) SelectivitySensitive to differences only in the cognitive construct of interest (e.g. cognitive demands as opposed to physical workload or emotional stress 1 ) IntrusivenessInterfere with performance on the primary task Implementation requirements / convenience Ease of implementing specific assessment technique (e.g., instrumentation requirements or operator training) Operator Acceptance Willingness on the part of the operators to follow instructions and actually utilize a particular assessment technique 1 Zhang & Luximon, 2005 Measurement Criteria

Workload Situation Awareness Decision Making Workload Situation Awareness Decision Making Workload Situation Awareness Decision Making Communication and Collaboration Interactions of Cognitive C2 Metrics

Workload  Workload is the portion of human’s limited capacity that is required to perform a particular task 1  In C2, cognitive workload is of most interest 2  Appropriate workload is most important  Too low or too high can both be bad 1 O’Donnell & Eggemeier, Zhang & Luximon

Workload Measurement Techniques  Primary-task measures: Quality (speed & accuracy) of primary task  Pro(s): Objective; related to performance  Con(s): May reflect data or system limitations vs. human, may have low sensitivity when task is easy  Secondary-task measures: Quality (speed & accuracy) of secondary task  Pro(s): High validity- helps predict residual resources in the event of a failure, can compare the workload of two different primary tasks  Con(s): Interference with primary task, must match the resource demands of the primary task  Subjective measures: Ratings by person doing task or observing subject matter expert  Pro(s): Low cost, ease of use, general non-intrusiveness, high validity, high sensitivity (at times more than objective measures)  Con(s): Confounding factors, short-term memory constraints, non-continuous (do not reflect changes in workload during the task)  Physiological measures: Use of physiological measures to objectively assess workload  Pro(s): Continuous, very sensitive, general non-intrusiveness  Con(s): New and immature area, confounds (e.g., individual differences, noise)

Situation Awareness (SA)  Situation awareness 1 is  the perception of elements in the environment within a volume of time and space (level 1)  the comprehension of their meaning (level 2)  the projection of their status in the near future (level 3)  Military terminology 2  Situational Awareness (level 1)  Situational Understanding (level 2)  Situational Assessment (level 3)  Sensemaking is the process to arrive at and maintain SA 1 Endsley, Gartska & Alberts, 2004

Situation Awareness (SA)  Technical and Cognitive components have been explored

Cognitive Situation Awareness Measurement Techniques  Explicit SA measures: Ask what SA is to determine level  Pro(s): Validated technique (e.g., SAGAT)  Con(s): Intrusive- may disrupt primary task, confound SA with probes, laborious to create probes  Implicit SA measures: Infer what SA is to determine level  Pro(s): Easy to obtain, less intrusive than explicit measures  Con(s): Only simple responses and behaviors, assumption driven  Subjective SA measures: Ratings by person doing task or observing subject matter expert  Pro(s): Does not need to be customized for different domains, easily employed, high validity  Con(s): May be confounded by performance and workload, usually post hoc- so rely on memory, individual differences, inter-rater reliability  Team SA measures: Multiple methods used  Pro(s): Most C2 environments involve teams  Con(s): Immaturity, complexity (e.g., status within the team, lack of control, differing expectations, prevented action by other team members)

Decision Making  Decision making is a complex process, not just the result.  Involves selecting options from alternatives, where:  some information pertaining to the option is available  time allotted is longer than a second  there is uncertainty or risk associated with the selection 1  Information component 2  provides “right” information, to the “right” person, at the “right” time  determine using cognitive engineering knowledge elicitation techniques (e.g., task analysis, cognitive task analysis, & cognitive work analysis, etc.)  Human component  selection of, or at least responsibility, for COA  rational or analytical decision making 3  naturalistic or intuitive decision making 4 1 Wickens & Hollands, Means & Burns, Azuma et al., e.g., Klein, Calderwood, & Macgregor, 1989

Decision Making Measurement Techniques  Complicated due to difficulties  in defining a “good” decision  influence of many factors (hard to equate decision making with mission effectiveness)  observing or eliciting strategies  continuous nature of some decisions  Result-based measures: Measure quality (accuracy and timeliness) of decision  Pro(s): Easy to employ, objective, observable, related to performance  Con(s): Doesn’t provide decision rationale, could be luck or chance, may not be “best” decision  Process-based measures: Measure appropriateness of strategies and evaluate information used  Pro(s): Understand “why”, lead to improved C2 processes  Con(s): Some processes not observable, difficult to represent, resource intensive, difficult to assess reliability and validity

Communication and Collaboration  Communication is expression and may include information sharing  prerequisite for collaboration  Collaboration involves leveraging the information of others to reach or meet a goal or objective Freeman et al., 2006

Collaboration Measurement Techniques  Technical-based measures: evaluate interconnectivity, modes available for communication, communication network, etc.  Pro(s): Easy to automate collection, trend analysis  Con(s): Implicit  Human-based measures  time-spent collaborating, content and frequency of collaboration, use of collaboration modes, etc.  goal of collaboration  social network and knowledge distribution diagrams 1  leadership and experience  Pro(s): Understand elements of team understanding and decision making, non-intrusive  Con(s): Resource intensive, can be difficult to represent, difficult to assess reliability and validity  Content analysis is usually very time and labor intensive 1 Freeman et al., 2006

Cognitive Metrics for C2 Research  Issues  Still much debate within the cognitive engineering community on appropriate definitions and metrics  Most metrics still focused on an individual  Workload, SA, DM, and collaboration are highly interdependent  Not a lot of automation available to collect and analyze collected data  Mitigations  Use suite of complimentary and overlapping measurement techniques  Design the evaluation and the analysis ahead of time  Use the automation that is available to collect and analyze data

JHU/APL Metrics Overview (FY07 and FY08) Secondary Task Subjective PhysiologicalTeam Subjective Implicit Explicit Process- based Result-based Human-based Technical- based Collaboration Decision Making Situation Awareness Workload Primary Task FY07 FY08 Key Physiological Exploring these metrics in two studies Physiological and Collaboration

Questions?