Download presentation
Presentation is loading. Please wait.
Published byElmer Fitzgerald Modified over 8 years ago
1
JAEC Assessment Initiatives and Implications Julia Loughran ThoughtLink, Inc. loughran@thoughtlink.com 703-281-5694 Presented to: NDIA’s Training Transformation Technologies Conference September 5, 2003
2
JAEC Assessment Initiatives Joint Lessons Learned Program 5 1Enhanced JTS JTS 3 Performance Assessment Architecture JKDDCJNTC 4 Linking Training to Readiness 2 Joint Performance Measures
3
1Enhanced JTS JTS Strengthen and broaden the focus of the JTS –“… by using the JTS methodology as the foundation for service, Federal agency, intergovernmental, and multinational training programs” Link combatant commander requirements with performance measures –Use this info to guide T2 assessments and investments Implications for JAEC: –Need to identify current JTS metrics, processes, AAR and lessons learned capabilities (establish baseline) –Broaden the existing scope Issue of acceptance outside DoD Consider other approaches (outside traditional education and training) –e.g., COPs, web portal, distributed trainings Example of JIACGs; interagency culture issues
4
Initial metrics based on existing tasks, conditions, and measures in the UJTL (“predictable capabilities”) Second level metrics focus on “unforeseen capabilities” (e.g., cognitive agility, innovative abilities) Implications for JAEC: –Definition of metrics, required data elements, standardization of assessment techniques –Identify issues related to data standardization, data interoperability –Need for innovative ways to measure typically hard to define characteristics (SSA, agility, adaptability, leadership) –Importance of developing measures for trainers 2 Joint Performance Measures
6
Architecture will help guide the integration of training simulations, instrumentation ranges, and distributed learning systems Incorporate common assessment tools and metrics in its operational software and global information environment JAEC Implications: –Need for new automated performance assessment tools/analysis techniques (e.g, PDAs) –Verify/validate metrics for overall system effectiveness E.g., strategy, technologies, trainers, teams, and individuals –Establish performance assessment repositories 3 Performance Assessment Architecture JKDDCJNTC
7
Enhance the Defense Readiness Reporting System –Develop more accurate/insightful methods for measuring the effectiveness of Joint education and training on performance of joint ops –Provide more robust reporting in terms of DOTMLPF (both predictable requirements and unforeseen exigencies) JAEC Implications: –Provide bridge between individual and collective measures –Need to link home station training to joint training events –Need to define readiness standards across the full range of missions –Enable performance trends analyses 4 Linking Training to Readiness
9
Integrate lessons learned into the development of new training processes and systems Systematically capture lessons learned from operational missions and inject into experimentation and concept development Build on the Joint Center for Lessons Learned JAEC Implications: –Need to more tightly couple the training community and the experimentation/concept development community –Highlight ways to improve lessons learned dissemination – focus on lessons learned vs. lessons lost Joint Lessons Learned Program 5
10
T2/JAEC’s Success Is Not Just About the Technology!!
11
Other Important Areas to Consider Communicating the vision Understanding roles Breaking down boundaries Dealing with resistance Timing Building trust Continuity of direction vs. accomplishments Cultural issues Addressing fears Cohesive top management Investment implications Preparing the system for change
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.