A Review Of Technical Performance and Technology Maturity Approaches for Improved Developmental Test and Evaluation Assessment

Similar documents
Program Success Through SE Discipline in Technology Maturity. Mr. Chris DiPetto Deputy Director Developmental Test & Evaluation October 24, 2006

Test & Evaluation Strategy for Technology Development Phase

Technology Transition Assessment in an Acquisition Risk Management Context

Reliability Growth Models Using System Readiness Levels

Air Force Research Laboratory

Closing the Knowledge-Deficit in the Defense Acquisition System: A Case Study

Reducing Manufacturing Risk Manufacturing Readiness Levels

DoDI and WSARA* Impacts on Early Systems Engineering

Jerome Tzau TARDEC System Engineering Group. UNCLASSIFIED: Distribution Statement A. Approved for public release. 14 th Annual NDIA SE Conf Oct 2011

WSARA Impacts on Early Acquisition

Defense Technical Information Center Compilation Part Notice

A New Way to Start Acquisition Programs

Are Rapid Fielding and Good Systems Engineering Mutually Exclusive?

Digital Engineering and Engineered Resilient Systems (ERS)

2 August 2017 Prof Jeff Craver So you are Conducting a Technology Readiness Assessment? What to Know

Update on R&M Engineering Activities: Rebuilding Military Readiness

The Role of CREATE TM -AV in Realization of the Digital Thread

Synopsis and Impact of DoDI

GAO Technology Readiness Assessment Guide: Best Practices for Evaluating and Managing Technology Risk in Capital Acquisition Programs

Technology Readiness Assessment of Department of Energy Waste Processing Facilities: When is a Technology Ready for Insertion?

Model Based Systems Engineering (MBSE) Business Case Considerations An Enabler of Risk Reduction

Engineered Resilient Systems NDIA Systems Engineering Conference October 29, 2014

Challenges and Innovations in Digital Systems Engineering

DEFENSE ACQUISITION UNIVERSITY EMPLOYEE SELF-ASSESSMENT. Outcomes and Enablers

Advancing the Use of the Digital System Model Taxonomy

Manufacturing Readiness Assessments of Technology Development Projects

DoD Modeling and Simulation Support to Acquisition

Autonomy Test & Evaluation Verification & Validation (ATEVV) Challenge Area

Fifteenth Annual INCOSE Region II Mini-Conference. 30 October 2010 San Diego

Defense Acquisition Guidebook (DAG) Chapter 4 Systems Engineering Update: Overview Briefing

An Assessment of Acquisition Outcomes and Potential Impact of Legislative and Policy Changes

Proposed Curriculum Master of Science in Systems Engineering for The MITRE Corporation

Presented at the 2017 ICEAA Professional Development & Training Workshop. TRL vs Percent Dev Cost Final.pptx

Large company practices. Small company responsiveness. Working for YOU.

Technology & Manufacturing Readiness RMS

Models, Simulations, and Digital Engineering in Systems Engineering Restructure (Defense Acquisition University CLE011)

Technology Roadmapping. Lesson 3

Software-Intensive Systems Producibility

RAPID FIELDING A Path for Emerging Concept and Capability Prototyping

The New DoD Systems Acquisition Process

OSD Engineering Enterprise: Digital Engineering Initiatives

Digital Engineering Support to Mission Engineering

Manufacturing Readiness Assessment (MRA) Deskbook

Engineering Autonomy

Best Practices for Technology Transition. Technology Maturity Conference September 12, 2007

Our Acquisition Challenges Moving Forward

Digital Engineering (DE) and Computational Research and Engineering Acquisition Tools and Environments (CREATE)

ROI of Technology Readiness Assessments Using Real Options: An Analysis of GAO Data from 62 U.S. DoD Programs by David F. Rico

Department of Energy Technology Readiness Assessments Process Guide and Training Plan

The Drive for Innovation in Systems Engineering

Michael Gaydar Deputy Director Air Platforms, Systems Engineering

2017 AIR FORCE CORROSION CONFERENCE Corrosion Policy, Oversight, & Processes

Stevens Institute of Technology & Systems Engineering Research Center (SERC)

INTRODUCTION TO THE DEVELOPMENT OF A MANUFACTURABILITY ASSESSMENT METHODOLOGY

Voluntary Education Program Readiness (Force Education & Training)

A Case Study to Examine Technical Data Relationships to the System Model Concept

Debrief of Dr. Whelan s TRL and Aerospace & R&D Risk Management. L. Waganer

Using the Streamlined Systems Engineering (SE) Method for Science & Technology (S&T) to Identify Programs with High Potential to Meet Air Force Needs

Department of Defense Instruction (DoDI) requires the intelligence community. Threat Support Improvement. for DoD Acquisition Programs

Integrated Transition Solutions

SUBJECT: Army Directive (Acquisition Reform Initiative #3: Improving the Integration and Synchronization of Science and Technology)

TECHNICAL RISK ASSESSMENT: INCREASING THE VALUE OF TECHNOLOGY READINESS ASSESSMENT (TRA)

R&M: Critical to Success in a Technology Reliant World

Four Conference Breakout Sessions

Image designed by Diane Fleischer

Manufacturing Readiness Level (MRL) Deskbook Version 2016

Vector-Based Metrics for Assessing Technology Maturity

Transitioning the Opportune Landing Site System to Initial Operating Capability

Gerald G. Boyd, Tom D. Anderson, David W. Geiser

Prototyping: Accelerating the Adoption of Transformative Capabilities

Defense Microelectronics Activity (DMEA) Advanced Technology Support Program IV (ATSP4) Organizational Perspective and Technical Requirements

Manufacturing Readiness Levels (MRLs) and Manufacturing Readiness Assessments (MRAs)

Comparing the State Estimates of a Kalman Filter to a Perfect IMM Against a Maneuvering Target

Acquisition Reform How are we Doing?

Manufacturing Readiness Assessment Overview

Presented at the 2007 ISPA/SCEA Joint Annual International Conference and Workshop - ISPA-SCEA 2007

Using System Architecture Maturity Artifacts to Improve Technology Maturity Assessment

System of Systems Software Assurance

Final Report of the Subcommittee on the Identification of Modeling and Simulation Capabilities by Acquisition Life Cycle Phase (IMSCALCP)

Manufacturing Readiness Level Deskbook

Rapid Fielding A Path for Emerging Concept and Capability Prototyping

Air Force Institute of Technology. A Quantitative Analysis of the Benefits of Prototyping Fixed-Wing Aircraft

Science & Technology for the Objective Force

Evaluating Complex System Development Maturity

Manufacturing Readiness Levels (MRLs) Manufacturing Readiness Assessments (MRAs) In an S&T Environment

UNCLASSIFIED. R-1 ITEM NOMENCLATURE PE S: Microelectronics Technology Development and Support (DMEA) FY 2013 OCO

Module 2 Lesson 201 Project Coordinator (PC) Duties

Follow the Yellow Brick Road

Management of Toxic Materials in DoD: The Emerging Contaminants Program

COLLECTIVE PROTECTION

Reconsidering the Role of Systems Engineering in DoD Software Problems

The Role of the Communities of Interest (COIs) March 25, Dr. John Stubstad Director, Space & Sensor Systems, OASD (Research & Engineering)

Analysis of Alternatives (AoAs) from a Cost Estimating Perspective

SERC Technical Overview: First-Year Results and Future Directions. Barry Boehm, USC Rich Turner, Stevens. 15 October 2009

Panel: Systems Engineering Considerations in Practicing Test & Evaluation A Perspective from DoD

Development of a Manufacturability Assessment Methodology and Metric

Engineered Resilient Systems DoD Science and Technology Priority

Administrative Change to AFRLI , Science and Technology (S&T) Systems Engineering (SE) and Technical Management

DOD Technology Innovation & Transition

A Knowledge-Centric Approach for Complex Systems. Chris R. Powell 1/29/2015

Transcription:

A Review Of Technical Performance and Technology Maturity Approaches for Improved Developmental Test and Evaluation Assessment Alethea Rucker Headquarters Air Force, Directorate of Test and Evaluation (AF/TE) Shahram Sarkani, Ph.D., P.E. Thomas A. Mazzuchi, D.Sc. The George Washington University NDIA Systems Engineering Conference Oct 22-25, 2012 This presentation is based on work leading to a dissertation submitted to The George Washington University in partial fulfillment of the Doctor of Philosophy Degree.

Outline Overview Current efforts Literature Review Technology Maturity Technical Performance Risk Observations Future Work Conclusion

Overview Purpose A review of technology maturity and technical performance approaches to improve DT&E assessment Discuss the mutual beneficial relationship between the SE and T&E communities Motivation A need for a tangible means to quantitatively assess technical readiness in order to transition from DT to IOT&E

Current efforts The DASD(DT&E) office is working to institutionalize the process and use of metrics to improve MDAP success in entering and exiting IOT&E. A framework, along with an initial set of performance criteria and associated metrics, was developed. Effort resulted in development of 14 performance criteria Department of Defense. (2011). DoD Developmental Test and Evaluation and Systems Engineering FY2011 Annual Report.

Performance Criteria 1. Key performance parameters (KPPs) are functionally traceable to Warfighter capabilities.* 2. KPPs are evaluated for mission capabilities. 3. Establish evaluation framework for KPPs and critical technical parameters (CTPs). 4. Execute evaluation framework for KPPs and CTPs.* 5. Demonstrated technical progress and system maturity. 6. Assess safety of the system.* 7. TEMP adequacy and currency. 8. DT&E resource management. 9. DT&E phase schedule performance. 10. Adherence to T&E policy and process.* 11. T&E program effectiveness and efficiency.* 12. AOTR accuracy. 13. T&E workforce certification status. 14. Fill identified T&E KLPs. (*) Requires further study to determine value and applicability Department of Defense. (2011). DoD Developmental Test and Evaluation and Systems Engineering FY2011 Annual Report.

Measurable Performance Criteria As part of the framework, the DASD(DT&E) developed a method for assessment. For each performance criterion, the Action Officer (AO) both assesses performance against the particular criterion and provides a confidence level in making the assessment. For the performance assessment, the DASD(DT&E) uses the stoplight colors of green, yellow, and red. The meaning of each stoplight assessment color was developed uniquely for each criterion to reflect the proper status. A Not Rated assessment is also available, as appropriate. Department of Defense. (2011). DoD Developmental Test and Evaluation and Systems Engineering FY2011 Annual Report.

Current Confidence Assessment High confidence is assessed when the presence and maturity of program T&E artifacts and documentation is consistent with expectations at the program s point in its life cycle. Medium confidence is assessed when the presence of program T&E artifacts and documentation is consistent with expectations at the program s point in its life cycle, but detail and maturity of documentation is lacking. Low confidence is assessed when there are omissions, gaps, inconsistencies, lack of expected detail, and/or conflicting data and information observed in program T&E artifacts and documentation. Subjective and oversimplifies performance criteria! Department of Defense. (2011). DoD Developmental Test and Evaluation and Systems Engineering FY2011 Annual Report.

Evaluation Framework Establish evaluation framework for KPPs and critical technical parameters (CTPs). Execute evaluation framework for KPPs and CTPs.

DT&E Assessment Assess technical progress and maturity against critical technical parameters (CTPs), key system attributes (KSAs), KPPs, and critical operational issues (COIs) as documented in the TEMP and test plans (DAU, 2012) CTPs can be used to assess completion of a major phase of developmental testing such as ground or flight testing; and determine readiness to enter the next phase of testing, whether developmental or operational (DAU, 2012)

Critical Technical Parameters Definition: CTPs measure critical system characteristics that, when achieved, enable the attainment of desired operational performance capabilities (DAU, 2012) Every technical parameter is NOT a CTP CTPs focuses on critical design features or risk areas (e.g., technical maturity, RAM issues, physical characteristics or measures) that if not achieved or resolved during development will preclude delivery of required operational capabilities (DAU, 2012)

How are CTPs derived? CTPs can be established from CTEs, TMPs, SE, etc. Mosser-Kerner, D. (2009). Test & Evaluation Strategy for Technology Development Phase. Presented at NDIA Systems Engineering Conference.

Technical Progress and Maturity Demonstrated technical progress and system maturity.

Problem Statement Objective and robust methods that can assess technology maturity accurately and provide insight into risks that lead to cost overruns, schedule delays, and performance degradation are imperative for making well-informed procurement decisions. (Azizian et al, 2009) The Weapon System Acquisition Reform Act of 2009 recognized that unrealistic performance expectations and immature technologies are among the root causes of trouble in defense programs (Gilmore, 2011) Reduce risk of immature technology in systems development (Stuckey, 2007) Programs that started development with immature technologies experienced an average acquisition unit cost increase of nearly 21 percent (GAO-05-301, 2005)

Technical versus Technology Maturity Assessing the maturity of a particular technology involves determining its readiness for operations across a spectrum of environments with a final objective of transitioning it to the user. Application to an acquisition program also includes determining the fitness of a particular technology to meet the customer's requirements and desired outcome for operations (MITRE, 2012) Stuckey, R. (2007). OSD DT&E Perspective: Technology Development and Maturation, Presented at AFRL Technology Maturity Conference.

Technology Maturity Immature technology is a primary source of cost and schedule risk (Stuckey, 2007) Recommendation was to add technology maturity focus into the SE and DT&E processes (Stuckey, 2007) TRL verification Not recommended due to numerous TRL limitations (Azizian et al, 2009) Subjective Focused on hardware Lacks accuracy and precision Not focused on system-to-system integration Does not communicate difficulty of maturing technology to higher TRL levels Increasing complexity of defense systems

Technical Performance Risk Little is available on how to integrate technical performance measures into a meaningful measure of system s overall performance risk (Garvey and Cho, 2003) 2003: Developed a Performance Risk Index Measure 2004: Extended previous work to measure and monitor a System-of-Systems Performance Risk TRLs does not measure how well the technology is performing against a set of performance criteria (Mahafza, 2005) 2006: Developed Technology Performance Risk Measure

Technology Performance Risk Measure Measures performance risk of technology in order to determine transition readiness Computed using performance requirements, the DD, and unmet performance Mahafza, S. (2004). Technology Performance Risk Measure. Presented at Multi-Dimensional Assessment of Technology Maturity Workshop.

SoS Technical Performance Risk Index Provides integrated measures of technical performance Measures technical performance as a function of the physical parameters of the TPMs Measures the degree of risk and monitors change over time Garvey, P.R. and Cho, C.C. (2004). An Index to Measure and Monitor a System-of-Systems Performance Risk, The MITRE Corporation.

MBSE Framework for T&E Uses MBSE framework and Monte Carlo Simulation to define uncertainty reduction goals for test planners to use in developing test strategies and detailed test designs for evaluating technical performance parameters Dr Bjorkman proposed a methodology to determine test value by estimating the amount of uncertainty reduction a particular test is expected to provide using Shannon s Information Entropy as a basis for the estimate Bjorkman, E.A., Sarkarni, S., and Mazzuchi, T.A. (2012). Using Model-Based Systems Engineering as a Framework for Improving Test and Evaluation Activities, Unpublished.

Observations Focus has been on cost and schedule; technical performance often an afterthought Recent emphasis on test planning and test design Need to redirect and increase focus on test analysis and reporting Immature technologies still an issue T&E interests need to be injected up front Critical technical parameter risks should be primary intent of research What do we go from here?

Future Work Assess system s progress and maturity against critical technical parameters as documented in the TEMP Integrate and quantify risk and uncertainty into CTPs Analyze and aggregate data using information theoretic approaches Shannon s information entropy; entropy as a risk measure Use Model-based Systems Engineering (MBSE) to continuously track and update various readiness levels Extend uncertainty reduction and MBSE research by Dr. Bjorkman into CTP reporting Roll values into a holistic decision making model Report decision making model at upcoming T&E conference and STAT Panel Meeting

Summary Need for critical technical performance risk index Inject technology maturity and uncertainty SE and T&E communities need to collaborate on development and tracking of technical performance, specifically CTPs, per DoDI 5134.17 T&E community needs to be involved as early as possible (pre-ms B) Keep moving forward with research.

Contact Information Alethea Rucker (703) 697-0296 alethea.rucker@pentagon.af.mil ShahramSarkani, Ph.D., P.E. sarkani@gwu.edu Thomas A. Mazzuchi, D.Sc. mazzu@gwu.edu

References Azizian, N., Sarkarni, S., and Mazzuchi, T. (2009) A Comprehensive Review and Analysis of Maturity Assessment Approaches for Improved Decision Support to Achieve Efficient Defense Acquisition, Proceedings of the World Congress on Engineering and Computer Science, Vol II. Bjorkman, E.A., Sarkarni, S., and Mazzuchi, T.A. (2012). Using Model-Based Systems Engineering as a Framework for Improving Test and Evaluation Activities, Unpublished. Defense Acquisition University. (2012) Defense Acquisition Guide, Chapter 9. Department of Defense. (2011). DoD Developmental Test and Evaluation and Systems Engineering FY2011 Annual Report. Department of Defense Instruction, DoDI 5134.17. (2011). Deputy Assistant Secretary of Defense for Developmental Test and Evaluation (DASD(DT&E)). Department of Defense. (2011). DOT&E FY 2011 Annual Report. http://www.dote.osd.mil/pub/reports/fy2011/. Garvey, P.R. and Cho, C.C. (2004). An Index to Measure and Monitor a System-of-Systems Performance Risk, The MITRE Corporation. Garvey, P.R. and Cho, C.C. (2003). An Index to Measure a System s Performance Risk, Acquisition Review Quarterl.y GAO. (2005). Assessments of Selected Major Weapon Programs. Gilmore, J.M. (2011). Key Issues Causing Program Delays in Defense Acquisition. ITEA Journal, 32, pages 389-391. Mahafza, S. (2004). Technology Performance Risk Measure. Presented at Multi-Dimensional Assessment of Technology Maturity Workshop. MITRE. (2012). Accessing Technical Maturity. http://www.mitre.org/work/systems_engineering/guide/acquisition_systems_engineering/acquisition_program_planning/assessing _technical_maturity.html Mosser-Kerner, D. (2008). New..Improved Test and Evaluation Master Plan. Presented at NDIA Systems Engineering Conference. Mosser-Kerner, D. (2009). Test & Evaluation Strategy for Technology Development Phase. Presented at NDIA Systems Engineering Conference. Stuckey, R. (2007). OSD DT&E Perspective: Technology Development and Maturation, Presented at AFRL Technology Maturity Conference.