+ All Categories
Home > Documents > Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4...

Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4...

Date post: 22-Jul-2020
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
18
Innovations in Engineering UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved Probability Driven Experimental Design for Autonomous Systems Troy Jones Autonomous Systems Capability Leader [email protected] (617) 258-2635 March 16, 2011 1 Team Members George Sass, Melissa Durfee, Nick Borer, Stephen York, Eric Nelson, Mike Ricard, Scott Ingleton
Transcript
Page 1: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

Innovations in Engineering

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Probability Driven Experimental Design for

Autonomous Systems

Troy Jones

Autonomous Systems Capability Leader

[email protected]

(617) 258-2635

March 16, 2011 1

Team Members George Sass, Melissa Durfee, Nick Borer, Stephen York, Eric

Nelson, Mike Ricard, Scott Ingleton

Page 2: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Motivation

March 16, 2011 2

Across DoD, lack of common vision for how to assess performance of decision-making systems

Need to meet needs of commanders, acquisition, and warfighter communities who need to trust system performance when needed, safely

Low confidence of performance in difficult conditions

Intractable to physically test every possible condition

Interesting Anecdotes

All deployed ground robots are tele-operated

Original iRobot Packbot had many autonomous driving features – they were removed

US Army tends to use automated Takeoff/Landing features of Predators, Air Force does not

Page 3: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Implementation

Concept of

Operations

Architecture

Requirements

Design

Integration, Test

& Verification

System

Verification &

Validation

Operations &

Maintenance

Behavioral

Model

Project Vision

March 16, 2011 3

Apply Draper experience in System Engineering, M&S, Reliability Analysis

Investigate use of Markov Reliability Analysis and DOE for System-Level test planning

Complementary with increasing emphasis on Model-Based design within DoD

Approach similar to human performance evaluation: Inject failure conditions during training to force off-nominal decisions

Feedback performance data to model over time to improve predictions of future reliability – continuous improvement

Selected Unmanned Underwater Vehicle (UUV) for Case Study

Highly autonomous operations in complex environment

Strong interest from community in testing improvements

Implementation

Concept of

Operations

Architecture

Requirements

Design Integration, Test

& Verification

System

Verification &

Validation

Operations &

Maintenance

Page 4: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Testing Robustness to Build Confidence

Increase Test Coverage with Failure & Environmental Conditions

March 16, 2011 4

Co

nfi

de

nc

e

0

Tests Defined Only

by Requirements

Increase Confidence

Gained/Time

t

Increased Coverage

by Failure &

Expanded

Environmental test

design

Page 5: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Behavioral Markov Reliability Analysis

System Markov Model

System component connections & logical dependencies

Reliability values for each system component (MTBF)

Model Outputs

Probabilities

– Any failure condition over system life

– System Loss

Reliability Metrics

– Overall Reliability (not directly used in this project)

– Sensitivity of Overall Reliability to failure rates of components (used to rank importance of failure modes)

Draper developed PARADyM Tool

1

2

3

4

Failure

rate = a

Failure

rate = b Failure

rate = c

a

b

c

Operational State

System Loss State

0 FL 1 FL

tcPdt

dP

tbPdt

dP

taPdt

dP

tPcbadt

dP

14

13

12

11

tcba

tcba

tcba

tcba

cetP

betP

aetP

etP

4

3

2

1

Solve

P(System Loss) = Σ(System Loss States)

Reliability = Σ(Operational States)

Motor

Pu

mp

Tank

March 16, 2011

Page 6: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Simulation

Test Matrix Simulation

Test Matrix

Required Inputs

Behavioral Markov Model

Extreme types and ranges of environmental conditions

Simulation Test Design

Perform Markov reliability sensitivity analysis

DOE for environmental conditions

Repeat all (or top subset) failure conditions for each experiment

Simulation Execution & Analysis

Parallel execution of test cases

Analysis of Variance to find Main & Interaction Effects

Rank significant factors according to reliability sensitivity

Final Results

Possible (not yet attempted) to extract confidence intervals for performance over bounds of operation

Highest significance subset of recommended tests to exercise in field

Process Summary

March 16, 2011 6

Markov Reliability Analysis Design of Experiments

High Fidelity Vehicle Simulation

with Failure Injection

Simulation

Test Matrix (i)

Test Results Test Results Simulation Test

Results (j)

Connectivity

Behavioral Model

Main Effects & Interaction

Analysis (DOE)

Environmental

Factors & Levels

Integrate Failure Cases

w/ Each Experiment

Field Test

Matrix

Ranking Against Reliability

Performance

Confidence

Intervals

Page 7: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Case Study: Generic UUV

Based on NUWC MARV UUV

1’ Diameter, 12’ Long

Max Speed: 5 knots

Prop Driven with 4 Control fins

Forward, Left, Right, Down Looking Sonars

ASTM F41 Software Architecture

Primary decision making in Autonomous Controller (AC)

Vehicle management by Vehicle Controller (VC)

Payload operations through Payload Controller (PC)

“Backseat Driver” Paradigm of control

March 16, 2011 7

PC Controls

Vehicle

Payload

Sensors

AC Performs Mission

Planning, Commands

Steering & Speed,

Payload Use

Scheduling, SA VC Performs Low-

Level Vehicle

Control and

Management

Page 8: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

UUV Simulation Based Testing

Draper Simulation Framework (DSF)

Govt. Open Framework

Dynamics/Physics simulation

Soft to Hard Real-Time and faster

Built for Hardware-in-Loop

MARV UUV Simulation

Validated vehicle dynamics

Simplified sensor models

Autonomy Controller running Software-in-Loop with simulated environment

New Extensions to Simulation

Created generalized failure injection nodes for DSF

Failure types: Omission/Constant, Noise, Bias

Parallel execution of simulations & Autonomy Controllers

March 16, 2011 8

Sink 1

Sink 2

Injected

Failure Source

New Failure Nodes Inserted

Page 9: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

UUV System Responses

March 16, 2011 9

Response Description Rationale

Position Error (t) Deviation from baseline

mission path over time

Position errors cause data

collection errors

Attitude Error (t)

[φ,θ,ψ]

Deviation from baseline

attitude over time

Attitude errors cause data

collection errors

Speed Error (t) Deviation from baseline speed

over time

Speed influences

execution time, stealth,

energy

Energy Consumption Energy consumption for

mission

Must operate within

available energy limits

Mission Time Total mission time Establish expectations for

recovery/communication

Surface Position Error Deviation from designated

end-of-mission surface point

Large errors on surfacing

impact recovery

Vehicle Recoverable TRUE if vehicle surfaced Lost at sea?

Page 10: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Case Study Evaluation Scenario

Scenario Goals

Short, rapid to iterate

Exercises terrain avoidance

Exercises waypoint following

Varies ocean currents, map quality

Case Study Scenario Design

Short mission, ~ 300 seconds

Approach & avoid terrain on way to waypoint

Basis of all case study simulations

Future Scenario Designs

Longer missions

More terrain complexity

Multiple time-varying objects of interest (ships, mines)

March 16, 2011 10

Page 11: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Environmental Experiment Design

Available Environmental Factors (3)

Uniform current magnitude & direction

Terrain under vehicle

DOE Design

2 Level, 3 Factor Full Factorial – using min/max levels, but adding median center point experiments

Center points show non-linearity in response, inform analysis

March 16, 2011

11

Min Median Max*

Current

Magnitude 0 Knots 2Knots 4Knots

Current

Direction 0° 90° 180°

Map

Mismatch 0% 50% 100%

180°

90°

50%

Mismatch

100%

Mismatch

RunOrder CenterPt

Current Magnitude

(knots)

Current Direction

(deg)

Map Mismatch

(%)

1 1 4 0 100

2 1 4 180 100

3 1 0 0 100

4 0 2 90 50

5 0 2 90 50

6 1 0 0 0

7 1 4 180 0

8 0 2 90 50

9 1 0 180 100

10 1 4 0 0

11 0 2 90 50

12 1 0 180 0

Experiment Design with Center Points

0%

Mismatch

Actual Terrain A priori Terrain Map

3/8/11 – Learned 4knot 0deg current cases too strong for vehicle

Page 12: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Example Results: Position Response

March 16, 2011

12

-400-300

-200-100

0100

200300

0

500

1000

1500

2000

2500

0

20

40

60

80

100

120

Crosstrack (ft)

UUV Path During Select Bathymetric Sonar Failures

Downrange (ft)

Dep

th (

ft)

Nominal Failure Case

Bathy Fails, Map Mismatch 100%

Bathy Fails, 50% Map Mismatch, 2 knot Side Current

Bathy Fails, 4 knot Tail Current, Mission Incomplete

System Baseline

Page 13: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Map Mismatch Significant Influence during Sonar Failures

Logical result

Almost 4 km Max Error in Surface Position

From Markov model, sonar failures drive reliability

Fin & attitude sensor failures much less probable

Failure effects same magnitude as environment only

Suspect impact cases and 4knot head currents biasing results

Need to set bounds on responses

Define overall PASS/FAIL limits

Summarize high level results more clearly

Example Results: Map Mismatch Effects

March 16, 2011

13

nominal

failure_sonar_rslsOmission

failure_sonar_lslsOmission

failure_sonar_flsOmission

failure_sonar_bathyOmission

failure_fins_fin7Omission

failure_fins_fin6Omission

failure_fins_fin4Omission

failure_attitude_qOmission

failure_attitude_pOmission

failure_attitude_phiOmission

4000

3500

3000

2500

2000

1500

1000

500

Su

rfa

ce

Po

sit

ion

Err

or

(m)

Boxplot of Surface Position Error

100 Map Mismatch

0 Map Mismatch(ft)

Page 14: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Demonstrated Reliability + DOE Test Planning method on Generic UUV case

Reliability analysis indicated sonars, battery monitor, VC, and AC primary drivers of system reliability

DOE Planning and analysis indicated Map Mismatch, Current, subset of failure modes significant

Need to complete analysis of simulated experiments

Review results with engineering, end-users, and customers to get feedback on usefulness

Rank effects and interactions against probability of failure conditions

Invest in method & tool improvements

Simulation Environment: Needs more fidelity in water properties, coupled with higher fidelity sensor models

Simulation Environment: Integrate reliability calculations with dynamic system model -> Avoid second model creation effort

Markov Analysis: Sources of reliability values (MTBF) for each component

Simulation Environment: Add failure mechanisms for VC and AC during simulation

Simulation Environment: Integrate autonomous controller decision logs with response data

Simulation Environment: Add time-varying failure and environmental perturbations during simulation

Design of Experiments: Also consider for integration with Simulation

Design of Experiments: Selection of best designs and analysis strategies for higher-order experiments

Summary & Future Work

March 16, 2011 14

Page 15: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

Innovations in Engineering

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Supplemental Slides

March 16, 2011 15

Page 16: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Assistant Secretary of the Navy (ASN) Research Development and Acquisition (RDA)

Large scale multi-unit test scenarios with many interoperating systems

Amy Markowich

Marine Corps Warfighting Lab

Extensive hands-on evaluation of aerial/ground robotics in relevant environments & missions

Jim Lasswell

NAVSEA (Combatant Craft Division)

In-Water testing of USV, advocates for division of testing at key interfaces – Perception, Effectors, Planning & Control

Eric Hansen

US Army Maneuver Battle Lab

Live/Virtual/Constructive testing with manned and unmanned systems

Harry Lubin

Army Research Laboratory (ARL)

Autonomous ground vehicle behavior testing with NIST partnership

Marshal Childers

MIT PATFrame

TRMC funded development of test planning framework for SoS

Ricardo Valerdi

Ongoing Testing Efforts of Note

March 16, 2011 16

Page 17: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

Example Results: Current Direction Effects

March 16, 2011 17

nominal

failure_sonar_rslsOmission

failure_sonar_lslsOmission

failure_sonar_flsOmission

failure_sonar_bathyOmission

failure_propulsion_omission

failure_pressure_omission

failure_fins_fin7Omission

failure_fins_fin6Omission

failure_fins_fin5Omission

failure_fins_fin4Omission

failure_attitude_thetaOmission

failure_attitude_qOmission

failure_attitude_pOmission

failure_attitude_phiOmission

7000

6000

5000

4000

3000

2000

1000

0Su

rfa

ce

Po

sit

ion

Err

or

(m)

Boxplot of Surface Position Error

180 Current Direction0 Current Direction

Current Direction Strong Effect

Logical result

Almost 4 km Max Error in Surface Position

From Markov model, sonar failures drive reliability

Fin, Prop, & attitude sensor failures much less probable

Failure effects same magnitude as environment

Suspect impact cases and 4knot head currents biasing results

Need to set bounds on responses

Define overall PASS/FAIL limits

Summarize high level results more clearly

(ft)

Page 18: Probability Driven Autonomous Systems Testing · 3 1 0 0 100 4 0 2 90 50 5 0 2 90 50 6 1 0 7 1 4 180 0 8 0 2 90 50 9 1 0 180 100 10 1 4 0 0 11 0 2 90 50 12 1 0 180 0 Experiment Design

UNCLASSIFIED PUBLIC RELEASE Copyright 2011 by the Charles Stark Draper Laboratory, Inc. all rights reserved

March 16, 2011 18

-200

-100

0

100

200

300

0

500

1000

1500

2000

2500

0

20

40

60

80

100

120

Crosstrack (ft)

UUV Path For Select Forward-Looking Sonar Failures

Downrange (ft)

Nominal FLS Failure Case

FLS Fails, Map Mismatch 100%

FLS Fails, Map Mismatch 50%, 2 knot Side Current

FLS Fails, 4 knot Tail Current, Mission Incomplete

System Baseline


Recommended