+ All Categories
Home > Documents > 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay....

1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay....

Date post: 27-Mar-2015
Category:
Upload: logan-roy
View: 213 times
Download: 0 times
Share this document with a friend
Popular Tags:
21
Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301 – Software Engineering Lecture # 32 – 2004-11-15 M. E. Kabay, PhD, CISSP Assoc. Prof. Information Assurance Division of Business & Management, Norwich University mailto:[email protected] V: 802.479.7937
Transcript
Page 1: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Critical Systems Validation

IS301 – Software EngineeringLecture # 32 – 2004-11-15

M. E. Kabay, PhD, CISSPAssoc. Prof. Information Assurance

Division of Business & Management, Norwich University

mailto:[email protected] V: 802.479.7937

Page 2: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

2 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Objectives

To explain how system reliability can be measured and how reliability growth models can be used for reliability prediction

To describe safety arguments and how these are used

To discuss the problems of safety assuranceTo introduce safety cases and how these are

used in safety validation

Page 3: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

3 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Topics covered

Reliability validationSafety assuranceSecurity assessmentSafety and dependability cases

Page 4: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

4 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Validation of critical systems

The verification and validation costs for critical systems involves more validation processes and analysis than for non-critical systems:The costs and consequences of failure are

high so it is cheaper to find and remove faults than to pay for system failure;

You may have to make a formal case to customers or to a regulator that the system meets its dependability requirements. This dependability case may require specific V & V activities to be carried out.

Page 5: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

5 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Validation costs

Because of the additional activities involved, the validation costs for critical systems are usually significantly higher than for non-critical systems.

Normally, V & V costs take up more than 50% of the total system development costs.

Page 6: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

6 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Reliability validation

Reliability validation involves exercising the program to assess whether or not it has reached the required level of reliability.

This cannot normally be included as part of a normal defect testing process because data for defect testing is (usually) atypical of actual usage data.

Reliability measurement therefore requires a specially designed data set that replicates the pattern of inputs to be processed by the system.

Page 7: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

7 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

The reliability measurement process

Computeobservedreliability

Apply tests tosystem

Prepare testdata set

Identifyoperational

profiles

Page 8: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

8 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Reliability validation activities

Establish the operational profile for the system.

Construct test data reflecting the operational profile.

Test the system and observe the number of failures and the times of these failures.

Compute the reliability after a statistically significant number of failures have been observed.

Page 9: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

9 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Statistical testing

Testing software for reliability rather than fault detection.

Measuring the number of errors allows the reliability of the software to be predicted. Note that, for statistical reasons, more errors than are allowed for in the reliability specification must be induced.

An acceptable level of reliability should be specified and the software tested and amended until that level of reliability is reached.

Page 10: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

10 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Reliability measurement problems

Operational profile uncertaintyThe operational profile may not be an

accurate reflection of the real use of the system.

High costs of test data generationCosts can be very high if the test data for

the system cannot be generated automatically.

Statistical uncertaintyYou need a statistically significant number

of failures to compute the reliability but highly reliable systems will rarely fail.

Page 11: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

11 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Operational profiles

An operational profile is a set of test data whose frequency matches the actual frequency of these inputs from ‘normal’ usage of the system. A close match with actual usage is necessary otherwise the measured reliability will not be reflected in the actual usage of the system.

It can be generated from real data collected from an existing system or (more often) depends on assumptions made about the pattern of usage of a system.

Page 12: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

12 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

An operational profile

...

Number ofinputs

Input classes

Page 13: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

13 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Operational profile generation

Should be generated automatically whenever possible.

Automatic profile generation is difficult for interactive systems.

May be straightforward for ‘normal’ inputs but it is difficult to predict ‘unlikely’ inputs and to create test data for them.

Page 14: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

14 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Reliability predictionA reliability growth model is a mathematical

model of the system reliability change as it is tested and faults are removed.

It is used as a means of reliability prediction by extrapolating from current dataSimplifies test planning and customer

negotiations.You can predict when testing will be

completed and demonstrate to customers whether or not the reliability growth will ever be achieved.

Prediction depends on the use of statistical testing to measure the reliability of a system version.

Page 15: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

15 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Equal-step reliability growth

Reliability(ROCOF)

t1 t2 t3 t4 t5Time

Page 16: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

16 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Observed reliability growth

The equal-step growth model is simple but it does not normally reflect reality.

Reliability does not necessarily increase with change as the change can introduce new faults.

The rate of reliability growth tends to slow down with time as frequently occurring faults are discovered and removed from the software.

A random-growth model where reliability changes fluctuate may be a more accurate reflection of real changes to reliability.

Page 17: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

17 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Random-step reliability growth

t1 t2 t3 t4 t5Time

Note different reliabilityimprovements

Fault repair adds new faultand decreases reliability(increases ROCOF)

Reliability(ROCOF)

Note different reliability improvements

Fault repair adds new faultsand decreases reliability(increases ROCOF)

Page 18: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

18 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Growth model selection

Many different reliability growth models have been proposed.

There is no universally applicable growth model.

Reliability should be measured and observed data should be fitted to several models.

The best-fit model can then be used for reliability prediction.

Page 19: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

19 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Reliability prediction

Reliability

Requiredreliability

Fitted reliabilitymodel curve

Estimatedtime of reliability

achievement

Time

= Measured reliabilityMeasured reliability

Fitted reliability model curveRe

liab

ilit

y

RequiredReliability

Estimated time ofreliability achievement

Page 20: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

41 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

Homework

RequiredBy Mon 29 Nov 2004For 20 points,

24.2 (@4), 24.4 (@12)—details!, 24.5 (@4)

OptionalBy Wed 1 Dec 2004For a maximum of 5 points,

24.11 – THINK about it before you write

Page 21: 1 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved. Critical Systems Validation IS301.

42 Note content copyright © 2004 Ian Sommerville. NU-specific content copyright © 2004 M. E. Kabay. All rights reserved.

DISCUSSION


Recommended