Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Assessing safety-critical systems from operational testing: : A study on autonomous vehicles

Published: 01 December 2020 Publication History

Abstract

Context

Demonstrating high reliability and safety for safety-critical systems (SCSs) remains a hard problem. Diverse evidence needs to be combined in a rigorous way: in particular, results of operational testing with other evidence from design and verification. Growing use of machine learning in SCSs, by precluding most established methods for gaining assurance, makes evidence from operational testing even more important for supporting safety and reliability claims.

Objective

We revisit the problem of using operational testing to demonstrate high reliability. We use Autonomous Vehicles (AVs) as a current example. AVs are making their debut on public roads: methods for assessing whether an AV is safe enough are urgently needed. We demonstrate how to answer 5 questions that would arise in assessing an AV type, starting with those proposed by a highly-cited study.

Method

We apply new theorems extending our Conservative Bayesian Inference (CBI) approach, which exploit the rigour of Bayesian methods while reducing the risk of involuntary misuse associated (we argue) with now-common applications of Bayesian inference; we define additional conditions needed for applying these methods to AVs.

Results

Prior knowledge can bring substantial advantages if the AV design allows strong expectations of safety before road testing. We also show how naive attempts at conservative assessment may lead to over-optimism instead; why extrapolating the trend of disengagements (take-overs by human drivers) is not suitable for safety claims; use of knowledge that an AV has moved to a “less stressful” environment.

Conclusion

While some reliability targets will remain too high to be practically verifiable, our CBI approach removes a major source of doubt: it allows use of prior knowledge without inducing dangerously optimistic biases. For certain ranges of required reliability and prior beliefs, CBI thus supports feasible, sound arguments. Useful conservative claims can be derived from limited prior knowledge.

References

[1]
B. Littlewood, L. Strigini, Validation of ultra-high dependability for software-based systems, Comm. ACM 36 (1993) 69–80.
[2]
R.W. Butler, G.B. Finelli, The infeasibility of quantifying the reliability of life-critical real-time software, IEEE Trans. Softw. Eng. 19 (1) (1993) 3–12.
[3]
R. Bloomfield, H. Khlaaf, P.R. Conmy, G. Fletcher, Disruptive innovations and disruptive assurance: Assuring machine learning and autonomy, Computer 52 (9) (2019) 82–89.
[4]
E. Alves, D. Bhatt, B. Hall, K. Driscoll, A. Murugesan, J. Rushby, Considerations in assuring safety of increasingly autonomous systems, Technical Report NASA/CR-2018-220080, NASA, 2018.
[5]
S. Burton, I. Habli, T. Lawton, J. McDermid, P. Morgan, Z. Porter, Mind the gaps: Assuring the safety of autonomous systems from an engineering, ethical, and legal perspective, Artif. Intell. 279 (2020) 103201.
[6]
C. Urmson, Hands off: The future of self-driving cars, Testimony, Committee on Commerce, Science and Transportation, Washington, D.C., USA, 2016.
[7]
S.S. Banerjee, S. Jha, J. Cyriac, Z.T. Kalbarczyk, R.K. Iyer, Hands off the wheel in autonomous vehicles?: A systems perspective on over a million miles of field data, 48th IEEE/IFIP Int. Conf. on Dependable Systems and Networks, 2018, pp. 586–597.
[8]
N. Kalra, S. Paddock, Driving to safety: How many miles of driving would it take to demonstrate autonomous vehicle reliability?, Transp. Res. Part A 94 (2016) 182–193.
[9]
F. Favarò, S. Eurich, N. Nader, Autonomous vehicles’ disengagements: Trends, triggers, and regulatory limitations, Accident Anal. Prevent. 110 (2018) 136–148.
[10]
V.V. Dixit, S. Chand, D.J. Nair, Autonomous vehicles: Disengagements, accidents and reaction times, PLoS ONE 11 (12) (2016) 1–14.
[11]
C. Lv, D. Cao, Y. Zhao, D.J. Auger, M. Sullman, H. Wang, L.M. Dutka, L. Skrypchuk, A. Mouzakitis, Analysis of autopilot disengagements occurring during autonomous vehicle testing, IEEE/CAA J. Autom. Sin. 5 (1) (2018) 58–68.
[12]
J.M. Anderson, K. Nidhi, K.D. Stanley, P. Sorensen, C. Samaras, O.A. Oluwatola, Autonomous vehicle technology: A guide for policymakers, Technical Report RR-443-2-RC, Rand Corporation, 2016.
[13]
Matthew Wood, Philipp Robbel et al, Safety first for automated driving, 2019. URL: https://www.daimler.com/documents/innovation/other/safety-first-for-automated-driving.pdf.
[14]
Waymo, Waymo safety report: On the road to fully self-driving, Technical Report, 2018.
[15]
A. Shashua, S. Shalev-Shwartz, A plan to develop safe autonomous vehicles. And prove it, Intel Newsroom (2017).
[16]
Y. Tian, K. Pei, S. Jana, B. Ray, DeepTest: Automated testing of deep-neural-network-driven autonomous cars, the 40th Int. Conf. on Software Engineering, New York, NY, USA, 2018, pp. 303–314.
[17]
X. Huang, M. Kwiatkowska, S. Wang, M. Wu, Safety verification of deep neural networks, Computer Aided Verification, LNCS, 10426, Springer International Publishing, Cham, 2017, pp. 3–29.
[18]
M. Fisher, L. Dennis, M. Webster, Verifying autonomous systems, Commun. ACM 56 (9) (2013) 84–93,.
[19]
M. Fisher, E. Collins, L. Dennis, M. Luckcuck, M. Webster, M. Jump, V. Page, C. Patchett, F. Dinmohammadi, D. Flynn, V. Robu, X. Zhao, Verifiable self-certifying autonomous systems, the 29th Int. Symp. on Software Reliability Engineering Workshops, IEEE, 2018, pp. 341–348.
[20]
P. Koopman, A. Kane, J. Black, Credible autonomy safety argumentation, the 27th Safety-Critical Systems Symposium, Safety-Critical Systems Club, Bristol, UK, 2019.
[21]
X. Zhao, V. Robu, D. Flynn, K. Salako, L. Strigini, Assessing the Safety and Reliability of Autonomous Vehicles from Road Testing, the 30th Int. Symp. on Software Reliability Engineering, IEEE, Berlin, Germany, 2019, pp. 13–23.
[22]
P. Bishop, R. Bloomfield, B. Littlewood, A. Povyakalo, D. Wright, Toward a formalism for conservative claims about the dependability of software-based systems, IEEE Trans. Softw. Eng. 37 (5) (2011) 708–717.
[23]
L. Strigini, A. Povyakalo, Software fault-freeness and reliability predictions, in: F. Bitsch, J. Guiochet, M. Kaniche (Eds.), Computer Safety, Reliability, and Security, in: LNCS, 8153, Springer Berlin Heidelberg, Berlin, Heidelberg, 2013, pp. 106–117.
[24]
X. Zhao, B. Littlewood, A. Povyakalo, L. Strigini, D. Wright, Modeling the probability of failure on demand (pfd) of a 1-out-of-2 system in which one channel is “quasi-perfect”, Reliab. Eng. Syst. Saf. 158 (2017) 230–245.
[25]
X. Zhao, B. Littlewood, A. Povyakalo, D. Wright, Conservative claims about the probability of perfection of software-based systems, 26th Int. Symp. on Software Reliability Eng., IEEE, 2015, pp. 130–140.
[26]
X. Zhao, B. Littlewood, A. Povyakalo, L. Strigini, D. Wright, Conservative claims for the probability of perfection of a software-based system using operational experience of previous similar systems, Reliab. Eng. Syst. Saf. 175 (2018) 265–282.
[27]
D.R. Miller, Exponential order statistic models of software reliability growth, IEEE Trans. Softw. Eng. 12 (01) (1986) 12–24.
[28]
IEC, IEC61508, Functional Safety of Electrical/ Electronic/Programmable Electronic Safety Related Systems, International Electrotechnical Commission (IEC), 2010.
[29]
CENELEC, EN 50129:2018: Railway applications - communication, signalling and processing systems - safety related electronic systems for signalling, European Committee for Electrotechnical Standardization (CENELEC), 2018.
[30]
C. Atwood, J. LaChance, H. Martz, D. Anderson, M. Englehardt, D. Whitehead, T. Wheeler, Handbook of parameter estimation for probabilistic risk assessment, Report NUREG/CR-6823, U.S. Nuclear Regulatory Commission, Washington, DC, 2003.
[31]
L. Strigini, B. Littlewood, Guidelines for Statistical Testing, Project Report PASCON/WO6-CCN2/TN12, City University London, 1997.
[32]
J. May, G. Hughes, A.D. Lunn, Reliability estimation from appropriate testing of plant protection software, Softw. Eng. J. 10 (6) (1995) 206–218.
[33]
G. Walter, L.J.M. Aslett, F.P.A. Coolen, Bayesian nonparametric system reliability using sets of priors, Int. J. Approx. Reason. 80 (2017) 67–88.
[34]
P. Bishop, A. Povyakalo, Deriving a frequentist conservative confidence bound for probability of failure per demand for systems with different operational and test profiles, Reliab. Eng. Syst. Saf. 158 (2017) 246–253.
[35]
L.V. Utkin, F.P.A. Coolen, Imprecise probabilistic inference for software run reliability growth models., J. Uncertain Syst. 12 (4) (2018) 292–308.
[36]
P. Popov, L. Strigini, Assessing asymmetric fault-tolerant software, Proc. of the 21st Int. Symp. on Software Reliability Engineering, IEEE Computer Society Press, San Jose, CA, USA, 2010, pp. 41–50.
[37]
F.M. Favarò, N. Nader, S.O. Eurich, M. Tripp, N. Varadaraju, Examining accident reports involving autonomous vehicles in California, PLoS ONE 12 (9) (2017) 1–20.
[38]
P. Liu, R. Yang, Z. Xu, How safe is safe enough for self-driving vehicles?, Risk Anal. 39 (2) (2019) 315–325.
[39]
A. Bertolino, L. Strigini, Assessing the risk due to software faults: Estimates of failure rate vs evidence of perfection, Softw. Test. Verificat. Reliab. 8 (3) (1998) 155–166.
[40]
B. Littlewood, J. Rushby, Reasoning about the reliability of diverse two-channel systems in which one channel is ‘possibly perfect’, IEEE Tran. Softw. Eng. 38 (5) (2012) 1178–1194.
[41]
J.O. Berger, Could Fisher, Jeffreys and Neyman have agreed on testing?, Stat. Sci. 18 (1) (2003) 1–32.
[42]
L. Strigini, D. Wright, Bounds on survival probability given mean probability of failure per demand; and the paradoxical advantages of uncertainty, Reliab. Eng. Syst. Saf. 128 (2014) 66–83.
[43]
S. Brocklehurst, B. Littlewood, Techniques for prediction analysis and recalibration, in: M. Lyu (Ed.), Handbook of Software Reliability Eng., McGraw-Hill & IEEE Computer Society Press, 1996, pp. 119–166.
[44]
S. Brocklehurst, P.Y. Chan, B. Littlewood, J. Snell, Recalibrating software reliability models, IEEE Trans. Softw. Eng. 16 (4) (1990) 458–470.
[45]
P. Koopman, B. Osyk, Safety argument considerations for public road testing of autonomous vehicles, SAE Int. J. Adv. Curr. Pract. Mobil. 1 (2) (2019) 512–523.
[46]
R.D. Sorkin, D.D. Woods, Systems with human monitors: A signal detection analysis, Human-Comput. Interact. 1 (1) (1985) 49–75.
[47]
L. Strigini, A. Povyakalo, E. Alberdi, Human-Machine diversity in the use of computerised advisory systems: A case study, Annual IEEE/IFIP Int. Conf. on Dependable Systems and Networks, DSN’03, IEEE Computer Society, San Francisco, CA, USA, 2003, pp. 249–258,.
[48]
B. Littlewood, K. Salako, L. Strigini, X. Zhao, On reliability assessment when a software-based system is replaced by a thought-to-be-better one, Reliab. Eng. Syst. Saf. 197 (2020) 106752.
[49]
S. Pathak, L. Pulina, A. Tacchella, Verification and repair of control policies for safe reinforcement learning, Appl. Intell. 48 (4) (2018) 886–908.
[50]
X. Zhao, V. Robu, D. Flynn, F. Dinmohammadi, M. Fisher, M. Webster, Probabilistic model checking of robots deployed in extreme environments, Proc. of the 33rd AAAI Conference on Artificial Intelligence, 33, 2019, pp. 8076–8084.
[51]
X. Zhao, A. Banks, J. Sharp, V. Robu, D. Flynn, M. Fisher, X. Huang, A safety framework for critical systems utilising deep neural networks, SafeComp2020, LNCS, Springer, 2020.
[52]
P. Koopman, M. Wagner, Autonomous vehicle safety: An interdisciplinary challenge, IEEE Intell. Transp. Syst. Mag. 9 (1) (2017) 90–96.
[53]
B. Cukic, D. Chakravarthy, Bayesian framework for reliability assurance of a deployed safety critical system, Proc. of the 5th Int. Symp. on High Assurance Systems Engineering, IEEE, Albuquerque, NM, USA, 2000, pp. 321–329.
[54]
C. Smidts, B. Cukic, E. Gunel, M. Li, H. Singh, Software reliability corroboration, Proc. of the 27th Annual NASA Goddard/IEEE Software Engineering Workshop, 2002, pp. 82–87.
[55]
C.W. Johnson, The increasing risks of risk assessment: on the rise of artificial intelligence and non-determinism in safety-critical systems, the 26th Safety-Critical Systems Symposium, Safety-Critical Systems Club, York, UK., 2018, p. 15.
[56]
B. Littlewood, L. Strigini, ‘Validation of ultra-high dependability... ’ – 20 years on, Saf. Syst. Newsletter Saf.-Critic. Syst. Club 20 (3) (2011).
[57]
L. Strigini, On testing process control software for reliability assessment: the effects of correlation between successive failures, Softw. Test. Verificat. Reliab. 6 (1) (1996) 33–48.
[58]
K. Goseva-Popstojanova, K.S. Trivedi, Failure correlation in software reliability models, IEEE Trans. Reliab. 49 (1) (2000) 37–48,.
[59]
L.A. Tomek, J.K. Muppala, K.S. Trivedi, Modeling correlation in software recovery blocks, IEEE Trans. Softw. Eng. 19 (11) (1993) 1071–1086.

Cited By

View all
  • (2024)AutoGrAN: Autonomous Vehicle LiDAR Contaminant Detection using Graph Attention NetworksCompanion of the 15th ACM/SPEC International Conference on Performance Engineering10.1145/3629527.3652896(112-119)Online publication date: 7-May-2024
  • (2023)Reliability Assessment and Safety Arguments for Machine Learning Components in System AssuranceACM Transactions on Embedded Computing Systems10.1145/357091822:3(1-48)Online publication date: 20-Apr-2023
  • (2022)Dependable learning-enabled multiagent systemsAI Communications10.3233/AIC-22012835:4(407-420)Online publication date: 1-Jan-2022
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Information and Software Technology
Information and Software Technology  Volume 128, Issue C
Dec 2020
155 pages

Publisher

Butterworth-Heinemann

United States

Publication History

Published: 01 December 2020

Author Tags

  1. Autonomous systems
  2. Safety assurance
  3. Statistical testing
  4. Safety-critical systems
  5. Ultra-high reliability
  6. Conservative Bayesian inference
  7. AI safety
  8. Proven in use
  9. Globally at least equivalent
  10. Software reliability growth models

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 03 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2024)AutoGrAN: Autonomous Vehicle LiDAR Contaminant Detection using Graph Attention NetworksCompanion of the 15th ACM/SPEC International Conference on Performance Engineering10.1145/3629527.3652896(112-119)Online publication date: 7-May-2024
  • (2023)Reliability Assessment and Safety Arguments for Machine Learning Components in System AssuranceACM Transactions on Embedded Computing Systems10.1145/357091822:3(1-48)Online publication date: 20-Apr-2023
  • (2022)Dependable learning-enabled multiagent systemsAI Communications10.3233/AIC-22012835:4(407-420)Online publication date: 1-Jan-2022
  • (2022)Autonomous Vehicles and Intelligent AutomationMobile Information Systems10.1155/2022/76328922022Online publication date: 1-Jan-2022

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media