Abstract
Growing AI technologies are a threat to safety and security in systems due to its obscurity and uncertainty. This study introduces a prevailing Deep Learning model, Convolutional Neural Network (CNN) and it’s deep weaknesses through a simple case study of the CNN model based on Keras for handwriting recognition. It reveals that CNN algorithms don’t adapt well to changes. Adding new cases to the training data may improve accuracy, but not to the same level as before. Synthetic training data may improve the accuracy superficially because of the similarity of data distributions between generated data and original data. Prevailing ML models such as Generative Adversarial Networks (GAN) have their limitations such as similarity-addiction and modality collapse. They could be toxic to safety engineering without domain expertise.
The study proposed four test strategies: 1) AI systems should be tested by the third parties, not the developers; 2) test datasets should be categorically different from training datasets; the test data should not be a part of the training data; the test data should be collected from independent sources to increase the “diversity” of data modality; 3) avoid fake data, or simulated data; and 4) don’t collect the data that are conveniently available, but actively collect disastrous event data, unexpected, or the worst scenarios that may destroy the model. The study also introduces a multidimensional checklist for AI safety analysis, including sensors, data and environments, default and recovery mode, system architectures, and human-system interaction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
https://ujjwalkarn.me/2016/08/11/intuitive-explanation-convnets/
Marko, K.: AI today and tomorrow is mostly about curve fitting, not intelligence, Diginomia, 3 June 2018
https://towardsdatascience.com/deep-learning-performance-cheat-sheet-21374b9c4
https://towardsdatascience.com/deep-learning-performance-cheat-sheet-21374b9c4f45
https://en.wikipedia.org/wiki/Generative_adversarial_network
Cai, Y.: Instinctive Computing. Springer, London (2016). https://doi.org/10.1007/978-1-4471-7278-9
DARPA Grand Challenge. (2016) https://en.wikipedia.org/wiki/DARPA_Grand_Challenge
Newman, M.E.J.: Networks: An Instruction. Oxford University Press, Oxford (2010)
Acknowledgement
The study is sponsored by ONR Summer Faculty Research Program in 2019. The author is grateful to Dr. Rani A. Kady for his discussions and Mr. Weizhe Sun for the experiments with the CNN model.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Cai, Y. (2020). Safety Analytics for AI Systems. In: Stephanidis, C., Kurosu, M., Degen, H., Reinerman-Jones, L. (eds) HCI International 2020 - Late Breaking Papers: Multimodality and Intelligence. HCII 2020. Lecture Notes in Computer Science(), vol 12424. Springer, Cham. https://doi.org/10.1007/978-3-030-60117-1_32
Download citation
DOI: https://doi.org/10.1007/978-3-030-60117-1_32
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60116-4
Online ISBN: 978-3-030-60117-1
eBook Packages: Computer ScienceComputer Science (R0)