Trustworthy AI Guidelines in Biomedical Decision-Making Applications: A Scoping Review
Abstract
:1. Introduction
- What are the existing practices that implement trustworthy AI principles in medical research?
- What are the principles from the ethics guidelines that are addressed by these practices?
- What are the common approaches found in the existing literature?
- What gaps prevent the coverage of all principles and, therefore, system trustworthiness?
2. Materials and Methods
2.1. Data Sources and Systematic Searches
(AI OR “Artificial Intelligence” OR “deep learning” OR “machine learning” OR “language model”) AND (medical OR biomedical) AND (trustworthy)
2.2. Inclusion and Exclusion Criteria
- Papers had to be written in English.
- Papers had to be published in a journal or a conference proceeding.
- Papers had to be explicit and intentional about their trustworthiness objectives.
- Papers had to describe, totally or partially, a practical case where one or multiple trustworthy AI principles were covered or answered.
- Papers had to be published between 1 January 2019 and 10 April 2024.
- Papers written in languages other than English.
- Papers that are descriptive or prescriptive in nature or propose guidelines or theoretical frameworks without associated use cases or ways to implement them in practice. Thus, review articles were also excluded.
2.3. Identification and Selection of Studies
2.4. Limitations
3. Results
3.1. Overview
3.2. Human Agency and Oversight
- Aboutalebi et al. [40] develop an explainability-driven framework for machine learning models with two human-guided phases; a first one called “clinician-guided design” phase, where the dataset is preprocessed using explainable AI and domain expert input, and a second one named “explainability-driven design refinement”, where they employ explainability methods not only for transparency but also “to gain a deeper understanding of the clinical validity of the decisions made” along an expert clinician, using such insights to refine the model iteratively. However, such human agency is limited to building the model and does not describe or design a control loop for the ongoing operation of the model.
- Similarly, Lu et al. [42] include a step in their proposed workflow where medical doctors “label the differential diagnosis features with medical guidelines and professional knowledge”; this is supposed to black list meaningless features extracted from electronic health records (EHR). The authors claim to “reduce workloads of clinicians in human-in-loop data mining” as they use oversight features instead of full predictions.
- Bruckert et al. [19] highlight the challenge for “human decision makers to develop trust, which is much needed in life-changing decision tasks” and answer such a challenge with an integrated framework aimed at medical diagnosis. Their inclusion of interactive ML and human-in-the-loop learning ideas enables them to integrate human expert knowledge into ML models so “that humans and machines act as companions within a critical decision task”, and, thus, represents a significant step towards trustworthy and human-overseen expert decision systems.
- Finally, the work “On Assessing Trustworthy AI in Healthcare” by Zicari et al. [18] presents a general translation of the AI HLEG trustworthy AI guidelines to the practice in the healthcare domain, and that includes human agency and oversight. In this regard, they delve into the challenges posed by this requirement. In particular, they illustrate with a practical case the issue of determining the appropriate level of human involvement in the decision-making process, how the AI system might reduce human agent agency and autonomous decision making (which are often critical in emergency call management, for instance), the balance between both actors and, most importantly, the need for an informed decision-making criteria.
3.3. Technical Robustness and Safety
3.3.1. Safety, Including Accuracy and Reliability
3.3.2. Security and Resilience to Attacks
3.4. Privacy and Data Governance
- Use distributed learning to keep control over the data. This implies that the data can stay in the owning organization, which should have privacy controls in place without exposing such protected data to the exterior. It is a privacy-preserving approach that also implies a certain degree of data governance (even though it may not cover the full spectrum of the requirement).
- Use agreement techniques for federated/distributed learning to ensure the integrity and robustness of the model, which mainly refers to the technical robustness and safety principles but also points to a data/model quality control.
- Finally, blockchain techniques are suggested to keep the model confidential (with the required level of privacy or access) and also for traceability purposes—which relates to the transparency principle.
3.5. Transparency
3.6. Diversity, Nondiscrimination, and Fairness
- The Ordinary Evidence Model “implies a clear division of labor between the designers of algorithms and medical doctors who are to apply the algorithms in clinical practice”. Even though this is the most commonly found model, the authors claim that due to the accountability and opacity issues inherent to this approach, such “reliance on AI in medicine challenges and even disrupts the professional accountability upon which the ordinary evidence model rests”.
- The Ethical Design Model addresses ethical concerns in the design process. This approach was found in a few works that involve some relevant stakeholders (e.g., medical doctors), partially in the model design [42] or deployment [13]. However, as Gundersen and Baeroe [53] argue, this approach might be insufficient since stakeholder engagement and ethical deliberation are required both in design and in use.
- The Collaborative Model states that “collaboration and mutual engagement between medical doctors and AI designers are required to align algorithms with medical expertise, bioethics, and medical ethics”. The idea is to use expert input into AI systems (to calibrate them, make them accessible, and increase the user’s trust in the system) while also helping AI designers understand and interpret the outputs appropriately according to medical practice. Aboutalebi et al. [40] follow this approach, integrating clinicians in both the design and refinement phases. On the other hand, Kumar et al. [30] integrate usability and expert evaluation with doctors (or final users) as a final step in the model implementation to see if it is considered trustable in terms of decision making by the direct stakeholders. This approach might be limited, but it is arguably the only practice related to accessibility that has been found.
- The Public Deliberation Model involves other relevant stakeholders besides AI designers, bioethicists, and medical experts; policymakers and the general public are included in this group for broader discussions about the transformative impact of a new AI technology and/or societal impact. No practices were found that involved this approach.
3.7. Accountability
4. Discussion
4.1. Explainability Is Key
4.2. Efforts on Model Accuracy
4.3. Privacy Preservation
4.4. Testimonial Human Agency and Stakeholder Integration
4.5. Missing Techniques
4.6. Legal Loopholes
4.7. Lack of Holistic Approaches
4.8. Future Work
5. Conclusions
- Most identified practices focus on explainability, aiming to make AI systems more understandable to healthcare professionals, which is crucial for ethical and informed decision-making. However, the review points out inconsistencies among various explainability tools, suggesting a need for more reliable methods.
- Regarding technical robustness and safety, the practices mostly focus on achieving a balance between system accuracy and trustworthiness, with less emphasis on integrating fail-safe mechanisms and specific security measures against data manipulation. This indicates a gap in comprehensive safety and security strategies essential for patient safety.
- Privacy and data governance are also predominantly focused on privacy aspects rather than comprehensive data governance. The common use of Federated Learning and blockchain suggests a focus on innovative privacy-preserving technologies, yet comprehensive data management practices remain underdeveloped.
- Regarding human agency and oversight, the review notes that most practices support human control indirectly through explainability efforts rather than through direct mechanisms that ensure meaningful human oversight throughout the AI lifecycle. This points to a need for more explicit integration of human control in the operational use of AI.
Author Contributions
Funding
Conflicts of Interest
References
- Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef] [PubMed]
- Directorate-General for Communications Networks; Content and Technology (European Commission); Grupa Ekspertów Wysokiego Szczebla ds. Sztucznej Inteligencji. Ethics Guidelines for Trustworthy AI; Publications Office: Brussels, Belgium, 2019. [Google Scholar] [CrossRef]
- Yu, K.H.; Beam, A.L.; Kohane, I.S. Artificial intelligence in healthcare. Nat. Biomed. Eng. 2018, 2, 719–731. [Google Scholar] [CrossRef] [PubMed]
- Goirand, M.; Austin, E.; Clay-Williams, R. Implementing ethics in healthcare AI-based applications: A scoping review. Sci. Eng. Ethics 2021, 27, 61. [Google Scholar] [CrossRef] [PubMed]
- Tricco, A.C.; Lillie, E.; Zarin, W.; O’Brien, K.K.; Colquhoun, H.; Levac, D.; Moher, D.; Peters, M.D.; Horsley, T.; Weeks, L.; et al. PRISMA extension for scoping reviews (PRISMA-ScR): Checklist and explanation. Ann. Intern. Med. 2018, 169, 467–473. [Google Scholar] [CrossRef] [PubMed]
- Rong, G.; Mendez, A.; Assi, E.B.; Zhao, B.; Sawan, M. Artificial intelligence in healthcare: Review and prediction case studies. Engineering 2020, 6, 291–301. [Google Scholar] [CrossRef]
- Silcox, C.; Zimlichmann, E.; Huber, K.; Rowen, N.; Saunders, R.; McClellan, M.; Kahn, C.N.; Salzberg, C.A.; Bates, D.W. The potential for artificial intelligence to transform healthcare: Perspectives from international health leaders. NPJ Digit. Med. 2024, 7, 88. [Google Scholar] [CrossRef] [PubMed]
- Federico, C.A.; Trotsyuk, A.A. Biomedical Data Science, Artificial Intelligence, and Ethics: Navigating Challenges in the Face of Explosive Growth. Annu. Rev. Biomed. Data Sci. 2024, 7. [Google Scholar] [CrossRef] [PubMed]
- Alves, V.M.; Auerbach, S.S.; Kleinstreuer, N.; Rooney, J.P.; Muratov, E.N.; Rusyn, I.; Tropsha, A.; Schmitt, C. Models Out: The Impact of Data Quality on the Reliability of Artificial Intelligence Models as Alternatives to Animal Testing. Altern. Lab. Anim. 2021, 49, 73–82. [Google Scholar] [CrossRef] [PubMed]
- Moradi, M.; Samwald, M. Improving the robustness and accuracy of biomedical language models through adversarial training. J. Biomed. Inf. 2022, 132, 104114. [Google Scholar] [CrossRef]
- Ma, K.; He, S.; Sinha, G.; Ebadi, A.; Florea, A.; Tremblay, S.; Wong, A.; Xi, P. Towards Building a Trustworthy Deep Learning Framework for Medical Image Analysis. Sensors 2023, 23, 8122. [Google Scholar] [CrossRef]
- Khanna, V.V.; Chadaga, K.; Sampathila, N.; Prabhu, S.; Bhandage, V.; Hegde, G.K. A Distinctive Explainable Machine Learning Framework for Detection of Polycystic Ovary Syndrome. Appl. Syst. Innov. 2023, 6, 32. [Google Scholar] [CrossRef]
- Fidon, L.; Aertsen, M.; Kofler, F.; Bink, A.; David, A.L.; Deprest, T.; Emam, D.; Guffens, F.; Jakab, A.; Kasprian, G.; et al. A Dempster-Shafer Approach to Trustworthy AI with Application to Fetal Brain MRI Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 3784–3795. [Google Scholar] [CrossRef] [PubMed]
- Nambiar, A.; Harikrishnaa, S.; Sharanprasath, S. Model-agnostic explainable artificial intelligence tools for severity prediction and symptom analysis on Indian COVID-19 data. Front. Artif. Intell. 2023, 6, 1272506. [Google Scholar] [CrossRef] [PubMed]
- Rashid, M.M.; Askari, M.R.; Chen, C.; Liang, Y.; Shu, K.; Cinar, A. Artificial Intelligence Algorithms for Treatment of Diabetes. Algorithms 2022, 15, 299. [Google Scholar] [CrossRef]
- Kumar, M.; Moser, B.A.; Fischer, L.; Freudenthaler, B. An Information Theoretic Approach to Privacy-Preserving Interpretable and Transferable Learning. Algorithms 2023, 16, 450. [Google Scholar] [CrossRef]
- Salahuddin, Z.; Woodruff, H.C.; Chatterjee, A.; Lambin, P. Transparency of deep neural networks for medical image analysis: A review of interpretability methods. Comput. Biol. Med. 2022, 140, 105111. [Google Scholar] [CrossRef] [PubMed]
- Zicari, R.V.; Brusseau, J.; Blomberg, S.N.; Christensen, H.C.; Coffee, M.; Ganapini, M.B.; Gerke, S.; Gilbert, T.K.; Hickman, E.; Hildt, E.; et al. On Assessing Trustworthy AI in Healthcare. Machine Learning as a Supportive Tool to Recognize Cardiac Arrest in Emergency Calls. Front. Hum. Dyn. 2021, 3, 673104. [Google Scholar] [CrossRef]
- Bruckert, S.; Finzel, B.; Schmid, U. The Next Generation of Medical Decision Support: A Roadmap toward Transparent Expert Companions. Front. Artif. Intell. 2020, 3, 507973. [Google Scholar] [CrossRef]
- Ma, J.; Schneider, L.; Lapuschkin, S.; Achtibat, R.; Duchrau, M.; Krois, J.; Schwendicke, F.; Samek, W. Towards Trustworthy AI in Dentistry. J. Dent. Res. 2022, 101, 1263–1268. [Google Scholar] [CrossRef]
- Imboden, S.; Liu, X.; Payne, M.C.; Hsieh, C.J.; Lin, N.Y.C. Trustworthy in silico cell labeling via ensemble-based image translation. Biophys. Rep. 2023, 3, 100133. [Google Scholar] [CrossRef]
- Karim, M.R.; Islam, T.; Lange, C.; Rebholz-Schuhmann, D.; Decker, S. Adversary-Aware Multimodal Neural Networks for Cancer Susceptibility Prediction from Multiomics Data. IEEE Access 2022, 10, 54386–54409. [Google Scholar] [CrossRef]
- Mu, J.; Kadoch, M.; Yuan, T.; Lv, W.; Liu, Q.; Li, B. Explainable Federated Medical Image Analysis through Causal Learning and Blockchain. IEEE J. Biomed. Health Inform. 2024, 28, 3206–3218. [Google Scholar] [CrossRef] [PubMed]
- Kamal, M.S.; Dey, N.; Chowdhury, L.; Hasan, S.I.; Santosh, K.C. Explainable AI for Glaucoma Prediction Analysis to Understand Risk Factors in Treatment Planning. IEEE Trans. Instrum. Meas. 2022, 71, 3171613. [Google Scholar] [CrossRef]
- Hassan, M.M.; Alqahtani, S.A.; Alrakhami, M.S.; Elhendi, A.Z. Transparent and Accurate COVID-19 Diagnosis: Integrating Explainable AI with Advanced Deep Learning in CT Imaging. CMES-Comput. Model. Eng. Sci. 2024, 139. [Google Scholar] [CrossRef]
- Tasnim, N.; Al Mamun, S.; Shahidul Islam, M.; Kaiser, M.S.; Mahmud, M. Explainable Mortality Prediction Model for Congestive Heart Failure with Nature-Based Feature Selection Method. Appl. Sci. 2023, 13, 6138. [Google Scholar] [CrossRef]
- El-Rashidy, N.; Sedik, A.; Siam, A.I.; Ali, Z.H. An efficient edge/cloud medical system for rapid detection of level of consciousness in emergency medicine based on explainable machine learning models. Neural Comput. Appl. 2023, 35, 10695–10716. [Google Scholar] [CrossRef] [PubMed]
- Prifti, E.; Chevaleyre, Y.; Hanczar, B.; Belda, E.; Danchin, A.; Clement, K.; Zucker, J.D. Interpretable and accurate prediction models for metagenomics data. Gigascience 2020, 9, giaa010. [Google Scholar] [CrossRef]
- Miao, J.; Thongprayoon, C.; Suppadungsuk, S.; Krisanapan, P.; Radhakrishnan, Y.; Cheungpasitporn, W. Chain of Thought Utilization in Large Language Models and Application in Nephrology. Medicina 2024, 60, 148. [Google Scholar] [CrossRef]
- Kumar, A.; Manikandan, R.; Kose, U.; Gupta, D.; Satapathy, S.C. Doctor’s Dilemma: Evaluating an Explainable Subtractive Spatial Lightweight Convolutional Neural Network for Brain Tumor Diagnosis. ACM Trans. Multimed. Comput. Commun. Appl. 2021, 17, 1–26. [Google Scholar] [CrossRef]
- Vijayvargiya, A.; Singh, P.; Kumar, R.; Dey, N. Hardware Implementation for Lower Limb Surface EMG Measurement and Analysis Using Explainable AI for Activity Recognition. IEEE Trans. Instrum. Meas. 2022, 71, 1–9. [Google Scholar] [CrossRef]
- Pintelas, E.; Livieris, I.E.; Pintelas, P. Explainable Feature Extraction and Prediction Framework for 3D Image Recognition Applied to Pneumonia Detection. Electronics 2023, 12, 2663. [Google Scholar] [CrossRef]
- Wang, Z.; Samsten, I.; Kougia, V.; Papapetrou, P. Style-transfer counterfactual explanations: An application to mortality prevention of ICU patients. Artif. Intell. Med. 2023, 135, 102457. [Google Scholar] [CrossRef]
- Lugan, S.; Desbordes, P.; Brion, E.; Tormo, L.X.R.; Legay, A.; Macq, B. Secure Architectures Implementing Trusted Coalitions for Blockchained Distributed Learning (TCLearn). IEEE Access 2019, 7, 181789–181799. [Google Scholar] [CrossRef]
- Shukla, S.; Birla, L.; Gupta, A.K.; Gupta, P. Trustworthy Medical Image Segmentation with improved performance for in-distribution samples. Neural Netw. 2023, 166, 127–136. [Google Scholar] [CrossRef]
- Bassiouny, R.; Mohamed, A.; Umapathy, K.; Khan, N.; IEEE. An Interpretable Object Detection-Based Model for the Diagnosis of Neonatal Lung Diseases Using Ultrasound Images. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Mexico City, Mexico, 1–5 November 2021. [Google Scholar] [CrossRef]
- Jiang, R.; Chazot, P.; Pavese, N.; Crookes, D.; Bouridane, A.; Celebi, M.E. Private Facial Prediagnosis as an Edge Service for Parkinson’s DBS Treatment Valuation. IEEE J. Biomed. Health Inform. 2022, 26, 2703–2713. [Google Scholar] [CrossRef] [PubMed]
- Abdelfattah, S.; Baza, M.; Mahmoud, M.; Fouda, M.M.; Abualsaud, K.; Yaacoub, E.; Alsabaan, M.; Guizani, M. Lightweight Multi-Class Support Vector Machine-Based Medical Diagnosis System with Privacy Preservation. Sensors 2023, 23, 9033. [Google Scholar] [CrossRef] [PubMed]
- De Paolis Kaluza, M.C.; Jain, S.; Radivojac, P. An Approach to Identifying and Quantifying Bias in Biomedical Data. In Proceedings of the Pacific Symposium on Biocomputing 2023, Kohala Coast, HI, USA, 3–7 January 2023; Volume 28, pp. 311–322. [Google Scholar]
- Aboutalebi, H.; Pavlova, M.; Shafiee, M.J.; Florea, A.; Hryniowski, A.; Wong, A. COVID-Net Biochem: An explainability-driven framework to building machine learning models for predicting survival and kidney injury of COVID-19 patients from clinical and biochemistry data. Sci. Rep. 2023, 13, 17001. [Google Scholar] [CrossRef] [PubMed]
- Uzunova, H.; Ehrhardt, J.; Kepp, T.; Handels, H. Interpretable Explanations of Black Box Classifiers Applied on Medical Images by Meaningful Perturbations Using Variational Autoencoders. In Medical Imaging 2019: Image Processing; SPIE: Bellingham, WA, USA, 2019. [Google Scholar] [CrossRef]
- Lu, K.; Tong, Y.; Yu, S.; Lin, Y.; Yang, Y.; Xu, H.; Li, Y.; Yu, S. Building a trustworthy AI differential diagnosis application for Crohn’s disease and intestinal tuberculosis. BMC Med. Inform. Decis. Mak. 2023, 23, 160. [Google Scholar] [CrossRef]
- Chen, Y.; Aleman, D.M.; Purdie, T.G.; McIntosh, C. Understanding machine learning classifier decisions in automated radiotherapy quality assurance. Phys. Med. Biol. 2022, 67, 025001. [Google Scholar] [CrossRef]
- Araujo, T.; Aresta, G.; Mendonca, L.; Penas, S.; Maia, C.; Carneiro, A.; Maria Mendonca, A.; Campilho, A. DR|GRADUATE: Uncertainty-aware deep learning-based diabetic retinopathy grading in eye fundus images. Med. Image Anal. 2020, 63, 101715. [Google Scholar] [CrossRef]
- Malik, H.; Anees, T.; Naeem, A.; Naqvi, R.A.; Loh, W.K. Blockchain-Federated and Deep-Learning-Based Ensembling of Capsule Network with Incremental Extreme Learning Machines for Classification of COVID-19 Using CT Scans. Bioengineering 2023, 10, 203. [Google Scholar] [CrossRef] [PubMed]
- Zerka, F.; Urovi, V.; Vaidyanathan, A.; Barakat, S.; Leijenaar, R.T.H.; Walsh, S.; Gabrani-Juma, H.; Miraglio, B.; Woodruff, H.C.; Dumontier, M.; et al. Blockchain for Privacy Preserving and Trustworthy Distributed Machine Learning in Multicentric Medical Imaging (C-DistriM). IEEE Access 2020, 8, 183939–183951. [Google Scholar] [CrossRef]
- Saleem, H.; Shahid, A.R.; Raza, B. Visual interpretability in 3D brain tumor segmentation network. Comput. Biol. Med. 2021, 133, 104410. [Google Scholar] [CrossRef] [PubMed]
- El Houda, Z.A.; Hafid, A.S.; Khoukhi, L.; Brik, B. When Collaborative Federated Learning Meets Blockchain to Preserve Privacy in Healthcare. IEEE Trans. Netw. Sci. Eng. 2023, 10, 2455–2465. [Google Scholar] [CrossRef]
- Stenwig, E.; Salvi, G.; Rossi, P.S.; Skjaervold, N.K. Comparative analysis of explainable machine learning prediction models for hospital mortality. BMC Med. Res. Methodol. 2022, 22, 53. [Google Scholar] [CrossRef] [PubMed]
- Ogbomo-Harmitt, S.; Muffoletto, M.; Zeidan, A.; Qureshi, A.; King, A.P.; Aslanidi, O. Exploring interpretability in deep learning prediction of successful ablation therapy for atrial fibrillation. Front. Physiol. 2023, 14, 1054401. [Google Scholar] [CrossRef] [PubMed]
- Alzubaidi, L.; Salhi, A.; A Fadhel, M.; Bai, J.; Hollman, F.; Italia, K.; Pareyon, R.; Albahri, A.S.; Ouyang, C.; Santamaria, J.; et al. Trustworthy deep learning framework for the detection of abnormalities in X-ray shoulder images. PLoS ONE 2024, 19, e0299545. [Google Scholar] [CrossRef] [PubMed]
- El-Sappagh, S.; Alonso, J.M.; Islam, S.M.R.; Sultan, A.M.; Kwak, K.S. A multilayer multimodal detection and prediction model based on explainable artificial intelligence for Alzheimer’s disease. Sci. Rep. 2021, 11, 2660. [Google Scholar] [CrossRef] [PubMed]
- Gundersen, T.; Baeroe, K. The Future Ethics of Artificial Intelligence in Medicine: Making Sense of Collaborative Models. Sci. Eng. Ethics 2022, 28, 17. [Google Scholar] [CrossRef]
- Mukhopadhyay, S. InfoGram and admissible machine learning. Mach. Learn. 2022, 111, 205–242. [Google Scholar] [CrossRef]
- Alamro, H.; Marzouk, R.; Alruwais, N.; Negm, N.; Aljameel, S.S.; Khalid, M.; Hamza, M.A.; Alsaid, M.I. Modeling of Blockchain Assisted Intrusion Detection on IoT Healthcare System Using Ant Lion Optimizer with Hybrid Deep Learning. IEEE Access 2023, 11, 82199–82207. [Google Scholar] [CrossRef]
- Soni, P.; Pradhan, J.; Pal, A.K.; Islam, S.K.H. Cybersecurity Attack-Resilience Authentication Mechanism for Intelligent Healthcare System. IEEE Trans. Ind. Inform. 2023, 19, 830–840. [Google Scholar] [CrossRef]
- Cho, K.H.; Kim, E.S.; Kim, J.W.; Yun, C.H.; Jang, J.W.; Kasani, P.H.; Jo, H.S. Comparative effectiveness of explainable machine learning approaches for extrauterine growth restriction classification in preterm infants using longitudinal data. Front. Med. 2023, 10, 1166743. [Google Scholar] [CrossRef] [PubMed]
- Stolte, S.E.; Volle, K.; Indahlastari, A.; Albizu, A.; Woods, A.J.; Brink, K.; Hale, M.; Fang, R. DOMINO: Domain-aware loss for deep learning calibration. Softw. Impacts 2023, 15, 100478. [Google Scholar] [CrossRef]
- Albahri, A.S.; Duhaim, A.M.; Fadhel, M.A.; Alnoor, A.; Baqer, N.S.; Alzubaidi, L.; Albahri, O.S.; Alamoodi, A.H.; Bai, J.; Salhi, A.; et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: Assessment of quality, bias risk, and data fusion. Inf. Fusion 2023, 96, 156–191. [Google Scholar] [CrossRef]
Requirement | Description |
---|---|
Human Agency and Oversight | AI systems should empower human beings, allowing them to make informed decisions and fostering their fundamental rights, while maintaining appropriate human oversight and intervention capabilities. |
Technical Robustness and Safety | AI systems should be resilient and secure, with reliable performance and fallback plans in case of issues. This includes managing risks related to cybersecurity and ensuring accuracy, reliability, and reproducibility. |
Privacy and Data Governance | AI systems must ensure that personal data are protected, incorporating mechanisms for data protection, consent, and privacy preservation. Proper data governance practices must be in place to manage data quality and integrity. |
Transparency | AI systems should operate transparently, with traceable processes and understandable outcomes. This involves clear communication about the system’s capabilities, limitations, and decision-making processes. |
Diversity, Nondiscrimination, and Fairness | AI systems should promote inclusion and diversity, preventing bias and ensuring that they do not discriminate against any individuals or groups. They must be designed and tested to ensure fair outcomes. |
Environmental and Societal Well-Being | AI systems should consider their environmental impact, promoting sustainability and ensuring that they contribute positively to societal well-being and the common good. |
Accountability | There must be mechanisms to ensure accountability for AI systems and their outcomes, with clear guidelines for responsibility and mechanisms for action in cases of harm or negative impacts. |
Publication | HUM. | TEC. | PRI. | TRA. | DIV. | ACC. |
---|---|---|---|---|---|---|
Alves et al. [9] | N | Y | Y | N | N | N |
Moradi and Samwald [10] | N | Y | N | N | N | N |
Alves et al. [9] | N | Y | N | Y | N | Y |
Moradi and Samwald [10] | Y | Y | N | Y | N | Y |
Ma et al. [11] | Y | Y | N | Y | Y | N |
Khanna et al. [12] | Y | Y | N | Y | N | N |
Fidon et al. [13] | N | Y | N | N | Y | Y |
Nambiar et al. [14] | N | N | N | Y | N | N |
Rashid et al. [15] | Y | Y | Y | Y | Y | N |
Kumar et al. [16] | N | N | Y | Y | N | N |
Salahuddin et al. [17] | N | N | N | Y | N | N |
Zicari et al. [18] | Y | Y | Y | Y | Y | Y |
Bruckert et al. [19] | Y | N | N | Y | N | Y |
Ma et al. [20] | Y | N | N | Y | N | N |
Imboden et al. [21] | N | Y | N | N | N | N |
Karim et al. [22] | N | Y | N | N | N | N |
Mu et al. [23] | N | Y | Y | Y | N | N |
Kamal et al. [24] | Y | N | N | Y | N | Y |
Hassan et al. [25] | N | N | N | Y | N | N |
Tasnim et al. [26] | N | N | N | Y | N | N |
El-Rashidy et al. [27] | N | N | N | N | N | N |
Prifti et al. [28] | N | N | N | Y | N | N |
Miao et al. [29] | N | N | N | N | N | N |
Kumar et al. [30] | Y | N | N | Y | Y | N |
Vijayvargiya et al. [31] | N | N | N | N | N | N |
Pintelas et al. [32] | N | N | N | Y | N | N |
Wang et al. [33] | N | N | N | Y | N | N |
Lugan et al. [34] | N | Y | Y | N | N | N |
Shukla et al. [35] | N | Y | N | N | N | N |
Bassiouny et al. [36] | Y | N | N | Y | N | N |
Jiang et al. [37] | N | Y | Y | N | N | N |
Abdelfattah et al. [38] | N | N | Y | N | N | N |
De Paolis Kaluza et al. [39] | N | N | N | N | Y | N |
Aboutalebi et al. [40] | Y | N | N | Y | Y | N |
Uzunova et al. [41] | N | N | N | Y | N | N |
Lu et al. [42] | Y | Y | N | Y | Y | N |
Chen et al. [43] | N | N | N | Y | N | Y |
Araujo et al. [44] | Y | N | N | Y | N | N |
Malik et al. [45] | N | N | Y | N | N | N |
Zerka et al. [46] | N | N | Y | N | N | N |
Saleem et al. [47] | N | N | N | Y | N | N |
El Houda et al. [48] | N | Y | Y | N | N | N |
Stenwig et al. [49] | N | N | N | Y | N | N |
Ogbomo-Harmitt et al. [50] | N | N | N | Y | N | N |
Alzubaidi et al. [51] | N | Y | N | Y | N | N |
El-Sappagh et al. [52] | N | N | N | Y | N | N |
Gundersen and Baeroe [53] | N | N | N | N | Y | N |
Mukhopadhyay [54] | N | Y | N | Y | Y | N |
Alamro et al. [55] | N | Y | N | N | N | N |
Soni et al. [56] | N | Y | N | N | N | N |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mora-Cantallops, M.; García-Barriocanal, E.; Sicilia, M.-Á. Trustworthy AI Guidelines in Biomedical Decision-Making Applications: A Scoping Review. Big Data Cogn. Comput. 2024, 8, 73. https://doi.org/10.3390/bdcc8070073
Mora-Cantallops M, García-Barriocanal E, Sicilia M-Á. Trustworthy AI Guidelines in Biomedical Decision-Making Applications: A Scoping Review. Big Data and Cognitive Computing. 2024; 8(7):73. https://doi.org/10.3390/bdcc8070073
Chicago/Turabian StyleMora-Cantallops, Marçal, Elena García-Barriocanal, and Miguel-Ángel Sicilia. 2024. "Trustworthy AI Guidelines in Biomedical Decision-Making Applications: A Scoping Review" Big Data and Cognitive Computing 8, no. 7: 73. https://doi.org/10.3390/bdcc8070073
APA StyleMora-Cantallops, M., García-Barriocanal, E., & Sicilia, M. -Á. (2024). Trustworthy AI Guidelines in Biomedical Decision-Making Applications: A Scoping Review. Big Data and Cognitive Computing, 8(7), 73. https://doi.org/10.3390/bdcc8070073