← Back to Assessment Tools

AI Risk Assessment Tool

Comprehensive clinical AI system risk evaluation framework based on FDA Software as a Medical Device (SaMD) guidelines, ISO 14971:2019 risk management standards, and FMEA methodology.

⚠️

Critical Disclaimer: Educational Use Only

This risk assessment tool is for educational, research, and preliminary evaluation purposes only. It does NOT constitute: formal regulatory submission support, legal compliance certification, professional safety assessment, or official risk management documentation.Results must be reviewed by qualified professionals (regulatory affairs specialists, clinical safety officers, legal counsel) before making deployment decisions or regulatory submissions. This tool provides guidance based on established frameworks but cannot replace comprehensive risk management processes required by FDA, ISO 14971, or other regulatory bodies. Use of this tool does not ensure regulatory compliance or clinical safety. High-risk medical AI systems require expert multidisciplinary review.

Assessment Methodology

Framework Basis

This assessment integrates multiple evidence-based frameworks:

  • FDA SaMD: Risk categorization based on intended use, clinical significance, and state of healthcare situation
  • ISO 14971:2019: Medical device risk management including severity classification and risk control measures
  • FMEA: Failure Mode and Effects Analysis principles for systematic hazard identification
  • WHO AI Ethics: Principles for responsible AI in healthcare including fairness, transparency, and accountability

Scoring System

The assessment uses weighted scoring across 6 dimensions:

  • Clinical Impact (30%): Intended use, potential harm, clinical criticality, human oversight
  • Data Quality (25%): Representativeness, bias testing, completeness, labeling quality
  • Clinical Validation (25%): Validation rigor, comparative effectiveness, appropriate metrics, temporal testing
  • Transparency (10%): Explainability, documentation of limitations and failure modes
  • Regulatory (5%): Regulatory status, data protection compliance
  • Post-Market (5%): Surveillance, adverse event reporting, update management

Note: Higher scores indicate better risk management and safety controls. The weighted approach prioritizes clinical impact and validation evidence as most critical to patient safety.

Interpretation Guidelines

  • 80-100% (Low Risk): Comprehensive risk management, ready for deployment with appropriate oversight
  • 60-79% (Moderate Risk): Acceptable with identified gaps requiring mitigation before or shortly after deployment
  • 40-59% (High Risk): Significant deficiencies requiring substantial additional work before deployment
  • 0-39% (Critical Risk): Not suitable for clinical use; major development and validation needed

Each question includes methodology notes explaining the scientific basis for the assessment dimension.

Questions answered: 0 / 18

Clinical Impact & Purpose (Weight: 30%)

1. What is the primary intended use of the AI system?

Clinical Impact & Purpose (Weight: 30%)

2. What is the potential severity of harm if the AI system fails or provides incorrect output?

Clinical Impact & Purpose (Weight: 30%)

3. What is the clinical state or situation of concern?

Clinical Impact & Purpose (Weight: 30%)

4. How does the AI system integrate into the clinical workflow?

Data Quality & Representativeness (Weight: 25%)

5. How representative is the training dataset of the intended patient population?

Data Quality & Representativeness (Weight: 25%)

6. Has the AI model been explicitly tested for fairness and bias across demographic subgroups?

Data Quality & Representativeness (Weight: 25%)

7. What is the quality and completeness of the training data?

Data Quality & Representativeness (Weight: 25%)

8. How was data labeling/ground truth established?

Clinical Validation & Performance (Weight: 25%)

9. What level of clinical validation has been completed?

Clinical Validation & Performance (Weight: 25%)

10. How does AI performance compare to the current standard of care?

Clinical Validation & Performance (Weight: 25%)

11. What metrics were used to evaluate performance, and are they clinically meaningful?

Clinical Validation & Performance (Weight: 25%)

12. Have temporal validation and/or performance monitoring over time been conducted?

Transparency & Explainability (Weight: 10%)

13. How interpretable and explainable is the AI model to clinicians?

Transparency & Explainability (Weight: 10%)

14. Is there comprehensive documentation of model limitations, failure modes, and contraindications?

Regulatory & Compliance (Weight: 5%)

15. What is the regulatory status and pathway for this AI system?

Regulatory & Compliance (Weight: 5%)

16. Does the system comply with data protection regulations (HIPAA, GDPR)?

Post-Market Surveillance (Weight: 5%)

17. Is there a system for post-market surveillance and adverse event reporting?

Post-Market Surveillance (Weight: 5%)

18. Is there a plan for model updates, retraining, and change management?

Please ensure all 18 questions are answered. Your responses will be used to calculate a weighted risk score and generate tailored recommendations.