按 Enter 到主內容區
:::

Enhancing transparency and efficiency in clinical AI applications while building public trust in the use of artificial intelligence in healthcare

Center for Responsible AI in Healthcare

Mission Statement

The Center for Responsible AI in Healthcare was established to enhance transparency and efficiency in clinical applications and to build public trust and acceptance of artificial (AI) technologies in the healthcare field.

With the rapid advancement of AI, its applications in healthcare have expanded from disease diagnosis, risk prediction, and medical image interpretation to include clinical documentation support, significantly reducing the time healthcare professionals spend on writing medical records.

However, the adoption of AI is not without risks. Ensuring autonomy, accountability, privacy, transparency, security, fairness, and sustainability has become a critical global concern. Both the World Health Organization (WHO) and Taiwan’s draft Fundamental Act of Artificial Intelligence emphasize that without proper guidance, AI could lead to ethical dilemmas and regulatory challenges.

To address these concerns, Dr. Chien-Chang Lee, MD, ScD, Director of the Department of Information Management at the Ministry of Health and Welfare (MOHW), launched the Center for Responsible AI in Healthcare. The Center is dedicated to developing a governance framework that reflects seven core ethical principles, ensuring that ethical and safety considerations evolve concurrently with innovations in smart healthcare technologies.

The seven major ethical risks of AI in healthcare under insufficient regulation

Figure 1. The seven major ethical risks of AI in healthcare under insufficient regulation

Core Strategies

1. Governance Measures Aligned with Security and Privacy Protection Principles

  • Privacy Protection

When AI applications involve the use of personal data, healthcare institutions must implement robust and comprehensive privacy protection mechanisms. Adherence to established frameworks such as the  U.S. Health Insurance Portability and Accountability Act (HIPAA) or the EU General Data Protection Regulation (GDPR) is strongly recommended. Identifiable personal data should be removed or securely encrypted to reduce the risks of re-identification and data leakage.

In addition, institutions should clearly define the data storage location and retention period, specify access permissions for external administrators, and establish a routine audit mechanism to ensure proper implementation. For AI applications that require data to be uploaded to external cloud systems for inference, additional safeguards should be used, including advanced protection protocols and a detailed breach response. These measures collectively help maintain high standards of data security and regulatory compliance, especially during inter-institutional or cross-site data transmission.

  • Information Security

When AI applications are hosted on external cloud platforms or managed by third parties, healthcare institutions must implement robust information security protocols to minimize the risk of system intrusions or data misuse.

Key recommendations include implementing multilayered cybersecurity defenses, such as firewalls and Intrusion Detection/Prevention Systems (IDS/IPS), to prevent unauthorized access; using network segmentation to limit access to sensitive systems and data; and conducting regular scans with automated vulnerability assessment tools. Additionally, the roles and access privileges of external administrators must be clearly defined.  Institutions should implement periodic audits to ensure that security policies are consistently enforced and continuously refined in daily operations.

Three core pillars of AI implementation governance - Information security and privacy meet international standards

Figure 2. Three core pillars of AI implementation governance - Information security and privacy meet international standards

 

2. Transparency and Explainability

  • Transparency Framework

Healthcare institutions deploying AI tools should ensure transparency by publicing disclosing essential information through dedicated online platforms. This includes the data sources used for training, model architecture, validation datasets, intended clinical use cases, and certification status from regulatory bodies such as the FDA or equivalent medical review authorities. These disclosures should align with a nine-principle transparency framework to promote accountability, trust, and informed decision-making among stakeholders.

  • Explainability Analysis

Explainability techniques are crucial in clinical settings, as they help clarify how AI models make predictions or decisions  to support physician interpretation and reinforce accountability. Tools like SHAP (SHapley Additive Explanations) and saliency maps offer insights into model behavior and decision logic, enhancing clinician confidence in AI outputs.

Three core pillars of AI implementation governance - Announcement transparency principle and explainability analysis

Figure 3. Three core pillars of AI implementation governance - Announcement transparency principle and explainability analysis

 

3. Monitoring AI Effectiveness Across the System Lifecycle

AI tools in clinical care require continuous monitoring and evaluation throughout their entire lifecycle,  which includes not only development and deployment but also real world use. This includes conducting routine performance audits (e.g., every 6 to 12 months) using independent datasets to ensure the model maintains its accuracy, reliability, and relevance in response  to changing medical data and clinical practices.

  • Effectiveness Monitoring

For example, an AI model designed to detect breast cancer may have been trained on a large set of imaging data. However, once deployed in a clinical setting, it must undergo regular evaluations d to ensure sustained diagnostic accuracy, reliability, and fairness. Ongoing monitoring should include bias detection and correction, as well as comparative analysis with expert radiologist interpretationsto validate clinical effectiveness

  • Retraining and Optimization

As new mammogram images are collected, the model must be periodically retrained to maintain high performance. Feedback from clinicians should be gathered to identify potential issues, which then inform system updates and model improvements.

Three core pillars of AI implementation governance-AI Lifecycle Management

Figure 4. Three core pillars of AI implementation governance-AI Lifecycle Management

 

Future Outlook

The Department of Information Management at MoHW is committed to promoting an innovative and forward-looking regulatory framework for smart healthcare. In 2024, it provided funding to 10 hospitals across northern, central, and southern Taiwan, covering medical centers, regional hospitals, and district hospitals to promote equitable access to AI technologies across all levels of care.

This six-month pilot initiative included the establishment of hospital-level task forces, regular on-site guidance, and the development of transparency mechanisms, explainability tools, and lifecycle monitoring systems. To foster public trust, each participating hospital was required to disclose its AI related policies, designated contact points, and successful use cases, allowing the public to understand how AI is applied and its impact on care quality.

Looking ahead, the Responsible AI Implementation Center will continue to foster international exchange of best practices and successful implementations, positioning Taiwan as a global leader  in the safe, transparent, and ethically governed application of AI in medicine.

 

center for Responsible AI in Healthcare

Figure 5. Center for Responsible AI in Healthcare

Project Team and Advisors

Project Initiator

Chien-Chang Lee, M.D., Sc.D.

Chien-Chang Lee, M.D., Sc.D. 

Chief Information Officer (CIO) for the Taiwan Ministry of Health and Welfare (MOHW) 
Professor of Emergency Medicine 
Deputy Director of the Center of Intelligent Healthcare at National Taiwan University (NTU) Hospital

Acknowledgements

We sincerely acknowledge Ming-Shiang Wu, MD, PhD, Convener of the Center for Responsible AI in Healthcare and Superintendent of National Taiwan University Hospital (Taipei, Taiwan); Kenneth Mandl, MD, MPH, Professor of Pediatrics and Biomedical Informatics at Harvard University and Director of the Computational Health Informatics Program at Boston Children’s Hospital (Boston, MA, USA); and David Rhew, MD, Global Chief Medical Officer (CMO) and Vice President of Healthcare at Microsoft (New York, NY, USA), as distinguished international experts for their invaluable professional guidance and contributions to our initiative in advancing responsible AI in healthcare.

Ming-Shiang Wu, MD, PhD Kenneth Mandl, MD, MPH David Rhew, MD

Ming-Shiang Wu, MD, PhD

Kenneth Mandl, MD, MPH David Rhew, MD

Convener of the Center for Responsible AI in Healthcare
Superintendent of National Taiwan University Hospital

Donald A.B. Lindberg Professor of Pediatrics and Biomedical Informatics at Harvard University
Director of the Computational Health Informatics Program at Boston Children’s Hospital

Global Chief Medical Officer (CMO) and VP of Healthcare for Microsoft

 

International Partner Matching Platform

International Partner Matching Platform

We support international organizations and innovators in identifying and connecting with the most appropriate partners in Taiwan, including the officially designated Center for Responsible AI in Healthcare.To learn more, please click here.

international partner matching platform

  • 發布日期:
  • 更新日期: