Table of Contents
Context
- India has developed a framework called Strategy for Artificial Intelligence in Healthcare for India (SAHI) for using AI in healthcare.
More about the Framework
- It is built by Union Health Ministry and the National Health Authority (NHA)
- It builds on the digital public infrastructure created under the Ayushman Bharat Digital Mission (ABDM)
- India becomes one of the first Southeast Asian countries with a comprehensive AI-health strategy.
- Model for Global South: India’s digital public infrastructure-based AI governance offers a scalable template for developing nations.
Key Features of the Guidelines (SAHI)
- Data Collection and Curation: High-quality, diverse and ethically sourced data must be collected carefully to prevent bias and ensure reliable training of AI systems.
- Model Training and Validation: AI models must be scientifically validated through rigorous testing to ensure accuracy, fairness and safety before deployment in real healthcare settings.
- Deployment and Integration: AI systems must integrate smoothly into hospital workflows without disrupting existing medical practices or overburdening healthcare professionals.
- Continuous Monitoring: AI tools must be regularly monitored after deployment to detect performance decline, bias or unexpected outcomes in real-world use.
- Decommissioning if Needed: If an AI system becomes unsafe, outdated or ineffective, there must be mechanisms to withdraw or replace it responsibly.
- Privacy-by-Design & Consent Architecture: It follows ABDM principles, ensuring patient consent, minimal data sharing, secure APIs, and complete transparency in how data is used.
- Federated Data Model: Health data remains stored at hospitals and labs, reducing risks of centralised breaches while allowing controlled data exchange when consent is given.
- Patient Consent Mechanism: Patients must explicitly approve sharing of their health data, and they can revoke consent at any time, strengthening data ownership rights.
- Standardised Evaluation Criteria: AI tools must meet defined technical and clinical standards before being approved for public healthcare procurement.
- Outcome-Based Purchasing: Payment models may link AI performance to measurable health outcomes rather than upfront technology costs
- Bias Assessment: AI must be tested across genders, rural populations and marginalised groups to avoid discriminatory outcomes.
- Periodic Re-certification: AI systems should undergo periodic re-evaluation to ensure safety as they evolve with new data.
- Benchmarking Open Data Platform for Health AI (BODH): BODH provides curated datasets to test and compare AI models fairly while protecting sensitive health data through secure access mechanisms.
| BODH (Benchmarking Open Data Platform for Health AI) |
| It is a structured platform to test and validate AI health solutions before large-scale implementation.
Key Features ● Evaluates performance and reliability of AI tools ● Ensures real-world readiness of solutions ● Developed through collaboration between government and academia Significance ● Prevents premature or unsafe AI deployment ● Encourages evidence-based innovation ● Strengthens India’s global competitiveness in health AI |
Why This Framework Was Needed
- Fragmented and Sensitive Health Data: India’s health data is scattered and sensitive. Without proper governance, AI deployment could worsen inequality or compromise privacy.
- Risk of Algorithmic Bias: AI trained on limited datasets may underperform for rural, female or marginalised populations, leading to unequal healthcare delivery.
- Safety & Clinical Accountability:AI systems directly influence diagnosis and treatment. Errors may cause serious harm, requiring strict validation and oversight mechanisms.
- Regulatory Vacuum: Existing medical regulations do not fully cover adaptive AI systems. The framework fills this policy gap.
- Procurement Uncertainty:Hospitals lacked clear mechanisms to evaluate and purchase AI tools, slowing responsible adoption.
Major Concerns and Challenges
- Data Privacy & Security Risks:Health data breaches can damage public trust and expose individuals to discrimination, making cybersecurity safeguards essential.
- Model Drift: AI systems may change performance over time as new data is added, requiring regular monitoring and recalibration.
- Over-Reliance on Technology: Doctors must remain decision-makers, ensuring AI assists rather than replaces human clinical judgment.
- Digital Divide: Rural hospitals may lack infrastructure and trained staff, limiting equitable AI deployment.
- Ethical and Legal Liability:Clear responsibility must be defined when AI-related errors occur in medical practice.
Way Forward
- Strengthen Regulatory Oversight:India may establish a dedicated AI-health regulatory cell to monitor compliance and enforce periodic safety reviews.
- Expand Federated Learning:Federated learning can allow AI training without moving raw data, improving privacy protection.
- Promote Public-Private Collaboration:Partnerships between government, academia and startups can accelerate innovation while ensuring accountability.
- Bridge the Digital Divide:Investment in rural digital infrastructure and low-bandwidth AI tools will ensure equitable healthcare access.
- Continuous Dataset Updating:BODH datasets must be regularly updated and diversified to prevent stagnation and bias.
- Clear Legal Framework:Clear laws on liability, transparency and grievance redressal will strengthen trust in AI systems.
| Global Best Practices |
| ● WHO Ethical Principles:WHO recommends transparency, accountability, inclusiveness and sustainability in AI use, which India’s framework incorporates.
● European Union – AI Act:The EU classifies health AI as high-risk and mandates strict testing, human oversight and post-market monitoring. ● United States – FDA AI Framework:The US FDA requires change control plans and continuous performance monitoring for AI-based medical tools. ● UK – NHS AI Lab Model: The NHS AI Lab promotes sandbox testing, evidence generation and structured procurement guidelines. ● Singapore’s AI Governance Framework: Singapore emphasises explainability, risk-based classification and strong cross-sector regulatory standards. |
|
Read More Notes |
|
| Environment Notes | Art and Culture Notes |
| Science and Tech | History Notes |
| Geography Notes | Indian Polity Notes |
| General Knowledge | International Relation |
|
Explore StudyIQ Courses |
|

Effectiveness of Tobacco Taxation in Ind...
Private Sector Participation in Defence ...
India–France Strategic Cooperation in ...








