Back to Insights
AI Compliance & RegulationGuidePractitioner

AI Regulations in Singapore: IMDA Guidelines and Compliance Requirements

October 21, 202511 min readMichael Lansdowne Hauge
For:Compliance OfficersLegal CounselIT DirectorsBusiness Leaders

Complete guide to Singapore AI governance. Covers IMDA Model Framework, PDPA requirements for AI, MAS guidelines, and practical implementation.

Muslim Man Lawyer Formal - ai compliance & regulation insights

Key Takeaways

  • 1.IMDA Model AI Governance Framework provides practical guidance for responsible AI deployment
  • 2.Singapore takes a principles-based approach rather than prescriptive regulation
  • 3.AI Verify is Singapore's testing toolkit for demonstrating responsible AI practices
  • 4.Financial services, healthcare, and government sectors face additional AI requirements
  • 5.Early adoption of governance frameworks provides competitive advantage in regulated sectors

AI Regulations in Singapore: IMDA Guidelines and Compliance Requirements

Singapore has positioned itself as a leader in AI governance through practical, business-friendly frameworks. While currently voluntary, these frameworks set expectations that organizations should meet—and are increasingly referenced in contracts, audits, and regulatory discussions.

Executive Summary

  • Singapore favors principles over prescriptions. The approach is practical and flexible, not rigidly prescriptive.
  • The Model AI Governance Framework is foundational. It sets expectations for responsible AI even without legal mandate.
  • PDPA applies to AI processing personal data. This is not optional—it's law.
  • Sector-specific rules add layers. Financial services (MAS) and other sectors have additional requirements.
  • Voluntary today may be mandatory tomorrow. Building governance now prepares for future requirements.
  • Regional leadership means global relevance. Singapore's approach influences ASEAN and beyond.
  • Practical implementation is expected. Frameworks emphasize actionable governance, not just documentation.
  • Accountability is the core principle. Organizations must be able to answer for their AI systems.

Why This Matters Now

Singapore's AI governance approach is maturing:

  • Model AI Governance Framework widely adopted by leading organizations
  • PDPC guidance on AI and personal data provides specific expectations
  • MAS has detailed AI requirements for financial institutions
  • Increasing customer and investor due diligence on AI governance
  • ASEAN coordination using Singapore frameworks as reference

Organizations operating in Singapore—or serving Singapore customers—should understand and implement these expectations.


Singapore's AI Governance Framework

Model AI Governance Framework

Published by IMDA and PDPC, now in its second edition, this framework establishes four key principles:

1. Internal Governance Structures and Measures

RequirementWhat It MeansImplementation
Clear roles and responsibilitiesSomeone is accountable for AIDesignate AI governance owner
Board and management oversightLeadership understands AI risksRegular AI reporting to leadership
Risk management integrationAI risks in enterprise riskInclude AI in risk framework
Operations managementAI lifecycle managedGovernance through development and deployment

2. Determining AI Decision-Making Model

The framework distinguishes:

  • Human-in-the-loop: Human makes all decisions with AI assistance
  • Human-over-the-loop: Human oversees AI decisions with intervention capability
  • Human-out-of-the-loop: AI operates autonomously (highest governance bar)

Select the appropriate model based on risk and impact.

3. Operations Management

AreaRequirements
Data managementQuality, accuracy, and relevance of training data
Model developmentRobust development practices
Model deploymentTesting, validation, and monitoring
Performance monitoringOngoing accuracy and effectiveness tracking

4. Stakeholder Interaction and Communication

StakeholderExpectation
UsersInformed about AI use; can seek clarification
Affected partiesRecourse available for adverse decisions
RegulatorsTransparency about AI governance
PublicOrganizational stance on AI ethics clear

PDPA Requirements for AI

Singapore's Personal Data Protection Act applies when AI processes personal data:

Key PDPA Principles Applied to AI

PDPA RequirementAI Application
ConsentObtain consent for AI processing of personal data
Purpose limitationUse data only for consented purposes
NotificationInform individuals about AI processing
Access and correctionEnable access to and correction of AI-processed data
AccuracyEnsure AI uses and produces accurate data
ProtectionSecure personal data in AI systems
Retention limitationDon't retain AI data longer than necessary
Transfer limitationCross-border AI processing compliance

PDPC Advisory Guidelines on AI

The PDPC has issued specific guidance:

Accountability for AI:

  • Organizations responsible for AI outcomes
  • Must demonstrate compliance ability
  • Cannot outsource accountability to vendors

Explainability:

  • Individuals should understand AI decisions affecting them
  • Level of explanation proportionate to impact
  • Technical accuracy not required; meaningful explanation is

Fairness:

  • AI should not unfairly discriminate
  • Testing for bias expected
  • Remediation when bias discovered

Sector-Specific: Financial Services (MAS)

The Monetary Authority of Singapore has detailed expectations:

FEAT Principles

Fairness:

  • AI decisions should be fair and non-discriminatory
  • Regular bias testing required
  • Remediation processes established

Ethics:

  • AI use aligned with ethical principles
  • Customer interests protected
  • Transparent about AI use

Accountability:

  • Clear ownership of AI systems
  • Governance structures in place
  • Ability to explain decisions

Transparency:

  • Customers informed about AI use
  • Regulators can understand AI operation
  • Documentation maintained

MAS Technology Risk Management Guidelines

  • AI systems subject to TRM requirements
  • Model risk management applies
  • Third-party AI vendor assessment required

Implementation Roadmap

Phase 1: Foundation (Weeks 1-4)

Governance structure:

  • Designate AI governance owner
  • Establish oversight mechanism
  • Define escalation procedures
  • Create AI governance policy

Inventory and assessment:

  • Catalog all AI systems
  • Map personal data in AI
  • Classify risk levels
  • Document purposes

Phase 2: Core Compliance (Weeks 5-8)

PDPA compliance:

  • Review consent mechanisms for AI
  • Update privacy notices
  • Implement access/correction for AI data
  • Establish retention policies

Human oversight:

  • Define oversight model per system
  • Implement intervention capabilities
  • Train oversight staff
  • Document procedures

Phase 3: Enhanced Governance (Weeks 9-12)

Documentation:

  • Complete technical documentation
  • Document governance decisions
  • Prepare audit-ready materials
  • Establish evidence repository

Monitoring:

  • Implement performance monitoring
  • Establish bias testing
  • Create reporting mechanisms
  • Schedule regular reviews

Common Failure Modes

1. Treating the framework as optional. While legally voluntary, it sets industry expectations. Customers and partners increasingly require it.

2. Documentation without implementation. Policies without practices don't satisfy governance requirements.

3. Ignoring PDPA for AI. PDPA is law, not guidance. AI processing personal data must comply.

4. Sector-blindness. Financial services and other regulated sectors have additional obligations beyond general frameworks.

5. One-time compliance. Governance requires ongoing maintenance, not point-in-time implementation.


Singapore AI Compliance Checklist

SINGAPORE AI COMPLIANCE CHECKLIST

Governance Structure
[ ] AI governance owner designated
[ ] Board/management oversight established
[ ] AI risk in enterprise risk framework
[ ] AI governance policy documented

Model AI Governance Framework
[ ] AI systems inventoried
[ ] Decision-making model selected per system
[ ] Human oversight appropriate to risk
[ ] Stakeholder communication approach defined

PDPA Compliance
[ ] Personal data in AI systems mapped
[ ] Consent obtained for AI processing
[ ] Privacy notices updated for AI
[ ] Access and correction processes include AI
[ ] Data protection measures implemented
[ ] Retention policies applied
[ ] Cross-border compliance verified

Sector-Specific (if applicable)
[ ] MAS FEAT principles addressed (financial services)
[ ] Industry-specific requirements identified
[ ] Sector regulator guidance reviewed

Documentation
[ ] Technical documentation complete
[ ] Governance decisions documented
[ ] Testing results maintained
[ ] Audit trail established

Monitoring
[ ] Performance monitoring active
[ ] Bias testing conducted
[ ] Regular review scheduled
[ ] Improvement process defined

Metrics to Track

MetricTargetFrequency
AI systems with governance assessment100%Quarterly
PDPA compliance for AI personal data100%Ongoing
Staff training completion>95%Annually
Governance review completion100%Annually
Bias testing for high-risk AI100%Semi-annually

FAQ

Q: Is the Model AI Governance Framework legally required? A: Not directly, but it sets industry expectations. PDPA compliance is legally required, and the framework helps demonstrate it.

Q: Does PDPA apply to AI that doesn't use personal data? A: PDPA applies only when personal data is processed. AI using only non-personal data isn't subject to PDPA but should still follow governance principles.

Q: What are the penalties for non-compliance? A: PDPA violations can result in penalties up to S$1 million. Sector-specific violations may have additional consequences.

Q: How does this compare to EU AI Act? A: Singapore's approach is less prescriptive. The EU AI Act has binding requirements with specific risk categories. Singapore emphasizes principles with organizational flexibility.

Q: Should we align with EU AI Act as well? A: If you serve EU customers or your AI affects EU residents, yes. Many organizations align with both frameworks.


Next Steps

Singapore compliance is part of regional governance:


Book an AI Readiness Audit

Need help with Singapore AI governance? Our AI Readiness Audit includes alignment with IMDA framework and PDPA requirements.

Book an AI Readiness Audit →


Disclaimer

This article provides general guidance on Singapore AI regulations. It does not constitute legal advice. Organizations should consult qualified Singapore legal counsel for specific compliance requirements.


References

  1. Singapore IMDA & PDPC. Model AI Governance Framework, Second Edition.
  2. Singapore PDPC. Advisory Guidelines on Use of Personal Data in AI Recommendation and Decision Systems.
  3. Monetary Authority of Singapore. Principles to Promote Fairness, Ethics, Accountability and Transparency (FEAT).
  4. Singapore PDPC. Guide to Data Protection by Design.
  5. Singapore Personal Data Protection Act 2012.

Frequently Asked Questions

Singapore takes a principles-based approach through the IMDA Model AI Governance Framework rather than prescriptive rules. It emphasizes practical guidance, voluntary adoption, and sector-specific requirements for regulated industries.

AI Verify is Singapore's testing framework and toolkit that allows organizations to demonstrate responsible AI practices. It provides standardized tests for fairness, explainability, and robustness of AI systems.

Financial services (MAS guidelines), healthcare, and government sectors face additional AI requirements including model risk management, explainability standards, and enhanced documentation.

References

  1. Singapore IMDA & PDPC. Model AI Governance Framework, Second Edition.. Singapore IMDA & PDPC Model AI Governance Framework Second Edition
  2. Singapore PDPC. Advisory Guidelines on Use of Personal Data in AI Recommendation and Decision Systems..
  3. Monetary Authority of Singapore. Principles to Promote Fairness, Ethics, Accountability and Transparency (FEAT)..
  4. Singapore PDPC. Guide to Data Protection by Design.. Singapore PDPC Guide to Data Protection by Design
  5. Singapore Personal Data Protection Act 2012.. Singapore Personal Data Protection Act (2012)
Michael Lansdowne Hauge

Founder & Managing Partner

Founder & Managing Partner at Pertama Partners. Founder of Pertama Group.

singapore ai regulationsimda guidelinessingapore ai compliance

Explore Further

Key terms:AI Regulation

Ready to Apply These Insights to Your Organization?

Book a complimentary AI Readiness Audit to identify opportunities specific to your context.

Book an AI Readiness Audit