Back to Insights
AI Compliance & RegulationGuide

AI Regulations in 2026: What Businesses Need to Know

October 20, 202511 min readMichael Lansdowne Hauge
Updated March 15, 2026
For:Legal/ComplianceCTO/CIOCISOConsultantCHROBoard MemberIT ManagerCEO/FounderHead of Operations

Comprehensive overview of AI regulatory landscape in 2026. Covers EU AI Act, ASEAN frameworks, sector-specific rules, and what to expect next.

Summarize and fact-check this article with:
Muslim Man Lawyer Formal - ai compliance & regulation insights

Key Takeaways

  • 1.EU AI Act is now in force with phased implementation through 2027
  • 2.Southeast Asian countries are developing AI governance frameworks aligned with regional standards
  • 3.Risk-based classification determines which AI systems face the strictest requirements
  • 4.Organizations using AI across borders must navigate multiple overlapping regulatory regimes
  • 5.Proactive compliance positioning provides competitive advantage as regulations mature

AI Regulations in 2026: What Businesses Need to Know

The AI regulatory landscape is shifting from voluntary guidelines to binding requirements. Organizations that wait for clarity before acting will find themselves playing catch-up with compliance obligations. This guide maps the current regulatory terrain and what's coming next.

Executive Summary

  • AI regulation is accelerating globally. What was guidance in 2024 is becoming law in 2026.
  • The EU AI Act is setting global standards. Even non-EU companies are affected through supply chains and customer requirements.
  • ASEAN is developing its own approach. Singapore, Malaysia, and Thailand are moving from voluntary frameworks toward clearer obligations.
  • Sector-specific regulations add layers. Financial services, healthcare, and other industries face additional AI requirements.
  • Cross-border operations complicate compliance. Organizations operating across jurisdictions must navigate multiple frameworks.
  • Enforcement is beginning. Early regulatory actions signal increased scrutiny ahead.
  • Preparation now reduces future pain. Organizations with mature governance will adapt more easily.
  • Risk-based approaches dominate. Higher-risk AI applications face stricter requirements.

Why This Matters Now

Several factors make 2026 a pivotal year for AI regulation:

EU AI Act implementation. The Act's requirements are phasing in, affecting global companies.

ASEAN alignment efforts. Regional coordination is producing more consistent expectations across Southeast Asia.

Enforcement precedents. Early regulatory actions establish how requirements will be interpreted.

Board and investor attention. Stakeholders increasingly ask about AI governance and regulatory readiness.

Customer due diligence. Enterprise customers require AI governance evidence from their vendors.


Global Regulatory Landscape

Regulatory Comparison by Jurisdiction

JurisdictionPrimary FrameworkApproachCurrent StatusKey Focus
European UnionEU AI ActRisk-based, bindingPhased implementationHigh-risk AI, prohibited practices
United StatesSector-specific, state lawsFragmentedEvolvingIndustry-specific, state variation
SingaporeIMDA Model FrameworkPrinciples-based, voluntary→mandatoryActive developmentGovernance, accountability
MalaysiaMDEC GuidelinesVoluntary guidanceDevelopingNational AI roadmap alignment
ThailandDEPA FrameworkEmergingEarly stageDigital economy integration
ChinaMultiple regulationsPrescriptiveActive enforcementContent, algorithms, deepfakes

EU AI Act: The Global Standard-Setter

Even if your organization isn't EU-based, the EU AI Act matters:

Risk Categories

Unacceptable Risk (Prohibited):

  • Social scoring systems
  • Real-time biometric identification (most cases)
  • Manipulation techniques exploiting vulnerabilities
  • Emotion recognition in workplaces/education

High Risk (Strict Requirements):

  • Employment decisions (recruitment, evaluation)
  • Education assessment
  • Credit scoring
  • Law enforcement applications
  • Infrastructure management

Limited Risk (Transparency Requirements):

  • Chatbots and AI assistants
  • Emotion detection systems
  • Content generation

Minimal Risk (No Specific Requirements):

  • Most business AI applications

Key Obligations for High-Risk AI

  • Risk management system
  • Data governance requirements
  • Technical documentation
  • Record-keeping
  • Transparency and information provision
  • Human oversight
  • Accuracy, robustness, cybersecurity

Extraterritorial Effect

The EU AI Act applies to:

  • AI systems placed on the EU market
  • AI systems whose outputs are used in the EU
  • Providers and deployers in the EU

This means many ASEAN companies serving European customers will need compliance.


ASEAN Regulatory Developments

Singapore

Current Framework:

  • IMDA Model AI Governance Framework (voluntary but influential)
  • PDPC guidance on AI and personal data
  • MAS guidelines for financial services AI

Direction:

  • Moving toward clearer accountability requirements
  • Increasing emphasis on AI transparency
  • Sector-specific obligations expanding

See for detailed Singapore guidance.

Malaysia

Current Framework:

  • MDEC AI guidelines
  • PDPA 2010 implications for AI
  • National AI Roadmap context

Direction:

  • Framework development ongoing
  • Alignment with regional approaches
  • Growing enforcement of data protection in AI contexts

See for detailed Malaysia guidance.

Thailand

Current Framework:

  • DEPA AI guidelines
  • Thailand PDPA (B.E. 2562) implications
  • National AI Strategy context

Direction:

  • Active framework development
  • Focus on responsible AI adoption
  • Building regulatory capacity

See for detailed Thailand guidance.

ASEAN Coordination

The ASEAN region is working toward:

  • Common AI governance principles
  • Interoperability between national frameworks
  • Shared approaches to cross-border AI services

Sector-Specific Regulations

Financial Services

JurisdictionRegulatorFocus Areas
SingaporeMASModel risk management, explainability, fairness
MalaysiaBNMRisk management, consumer protection
ThailandBOTDigital lending, algorithmic trading
GlobalBasel CommitteeAI in banking risk management

Healthcare

  • Patient data protection heightened for AI
  • Clinical AI requiring approval processes
  • Diagnostic AI facing device regulations

Education

  • Student data protection for AI tools
  • Fairness in AI-assisted assessment
  • Transparency to parents and students

Cross-Border Considerations

Organizations operating across jurisdictions face complexity:

Data transfer requirements:

  • AI training data may face transfer restrictions
  • Inference data flows need compliance
  • Model hosting location matters

Jurisdictional triggers:

  • Where is the AI developed?
  • Where is it deployed?
  • Where are users located?
  • Where are effects felt?

Practical approach:

  • Map AI systems to jurisdictions
  • Identify highest common denominator
  • Build flexible compliance frameworks

Early enforcement actions signal regulatory priorities:

Areas of focus:

  • Data protection violations in AI
  • Transparency failures
  • Discriminatory AI outcomes
  • Consumer harm

Penalty ranges:

  • EU AI Act: Up to €35M or 7% of global turnover
  • PDPA (Singapore): Up to S$1M per breach
  • PDPA (Malaysia): Up to RM500,000 and imprisonment

What to Expect Next

2026-2027 Outlook

Likely developments:

  • Full EU AI Act implementation
  • ASEAN framework solidification
  • Increased enforcement across regions
  • Industry-specific requirements expanding
  • International coordination efforts

Emerging areas:

  • Foundation model regulation
  • AI-generated content labeling
  • Environmental impact of AI
  • AI in critical infrastructure

Common Failure Modes

1. Waiting for final regulations. By the time rules are final, compliance timelines are short.

2. Ignoring extraterritorial reach. Serving international customers triggers foreign requirements.

3. Treating AI governance as optional. Voluntary frameworks are becoming mandatory expectations.

4. Sector-specific blindness. Industry regulations layer on top of general AI requirements.

5. Compliance as checkbox. Regulations require genuine governance, not just documentation.


AI Regulatory Readiness Checklist

AI REGULATORY READINESS CHECKLIST

Awareness
[ ] Key regulations identified for operating jurisdictions
[ ] Extraterritorial requirements understood
[ ] Sector-specific regulations mapped
[ ] Regulatory change monitoring in place

Assessment
[ ] AI systems inventoried
[ ] Risk classification applied (especially for EU AI Act)
[ ] Data flow mapping for AI completed
[ ] Gap analysis against requirements conducted

Governance
[ ] AI governance framework established
[ ] Accountability roles defined
[ ] Documentation practices implemented
[ ] Human oversight mechanisms in place

Technical
[ ] Risk management for AI systems
[ ] Transparency mechanisms implemented
[ ] Bias testing and monitoring
[ ] Security controls appropriate to risk

Operational
[ ] Training programs for AI compliance
[ ] Incident response includes AI
[ ] Vendor management addresses AI requirements
[ ] Regular review and update process

Metrics to Track

MetricTargetFrequency
AI systems with compliance assessment100%Quarterly
High-risk AI with full documentation100%Ongoing
Regulatory change reviewCompleteMonthly
Staff training completion>95%Annually
Compliance gaps openZero criticalMonthly

FAQ

Q: Does the EU AI Act apply to companies in ASEAN? A: If you serve EU customers or your AI affects EU residents, yes. The Act has extraterritorial reach.

Q: Are ASEAN AI regulations mandatory? A: Currently mostly voluntary, but moving toward mandatory. Data protection laws (PDPA) already apply to AI.

Q: What's the penalty for non-compliance? A: Varies by jurisdiction. EU AI Act penalties can reach €35M or 7% of global turnover. PDPA penalties in ASEAN range from hundreds of thousands to millions in local currency.

Q: How do I know which regulations apply? A: Map your AI systems by: where developed, where deployed, who uses them, where effects occur. Each may trigger different requirements.

Q: Should we wait for regulations to finalize? A: No. Build governance now. It's easier to adjust a framework than build one under deadline pressure.


Next Steps

Start preparing with jurisdiction-specific guidance:

  • [AI Compliance Checklist: Preparing for Regulatory Requirements]
  • [AI Regulations in Singapore: IMDA Guidelines and Compliance Requirements]
  • [AI Regulations in Malaysia: Current Framework and Future Directions]

Disclaimer

This article provides general information on AI regulatory developments. It does not constitute legal advice. Regulations are evolving rapidly. Organizations should consult qualified legal counsel for specific compliance requirements.


Building Regulatory Resilience: Preparing for Unknown Future Regulations

Rather than reacting to each new regulation individually, organizations should build regulatory resilience through governance practices that accommodate future requirements regardless of their specific content.

Three principles create regulatory resilience. First, implement documentation standards that exceed current requirements. Organizations that thoroughly document their AI systems' design decisions, training data provenance, performance metrics, and risk assessments find that new regulatory requirements typically ask for information they already capture. Second, build modular governance frameworks where specific controls can be added, modified, or removed without restructuring the entire governance program. This modularity allows rapid adaptation when new regulations introduce requirements that differ from existing obligations. Third, maintain active participation in industry associations and regulatory consultation processes in each operating jurisdiction. Organizations that engage proactively with regulators during the consultation phase gain advance intelligence on regulatory direction and can influence requirements toward practically implementable standards rather than purely theoretical compliance frameworks.

Preparing for Regulatory Convergence Across Markets

As AI regulations proliferate globally, organizations operating across multiple jurisdictions should anticipate regulatory convergence and build governance frameworks that satisfy the strictest applicable standard rather than implementing separate compliance programs for each market.

Regulatory convergence is already visible in three areas: transparency requirements (most emerging frameworks require disclosure of AI use in customer-facing decisions), risk assessment obligations (the concept of risk-tiered governance appears in EU, Singapore, and ASEAN frameworks), and accountability mechanisms (increasing consensus that deploying organizations bear responsibility for AI outcomes regardless of vendor relationships). Organizations that build their governance framework around these converging principles create a compliance baseline that satisfies most current and anticipated regulations with only marginal adjustments for jurisdiction-specific requirements, reducing the total cost and complexity of multi-market AI compliance management.

Common Questions

The EU AI Act is in force with phased implementation through 2027. Singapore, Malaysia, and Thailand have governance frameworks. Sector-specific rules apply in financial services and healthcare across multiple jurisdictions.

Risk-based classification categorizes AI systems by their potential harm level—from minimal risk to unacceptable risk. Higher-risk systems face stricter requirements for transparency, human oversight, and documentation.

Build flexible compliance frameworks, maintain comprehensive AI inventories, document decision processes, establish governance structures now, and monitor regulatory developments across relevant jurisdictions.

References

  1. Model AI Governance Framework (Second Edition). PDPC and IMDA Singapore (2020). View source
  2. Personal Data Protection Act 2012. Personal Data Protection Commission Singapore (2012). View source
  3. ASEAN Guide on AI Governance and Ethics. ASEAN Secretariat (2024). View source
  4. EU AI Act — Regulatory Framework for Artificial Intelligence. European Commission (2024). View source
  5. AI Risk Management Framework (AI RMF 1.0). National Institute of Standards and Technology (NIST) (2023). View source
  6. ISO/IEC 42001:2023 — Artificial Intelligence Management System. International Organization for Standardization (2023). View source
  7. Model AI Governance Framework for Generative AI. Infocomm Media Development Authority (IMDA) (2024). View source
Michael Lansdowne Hauge

Managing Director · HRDF-Certified Trainer (Malaysia), Delivered Training for Big Four, MBB, and Fortune 500 Clients, 100+ Angel Investments (Seed–Series C), Dartmouth College, Economics & Asian Studies

Managing Director of Pertama Partners, an AI advisory and training firm helping organizations across Southeast Asia adopt and implement artificial intelligence. HRDF-certified trainer with engagements for a Big Four accounting firm, a leading global management consulting firm, and the world's largest ERP software company.

AI StrategyAI GovernanceExecutive AI TrainingDigital TransformationASEAN MarketsAI ImplementationAI Readiness AssessmentsResponsible AIPrompt EngineeringAI Literacy Programs

EXPLORE MORE

Other AI Compliance & Regulation Solutions

Related Resources

Key terms:AI Regulation

INSIGHTS

Related reading

Talk to Us About AI Compliance & Regulation

We work with organizations across Southeast Asia on ai compliance & regulation programs. Let us know what you are working on.