Abstract
Singapore's Infocomm Media Development Authority proposed governance framework for agentic AI systems. Examines unique risks of autonomous AI agents including loss of human oversight, cascading errors in multi-agent systems, and accountability gaps. Proposes principle-based guardrails for deploying AI agents in enterprise settings.
About This Research
Publisher: Singapore IMDA Year: 2025 Type: Governance Framework
Source: Proposed Framework for Governing Agentic AI Systems
Relevance
Industries: Cross-Industry Pillars: AI Compliance & Regulation, AI Governance & Risk Management, Board & Executive Oversight Use Cases: AI Agents & Autonomous Systems Regions: Singapore
Autonomy Boundaries and Action Spaces
The framework introduces the concept of formally defined autonomy boundaries that constrain agentic AI systems to pre-approved action spaces. These boundaries operate at multiple levels: hard constraints that cannot be overridden regardless of the agent's objective assessment, soft constraints that can be escalated to human supervisors for exception approval, and contextual constraints that adjust dynamically based on environmental risk levels. This layered architecture enables productive autonomy while maintaining meaningful guardrails against catastrophic or irreversible actions.
Accountability Attribution in Multi-Agent Environments
As agentic systems increasingly operate in environments where multiple AI agents interact, traditional single-point accountability models become inadequate. The framework proposes a distributed accountability architecture that assigns proportional responsibility based on each agent's contribution to an outcome, the foreseeability of the interaction effects, and the adequacy of safety measures implemented by each agent's operator. This approach draws on established principles from tort law and organisational liability theory, adapting them for the novel characteristics of autonomous AI interaction.
Intervention Protocols and Human Override
Meaningful human oversight of agentic systems requires more than a theoretical kill switch. The framework mandates intervention protocols that ensure human operators can understand an agent's current state, predict its planned actions, and execute override commands within timeframes sufficient to prevent unacceptable outcomes. These protocols include mandatory state transparency interfaces, action preview mechanisms, and graduated intervention levels ranging from pace reduction to complete suspension of autonomous operation.