Pilot phase: CAIM is under construction. Records are provisional, based on public sources, and have not yet been peer-reviewed. Feedback welcome.
Escalating Significant Confidence: high

AI agents are being deployed at scale in Canada — TD Bank (25,000+ Copilot users), Scotiabank, CGI, Telus, federal government (Coveo MOU) — while safety research documents systemic risks. The 2025 AI Agent Index found 25/30 deployed agents disclose no safety results. KPMG Canada: 27% of businesses have deployed agentic AI. The first large-scale AI-orchestrated cyberattack occurred in November 2025. Canada has no governance framework for agentic AI.

Identified: October 1, 2024 Last assessed: March 10, 2026

AI systems are increasingly deployed as autonomous agents — executing multi-step tasks, browsing the web, writing and running code, making purchases, interacting with APIs, and operating computer interfaces — with minimal human oversight between steps. This represents a qualitative shift from AI as a tool that responds to individual prompts to AI as an actor that pursues goals across extended action sequences, where errors compound and unintended consequences accumulate.

The deployment of agentic AI accelerated rapidly in 2025-2026. Anthropic released Claude computer use capabilities (October 2024), enabling AI to operate computer interfaces autonomously. OpenAI launched Operator (January 2025), an agent that performs web-based tasks on behalf of users. Google DeepMind deployed Mariner for web browsing and Jules for coding. Coding agents achieved dramatic capability gains: on the SWE-bench benchmark (resolving real GitHub issues), performance rose from under 5% in early 2024 to over 50% by mid-2025, with the best systems now resolving issues that would take experienced developers hours. Companies like Cognition (Devin), Factory, and others raised hundreds of millions of dollars for autonomous coding products.

The International AI Safety Report 2026 explicitly identifies agentic AI as an emerging risk category, noting that "the deployment of AI systems as autonomous agents introduces novel risk vectors including compounding errors across action sequences, difficulty of attributing responsibility for agent actions, and potential for agents to take actions with irreversible real-world consequences." The report notes that safety evaluation methodologies developed for single-turn interactions are inadequate for agentic systems that operate over extended time horizons.

The risk structure is distinct from other AI hazards. In a standard AI deployment, the human reviews each output before acting on it. In agentic deployment, the AI takes a sequence of actions — searching, reading, clicking, typing, submitting — with the human seeing only the final result. Each step has some probability of error or misinterpretation; across a sequence of dozens or hundreds of steps, errors compound. An agent that misunderstands a task specification may take confident, well-executed actions in the wrong direction — purchasing the wrong items, sending incorrect communications, modifying the wrong files, or interacting with the wrong systems — before the human can intervene.

Accountability gaps are structural. When an AI agent sends an email, makes a purchase, modifies a database, or files a form on behalf of a user or organization, who bears responsibility if the action is incorrect, harmful, or unauthorized? Existing legal frameworks assume a human decision-maker at each point of action. Agentic AI disrupts this assumption without providing an alternative accountability structure.

Multi-agent dynamics add complexity. As organizations deploy multiple AI agents that interact with each other — one agent's output becoming another's input — emergent behaviours can arise that no individual agent was designed to produce. Market dynamics, information cascades, and coordination failures become possible at machine speed without human intervention points.

Harms

Agentic AI systems execute multi-step tasks (browsing, coding, purchasing, API interactions) with minimal human oversight between steps. Errors compound across action sequences — each step has some error probability, and across dozens of steps, unintended consequences accumulate without human review.

Autonomy UnderminedService DisruptionModeratePopulation

Existing legal liability frameworks assume a human decision-maker at each consequential step. Agentic AI that autonomously takes actions (making purchases, sending communications, modifying systems) creates an accountability gap where no entity bears clear responsibility for the agent's autonomous actions.

Autonomy UnderminedSignificantPopulation

Evidence

10 reports

  1. Introducing Computer Use Primary source
    Official — Anthropic (Oct 22, 2024)

    Anthropic released Claude computer use capabilities enabling AI to operate computer interfaces

  2. Introducing Operator Primary source
    Official — OpenAI (Jan 23, 2025)

    OpenAI launched Operator for autonomous web-based tasks

  3. The 2025 AI Agent Index Primary source
    Academic — MIT / Cambridge / Harvard / Stanford (Feb 1, 2025)

    25/30 deployed agents disclose no internal safety results; 23/30 have no third-party testing

  4. Official — Anthropic (Nov 1, 2025)

    First documented large-scale AI-orchestrated cyberattack: Claude Code used to perform 80-90% of attack work autonomously against ~30 targets

  5. Academic — International AI Safety Report (Feb 3, 2026)

    IASR 2026 identifies agentic AI as an emerging risk category with novel risk vectors

  6. Official — Treasury Board of Canada Secretariat (Jan 1, 2023)

    Canada's Directive on Automated Decision-Making does not cover broader agentic AI deployment

  7. Academic — Princeton NLP / SWE-bench (Jan 1, 2024)

    SWE-bench performance rose from under 5% to over 50% between early 2024 and mid-2025

  8. Academic — DeepMind / Anthropic / CMU / Harvard (Feb 1, 2025)

    Taxonomy of multi-agent failure modes: miscoordination, conflict, collusion; 50+ researchers

  9. Media — Microsoft Source Canada (Jan 1, 2026)

    TD Bank deployed Copilot to 25,000+ colleagues; Scotiabank pioneering agentic AI with EY and Microsoft

  10. Media — MSP Corp / KPMG Canada (Jan 1, 2026)

    KPMG Canada: 27% deployed agentic AI, 64% experimenting, 57% planning investment within 6 months

Record details

Policy Recommendationsassessed

Develop a legal liability framework for actions taken by AI agents on behalf of persons or organizations

International AI Safety Report 2026

Require mandatory disclosure when AI agents interact with third parties on behalf of users

IASR 2026 / EU AI Act

Establish human oversight checkpoint requirements for AI agent actions with financial, legal, or safety consequences

IASR 2026

Editorial Assessment assessed

Agentic AI is the defining capability shift in AI deployment and CAIM's schema includes multi_agent_dynamics as an AI pathway — but no hazard used it. AI agents are taking real-world actions (sending messages, making purchases, modifying systems) with minimal human oversight. Performance on coding tasks grew from <5% to >50% in 18 months. The IASR 2026 explicitly identifies agentic AI as an emerging risk. Canada has no liability framework, no disclosure requirement, and no oversight standards for AI agents — a gap that will matter increasingly as organizations delegate consequential tasks to these systems.

Entities Involved

Anthropic
developer
OpenAI
developer

Related Records

Taxonomyassessed

Domain
Public ServicesRetail & CommerceFinance & Banking
Harm type
Safety IncidentEconomic HarmService Disruption
AI pathway
Deployment ContextMonitoring AbsentUse Beyond Intended ScopeMulti-Agent Dynamics
Lifecycle phase
DeploymentMonitoring

Changelog

Changelog
VersionDateChange
v1Mar 10, 2026Initial publication

Version 1