Pilot phase: CAIM is under construction. Records are provisional, based on public sources, and have not yet been peer-reviewed. Feedback welcome.
Active Critical Confidence: medium

Frontier AI models are demonstrating capabilities relevant to biological and chemical weapon development that multiple developers cannot confidently exclude as providing meaningful uplift. Canada hosts BSL-4 infrastructure with proven insider-threat history, chairs the international assessment identifying this risk, and signed commitments recognizing it — ; it has no dedicated AI-biosecurity assessment or evaluation mandate.

Identified: January 25, 2024 Last assessed: March 10, 2026

Frontier AI systems are demonstrating capabilities relevant to biological and chemical weapon development that multiple AI developers have been unable to confidently rule out as providing meaningful uplift to non-expert actors. In May 2025, Anthropic activated ASL-3 protections — its second-highest safety tier — for Claude Opus 4 because it could not "confidently rule out the ability of their most advanced model to uplift people with basic STEM backgrounds" for bio/chem weapons development. This was the first time any AI developer deployed a model under its highest activated safety level specifically due to biosecurity concerns.

In June 2025, RAND Corporation tested three frontier models (Llama 3.1 405B, ChatGPT-4o, and Claude 3.5 Sonnet) and found that all three "successfully provide accurate instructions and guidance for recovering a live poliovirus from a construct built from commercially obtained synthetic DNA." RAND argued that existing safety assessments "underestimate this risk" due to flawed assumptions about the tacit knowledge barriers that remain. The IASR 2026, chaired by Canadian researcher Yoshua Bengio, reports that "in one study a recent model outperformed 94% of domain experts at troubleshooting virology laboratory protocols" — referring to OpenAI's o3 achieving 43.8% accuracy on SecureBio's Virology Capabilities Test versus 22.1% average for human experts in their sub-specialties.

A separate line of evidence concerns AI protein design tools. In October 2025, research published in Science found that AI protein design tools generated over 70,000 DNA sequences for variant forms of controlled toxic proteins, and one screening tool missed more than 75% of potential toxins. After a 10-month remediation effort, screening was improved to catch 97% of high-risk sequences — but the episode demonstrated that biosecurity controls have not kept pace with capability.

Canada's exposure to this risk is direct and multi-layered. Canada hosts the sole BSL-4 facility in the country — the Canadian Science Centre for Human and Animal Health in Winnipeg — which works with Ebola, Marburg, and other high-consequence pathogens. The Qiu/Cheng incident (scientists terminated 2019-2021, CSIS investigation confirmed 2024) demonstrated that insider threats at this facility are real: CSIS found intentional transfer of scientific knowledge and materials related to Ebola and Marburg viruses. Canada has 17+ academic BSL-3 facilities through the CCABL3 consortium. VIDO at the University of Saskatchewan is constructing a second BSL-4, which will be the only non-government CL4 in Canada.

Canada's Sensitive Technology List (published February 2025) explicitly identifies the convergence: "advancements in nanotechnology, synthetic biology, artificial intelligence and sensing technologies could provide enhancements to existing weapons, such as biological/chemical weapons." Canada signed the Seoul Declaration (May 2024) recognizing that frontier AI could "meaningfully assist non-state actors in advancing the development, production, acquisition or use of chemical or biological weapons."

Yet Canada has published no dedicated assessment of AI-enabled bio/chem weapon risk. The Canadian AI Safety Institute (CAISI, launched November 2024, $50M over five years) does not yet explicitly include biosecurity evaluations in its public mandate. The gap between Canada's international commitments acknowledging this risk and its domestic institutional capacity to evaluate it is the core governance concern.

Anthropic's activation of ASL-3 protections — the first such deployment by any AI developer — represents a case where voluntary safety frameworks functioned as designed: the company identified a risk during pre-deployment evaluation and applied its highest activated safety tier in response. Several other frontier AI developers have also implemented pre-deployment biosecurity evaluations. The debate centers on whether voluntary measures are sufficient or whether mandatory requirements are needed to ensure consistent evaluation across all developers.

Harms

Frontier AI models provide actionable knowledge for biological weapon development. RAND found that three frontier models 'successfully provide accurate instructions and guidance for recovering a live poliovirus from a construct built from commercially obtained synthetic DNA.' Anthropic activated ASL-3 protections for the first time due to inability to rule out meaningful biosecurity uplift.

CBRN UpliftCriticalPopulation

AI protein design tools create dual-use risks. De novo protein design enables creation of novel molecular structures with potential applications including toxin engineering, with the IASR 2026 warning of 'dual-use risks' from such tools.

CBRN UpliftCriticalPopulation

Canada lacks AI-specific biosecurity governance: no mandatory pre-deployment biosecurity evaluation for frontier AI models, no national assessment of AI-enabled biosecurity risk, and no regulatory framework connecting AI safety evaluation to biosecurity oversight despite Canada's BSL-4 infrastructure and role chairing the IASR.

Editorial note:This is a governance gap, not a materialized harm. Its severity is assessed against the potential consequences if the gap persists as AI biosecurity capabilities advance.

CBRN UpliftSignificantSector

Evidence

9 reports

  1. Official — Anthropic (May 1, 2025)

    First model deployed with ASL-3 protections due to biosecurity concerns

  2. Academic — RAND Corporation (Jun 1, 2025)

    Three frontier models provided accurate poliovirus recovery instructions from synthetic DNA

  3. Academic — Science (Oct 1, 2025)

    AI protein design tools generated 70K+ toxic protein sequences; screening missed 75%+

  4. Academic — International AI Safety Report (Feb 3, 2026)

    AI model outperformed 94% of domain experts on virology lab protocols

  5. Academic — Centre for International Governance Innovation (Jan 1, 2024)

    LLMs showed 80% improvement in instructions for releasing lethal substances in 2024

  6. Official — OpenAI (Aug 1, 2024)

    AI access brought student bio/chem performance to expert baseline on magnification/formulation

  7. Official — Anthropic (Jan 1, 2025)

    Claude exceeded expert baselines on molecular biology and cloning workflow benchmarks

  8. Academic — Centre for International Governance Innovation (Jan 1, 2025)

    Canada lacks a dedicated biosecurity strategy

  9. Official — Government of Canada (Feb 6, 2025)

    Canada's Sensitive Technology List identifies AI-biosecurity convergence

Record details

Policy Recommendationsassessed

Include biosecurity evaluation explicitly in CAISI's mandate and research priorities

Centre for International Governance Innovation

Develop a comprehensive Canadian biosecurity strategy addressing AI-enabled threats

Centre for International Governance Innovation

Require mandatory pre-deployment biosecurity assessment for frontier models deployed in Canada

International AI Safety Report 2026

Editorial Assessment assessed

Multiple AI developers have activated their highest safety protocols because they cannot rule out that their models provide meaningful assistance for bio/chem weapon development. The IASR 2026, chaired by a Canadian researcher, identifies this as a key emerging risk. Canada hosts BSL-4 infrastructure, 17+ BSL-3 facilities, and has signed international commitments recognizing AI-CBRN risk. CAISI's current mandate does not explicitly include biosecurity evaluation, though some frontier AI developers have implemented voluntary pre-deployment biosecurity assessments.

Entities Involved

Taxonomyassessed

Domain
HealthcareDefence & Security
Harm type
Safety IncidentCBRN Uplift
AI pathway
Use Beyond Intended ScopeCapability Beyond SpecificationSafety Mechanism Ineffective
Lifecycle phase
DeploymentEvaluation

Changelog

Changelog
VersionDateChange
v1Mar 10, 2026Initial publication

Version 1