AI-Enabled Biological and Chemical Weapon Development Risk
Frontier AI models are demonstrating capabilities relevant to biological and chemical weapon development that multiple developers cannot confidently exclude as providing meaningful uplift. Canada hosts BSL-4 infrastructure with proven insider-threat history, chairs the international assessment identifying this risk, and signed commitments recognizing it — ; it has no dedicated AI-biosecurity assessment or evaluation mandate.
Frontier AI systems are demonstrating capabilities relevant to biological and chemical weapon development that multiple AI developers have been unable to confidently rule out as providing meaningful uplift to non-expert actors. In May 2025, Anthropic activated ASL-3 protections — its second-highest safety tier — for Claude Opus 4 because it could not "confidently rule out the ability of their most advanced model to uplift people with basic STEM backgrounds" for bio/chem weapons development. This was the first time any AI developer deployed a model under its highest activated safety level specifically due to biosecurity concerns.
In June 2025, RAND Corporation tested three frontier models (Llama 3.1 405B, ChatGPT-4o, and Claude 3.5 Sonnet) and found that all three "successfully provide accurate instructions and guidance for recovering a live poliovirus from a construct built from commercially obtained synthetic DNA." RAND argued that existing safety assessments "underestimate this risk" due to flawed assumptions about the tacit knowledge barriers that remain. The IASR 2026, chaired by Canadian researcher Yoshua Bengio, reports that "in one study a recent model outperformed 94% of domain experts at troubleshooting virology laboratory protocols" — referring to OpenAI's o3 achieving 43.8% accuracy on SecureBio's Virology Capabilities Test versus 22.1% average for human experts in their sub-specialties.
A separate line of evidence concerns AI protein design tools. In October 2025, research published in Science found that AI protein design tools generated over 70,000 DNA sequences for variant forms of controlled toxic proteins, and one screening tool missed more than 75% of potential toxins. After a 10-month remediation effort, screening was improved to catch 97% of high-risk sequences — but the episode demonstrated that biosecurity controls have not kept pace with capability.
Canada's exposure to this risk is direct and multi-layered. Canada hosts the sole BSL-4 facility in the country — the Canadian Science Centre for Human and Animal Health in Winnipeg — which works with Ebola, Marburg, and other high-consequence pathogens. The Qiu/Cheng incident (scientists terminated 2019-2021, CSIS investigation confirmed 2024) demonstrated that insider threats at this facility are real: CSIS found intentional transfer of scientific knowledge and materials related to Ebola and Marburg viruses. Canada has 17+ academic BSL-3 facilities through the CCABL3 consortium. VIDO at the University of Saskatchewan is constructing a second BSL-4, which will be the only non-government CL4 in Canada.
Canada's Sensitive Technology List (published February 2025) explicitly identifies the convergence: "advancements in nanotechnology, synthetic biology, artificial intelligence and sensing technologies could provide enhancements to existing weapons, such as biological/chemical weapons." Canada signed the Seoul Declaration (May 2024) recognizing that frontier AI could "meaningfully assist non-state actors in advancing the development, production, acquisition or use of chemical or biological weapons."
Yet Canada has published no dedicated assessment of AI-enabled bio/chem weapon risk. The Canadian AI Safety Institute (CAISI, launched November 2024, $50M over five years) does not yet explicitly include biosecurity evaluations in its public mandate. The gap between Canada's international commitments acknowledging this risk and its domestic institutional capacity to evaluate it is the core governance concern.
Anthropic's activation of ASL-3 protections — the first such deployment by any AI developer — represents a case where voluntary safety frameworks functioned as designed: the company identified a risk during pre-deployment evaluation and applied its highest activated safety tier in response. Several other frontier AI developers have also implemented pre-deployment biosecurity evaluations. The debate centers on whether voluntary measures are sufficient or whether mandatory requirements are needed to ensure consistent evaluation across all developers.
Harms
Frontier AI models provide actionable knowledge for biological weapon development. RAND found that three frontier models 'successfully provide accurate instructions and guidance for recovering a live poliovirus from a construct built from commercially obtained synthetic DNA.' Anthropic activated ASL-3 protections for the first time due to inability to rule out meaningful biosecurity uplift.
AI protein design tools create dual-use risks. De novo protein design enables creation of novel molecular structures with potential applications including toxin engineering, with the IASR 2026 warning of 'dual-use risks' from such tools.
Canada lacks AI-specific biosecurity governance: no mandatory pre-deployment biosecurity evaluation for frontier AI models, no national assessment of AI-enabled biosecurity risk, and no regulatory framework connecting AI safety evaluation to biosecurity oversight despite Canada's BSL-4 infrastructure and role chairing the IASR.
Editorial note:This is a governance gap, not a materialized harm. Its severity is assessed against the potential consequences if the gap persists as AI biosecurity capabilities advance.
Evidence
9 reports
- Activating ASL-3 Protections Primary source
First model deployed with ASL-3 protections due to biosecurity concerns
-
Three frontier models provided accurate poliovirus recovery instructions from synthetic DNA
-
AI protein design tools generated 70K+ toxic protein sequences; screening missed 75%+
- International AI Safety Report 2026 Primary source
AI model outperformed 94% of domain experts on virology lab protocols
-
LLMs showed 80% improvement in instructions for releasing lethal substances in 2024
-
AI access brought student bio/chem performance to expert baseline on magnification/formulation
-
Claude exceeded expert baselines on molecular biology and cloning workflow benchmarks
-
Canada lacks a dedicated biosecurity strategy
-
Canada's Sensitive Technology List identifies AI-biosecurity convergence
Record details
Policy Recommendationsassessed
Include biosecurity evaluation explicitly in CAISI's mandate and research priorities
Centre for International Governance InnovationDevelop a comprehensive Canadian biosecurity strategy addressing AI-enabled threats
Centre for International Governance InnovationRequire mandatory pre-deployment biosecurity assessment for frontier models deployed in Canada
International AI Safety Report 2026Editorial Assessment assessed
Multiple AI developers have activated their highest safety protocols because they cannot rule out that their models provide meaningful assistance for bio/chem weapon development. The IASR 2026, chaired by a Canadian researcher, identifies this as a key emerging risk. Canada hosts BSL-4 infrastructure, 17+ BSL-3 facilities, and has signed international commitments recognizing AI-CBRN risk. CAISI's current mandate does not explicitly include biosecurity evaluation, though some frontier AI developers have implemented voluntary pre-deployment biosecurity assessments.
Entities Involved
Taxonomyassessed
Changelog
| Version | Date | Change |
|---|---|---|
| v1 | Mar 10, 2026 | Initial publication |