This site is a work-in-progress prototype.
Reported Severity: Significant Version 1

A joint investigation by federal and provincial privacy commissioners — the first into a large language model in Canada — is examining whether OpenAI's collection and generation of personal information about Canadians violates Canadian privacy law.

Occurred: April 4, 2023 Reported: April 4, 2023

Narrative

In April 2023, Canada’s Privacy Commissioner launched an investigation into OpenAI after receiving a complaint about ChatGPT’s handling of personal information. The investigation was subsequently joined by privacy commissioners in Quebec, British Columbia, and Alberta in May 2023, making it one of the first joint federal-provincial privacy investigations into a large language model.

The investigation is examining whether OpenAI violated the Personal Information Protection and Electronic Documents Act (PIPEDA) on multiple grounds: collecting personal information of Canadians without consent through web scraping to build training datasets, failing to ensure the accuracy of personal information generated by ChatGPT, and lacking transparency about how personal data was collected, used, and processed. The scope includes ChatGPT’s generation of false biographical statements about identifiable Canadians and whether this constitutes a failure to meet accuracy obligations under Canadian privacy law.

As of early 2026, the investigation remains ongoing. Privacy Commissioner Philippe Dufresne described it as his “ongoing investigation into OpenAI” in a February 2026 statement to Parliament. The investigation is one of the Privacy Commissioner’s most significant regulatory actions concerning AI systems and is expected to address whether companies deploying large language models in Canada bear privacy obligations for the outputs those systems generate — not just the data they consume.

The investigation addresses a tension in generative AI: systems trained on vast internet data inevitably absorb personal information about real people, and their probabilistic text generation can produce confidently stated falsehoods about identifiable individuals. The outcome of this investigation will help determine whether current Canadian privacy frameworks have applicability to these novel AI harms.

Harms

OpenAI collected personal information of Canadians without consent through web scraping to build ChatGPT's training datasets, and failed to provide transparency about how personal data was collected, used, and processed.

Significant Population

ChatGPT generates false biographical statements about identifiable Canadians, presenting fabricated personal details with apparent confidence, constituting a potential failure to meet accuracy obligations under Canadian privacy law.

Moderate Population

Affected Populations

  • Canadian ChatGPT users
  • individuals about whom ChatGPT generates false information
  • privacy rights advocates

Entities Involved

OpenAI
developer

Developed and operates ChatGPT; under joint investigation by federal and provincial privacy commissioners for allegedly collecting personal information of Canadians without consent and generating false biographical statements about identifiable individuals

Launched the investigation into OpenAI in April 2023 and coordinated with provincial privacy commissioners in Quebec, BC, and Alberta to conduct a joint federal-provincial investigation — the first into a large language model in Canada

AI Systems Involved

ChatGPT

The AI system under investigation for its training data collection practices and its generation of false personal information about identifiable Canadians

Responses & Outcomes

Office of the Privacy Commissioner of Canada

Launched formal investigation into OpenAI's ChatGPT after receiving a complaint about its handling of personal information

Office of the Privacy Commissioner of Canada

Expanded investigation into a joint federal-provincial effort with privacy commissioners of Quebec, British Columbia, and Alberta

AI System Context

OpenAI's ChatGPT large language model, trained on data scraped from the internet including personal information of Canadians, which generates text that can include false or fabricated biographical details about real individuals.

Preventive Measures

  • Require AI companies operating in Canada to implement accessible mechanisms for Canadians to identify, challenge, and correct false personal information generated by their systems
  • Mandate transparency about the personal information used to train AI models, including data sourced from Canadian individuals and institutions
  • Establish accuracy obligations for AI systems that generate statements about identifiable individuals
  • Require AI companies to conduct privacy impact assessments under Canadian law before deploying systems trained on data that includes personal information of Canadians

Related Records

Taxonomy

Domain
Telecommunications
Harm type
Privacy & Data ExposureMisinformation
AI involvement
Training Data IssueModel ConfabulationMonitoring Gap
Lifecycle phase
TrainingDeploymentMonitoring

Sources

  1. Privacy Commissioner launches investigation into ChatGPT Official — Office of the Privacy Commissioner of Canada (Apr 4, 2023)
  2. Joint investigation of ChatGPT by privacy commissioners Official — Office of the Privacy Commissioner of Canada (May 25, 2023)
  3. Canada's privacy watchdog launches probe into ChatGPT Media — CBC News (Apr 4, 2023)

Changelog

VersionDateChange
v1 Mar 8, 2026 Initial publication