Pilot phase: CAIM is under construction. Records are provisional, based on public sources, and have not yet been peer-reviewed. Feedback welcome.
Active Significant Confidence: medium

Multiple LLMs — including ChatGPT, Claude, and Llama — systematically recommend lower salaries for women, minorities, and refugees; in one scenario ChatGPT's o3 recommended $120K less for a woman than an identical male profile. OHRC formally cited findings in Canada's AI strategy consultations.

Identified: January 1, 2026 Last assessed: March 10, 2026

Two independent studies have demonstrated that large language models systematically recommend lower salaries for women, ethnic minorities, and refugees when asked for salary negotiation advice.

A peer-reviewed study by Geiger et al. (PLOS ONE, February 2025) submitted 395,200 prompts to four ChatGPT versions, systematically varying gender, university, and major. All four models showed statistically significant bias advantaging men over women, with an average gender gap of approximately $1,060 — about 1% of recommended salary. Gender was the smallest disparity found; model version and prompt framing produced larger variation.

A separate study by Sorokovikova et al. (presented at the GeBNLP workshop at ACL 2025) tested five LLMs — GPT-4o Mini, Claude 3.5 Haiku, Llama 3.1 8B, Qwen 2.5 Plus, and Mixtral 8x22B — and found consistent gender-based salary bias across all models. In the most extreme example, ChatGPT's o3 model suggested a male medical specialist in Denver ask for $400,000 while suggesting $280,000 for an identical female persona — a $120,000 gap. Pay gaps were most pronounced in law and medicine. The biases were found to be compounding: individuals at the intersection of multiple marginalized identities (e.g., a female refugee from an ethnic minority) received the lowest salary recommendations.

The Ontario Human Rights Commission formally cited these findings in its submission to Canada's renewed AI Strategy consultations, as evidence that AI systems perpetuate and amplify existing patterns of discrimination. On January 21, 2026, the OHRC and the Information and Privacy Commissioner of Ontario jointly published AI principles addressing algorithmic discrimination in employment contexts.

Ontario's Working for Workers Five Act, which took effect January 1, 2026, requires employers to disclose AI use in hiring — but does not cover AI systems used by job seekers for salary negotiation advice, leaving this discrimination pathway unaddressed by current law.

The hazard is structural: as millions of workers increasingly rely on AI chatbots for career and salary advice, systematically biased recommendations create a pathway to widening pay gaps at population scale — even without any employer deploying a biased system.

The findings are based on two studies: one peer-reviewed (Geiger et al., ChatGPT only) and one workshop paper (Sorokovikova et al., multi-model). AI models are regularly updated, and bias patterns documented in one model version may not persist in subsequent versions. Some AI developers have implemented bias testing and mitigation procedures for their models. The OHRC's citation of the findings in policy consultations reflects the study's relevance to ongoing regulatory discussions, though broader replication would strengthen the evidence base.

Harms

Systematic salary recommendation bias across major LLMs: identical profiles receive significantly lower salary advice based on sex, ethnicity, or refugee status. The $120,000 gap for a medical specialist demonstrates the magnitude. Biases compound at intersections of marginalized identities.

Discrimination & RightsEconomic HarmSignificantPopulation

Evidence

6 reports

  1. Academic — arXiv (Sep 23, 2024)

    Preprint of Geiger et al. ChatGPT salary bias study with full methodology

  2. Academic — GeBNLP Workshop, ACL 2025 (Jun 1, 2025)

    Multi-model study (GPT-4o Mini, Claude 3.5 Haiku, Llama 3.1 8B, Qwen, Mixtral) found systematic gender salary bias; o3 example: $400K male vs $280K female ($120K gap) for medical specialist

  3. Regulatory — Ontario Human Rights Commission (Jan 21, 2026)

    OHRC formally citing salary discrimination research in AI strategy submission

  4. Academic — PubMed / peer-reviewed journal (Feb 1, 2026)

    Peer-reviewed study of 395,200 ChatGPT prompts found ~$1,060 average gender salary gap (~1%); ChatGPT-only, not multi-model

  5. Media — TechRadar (Feb 10, 2026)

    TechRadar reporting on AI salary advice bias: LLMs systematically recommend lower salaries for women and minorities

  6. Media — Inc. (Feb 15, 2026)

    Inc. reporting on discriminatory salary recommendations from AI tools; documents the practical impact on job seekers

Record details

Editorial Assessment assessed

Unlike most AI discrimination cases — where an employer deploys a biased tool — this hazard operates through individual workers seeking advice from consumer-facing chatbots. A peer-reviewed study found that major LLMs systematically recommend lower salaries for women, minorities, and refugees. The OHRC cited these findings in Canada's AI strategy consultations. Ontario's 2026 AI-in-hiring disclosure law does not cover consumer-facing AI advisory services. Broader replication of the study's findings would strengthen the evidence base for policy responses.

Entities Involved

AI Systems Involved

ChatGPT

Showed largest documented salary recommendation bias: $400K for male vs $280K for identical female medical specialist profile (o3 model)

Claude

Claude 3.5 Haiku tested and found to exhibit systematic salary recommendation bias across sex, ethnicity, and refugee status

Related Records

Taxonomyassessed

Domain
EmploymentPublic Services
Harm type
Discrimination & RightsEconomic Harm
AI pathway
Training Data OriginDeployment Context
Lifecycle phase
TrainingDeployment

Changelog

Changelog
VersionDateChange
v1Mar 10, 2026Initial publication

Version 2