Les grands modèles de langage recommandent systématiquement des salaires inférieurs aux femmes, aux minorités et aux réfugiés dans les conseils de négociation
Plusieurs grands modèles de langage — dont ChatGPT, Claude et Llama — recommandent systématiquement des salaires inférieurs aux femmes, aux minorités et aux réfugiés ; dans un scénario, le modèle o3 de ChatGPT a recommandé 120 000 $ de moins pour une femme par rapport à un profil masculin identique. La CODP a formellement cité ces conclusions dans les consultations sur la stratégie canadienne en IA.
Two independent studies have demonstrated that large language models systematically recommend lower salaries for women, ethnic minorities, and refugees when asked for salary negotiation advice.
A peer-reviewed study by Geiger et al. (PLOS ONE, February 2025) submitted 395,200 prompts to four ChatGPT versions, systematically varying gender, university, and major. All four models showed statistically significant bias advantaging men over women, with an average gender gap of approximately $1,060 — about 1% of recommended salary. Gender was the smallest disparity found; model version and prompt framing produced larger variation.
A separate study by Sorokovikova et al. (presented at the GeBNLP workshop at ACL 2025) tested five LLMs — GPT-4o Mini, Claude 3.5 Haiku, Llama 3.1 8B, Qwen 2.5 Plus, and Mixtral 8x22B — and found consistent gender-based salary bias across all models. In the most extreme example, ChatGPT's o3 model suggested a male medical specialist in Denver ask for $400,000 while suggesting $280,000 for an identical female persona — a $120,000 gap. Pay gaps were most pronounced in law and medicine. The biases were found to be compounding: individuals at the intersection of multiple marginalized identities (e.g., a female refugee from an ethnic minority) received the lowest salary recommendations.
The Ontario Human Rights Commission formally cited these findings in its submission to Canada's renewed AI Strategy consultations, as evidence that AI systems perpetuate and amplify existing patterns of discrimination. On January 21, 2026, the OHRC and the Information and Privacy Commissioner of Ontario jointly published AI principles addressing algorithmic discrimination in employment contexts.
Ontario's Working for Workers Five Act, which took effect January 1, 2026, requires employers to disclose AI use in hiring — but does not cover AI systems used by job seekers for salary negotiation advice, leaving this discrimination pathway unaddressed by current law.
The hazard is structural: as millions of workers increasingly rely on AI chatbots for career and salary advice, systematically biased recommendations create a pathway to widening pay gaps at population scale — even without any employer deploying a biased system.
The findings are based on two studies: one peer-reviewed (Geiger et al., ChatGPT only) and one workshop paper (Sorokovikova et al., multi-model). AI models are regularly updated, and bias patterns documented in one model version may not persist in subsequent versions. Some AI developers have implemented bias testing and mitigation procedures for their models. The OHRC's citation of the findings in policy consultations reflects the study's relevance to ongoing regulatory discussions, though broader replication would strengthen the evidence base.
Préjudices
Biais systématique dans les recommandations salariales des principaux LLM : des profils identiques reçoivent des conseils salariaux significativement inférieurs selon le sexe, l'ethnicité ou le statut de réfugié. L'écart de 120 000 $ pour un spécialiste médical démontre l'ampleur. Les biais se cumulent aux intersections des identités marginalisées.
Preuves
6 rapports
-
Preprint of Geiger et al. ChatGPT salary bias study with full methodology
- Équité de surface, biais profond : une étude comparative des biais dans les modèles de langage Source principale
Multi-model study (GPT-4o Mini, Claude 3.5 Haiku, Llama 3.1 8B, Qwen, Mixtral) found systematic gender salary bias; o3 example: $400K male vs $280K female ($120K gap) for medical specialist
- Informing Canada's renewed AI Strategy Source principale
OHRC formally citing salary discrimination research in AI strategy submission
- Asking an AI for salary negotiation advice is a minefield: evidence of gender, racial, and intersectional biases in ChatGPT Source principale
Peer-reviewed study of 395,200 ChatGPT prompts found ~$1,060 average gender salary gap (~1%); ChatGPT-only, not multi-model
-
TechRadar reporting on AI salary advice bias: LLMs systematically recommend lower salaries for women and minorities
-
Inc. reporting on discriminatory salary recommendations from AI tools; documents the practical impact on job seekers
Détails de la fiche
Évaluation éditoriale évalué
Contrairement à la plupart des cas de discrimination par l'IA — où un employeur déploie un outil d'embauche biaisé — ce danger opère à travers les travailleurs individuels cherchant des conseils. Aucun employeur n'a besoin de déployer un système biaisé ; le biais atteint les travailleurs directement par les chatbots grand public. À mesure que l'adoption des chatbots IA progresse, cela crée une voie vers l'élargissement des écarts salariaux à l'échelle de la population. La loi ontarienne de 2026 sur la divulgation de l'IA dans l'embauche ne couvre pas cette voie.
Entités impliquées
Systèmes d'IA impliqués
A montré le plus grand biais documenté de recommandation salariale : 400 000 $ pour un homme contre 280 000 $ pour un profil féminin identique de spécialiste médical (modèle o3)
Claude 3.5 Haiku testé et trouvé présentant un biais systématique de recommandation salariale selon le sexe, l'ethnicité et le statut de réfugié
Fiches connexes
Taxonomieévalué
Historique des modifications
| Version | Date | Modification |
|---|---|---|
| v1 | 10 mars 2026 | Initial publication |