"AI systems show measurable political bias."
Related Claims
Evidence9
A 2026 audit of 26 LLMs across three political tests and a ~27,000-article news-labeling task found 96.3% clustered in the libertarian-left, and models labeled neutral news as left-leaning while spotting far-left content 19.2% vs far-right 2.0.
Researchers audited 26 prominent LLMs with three political orientation inventories and a large news-labeling task of about 27,000 articles.
Across the inventories, 96.3% of models clustered in the libertarian-left. In the news task, models systematically labeled neutral articles as left-leaning and detected far-left content more often than far-right (19.2% vs 2.0%).
Researchers audited 26 prominent LLMs with three political orientation inventories and a large news-labeling task of about 27,000 articles.
Across the inventories, 96.3% of models clustered in the libertarian-left. In the news task, models systematically...
Using benchmarks built from 2,701 Dutch, 10,584 Norwegian, and 2,480 Spanish parliamentary votes, a 2026 study found LLMs show left-leaning or centrist tendencies and negative bias toward right-conservative parties.
The paper aligns model-predicted votes with verified parliamentary records and places model positions in the same two-dimensional CHES space as real parties.
Across the three national benchmarks, state-of-the-art LLMs showed left-leaning or centrist tendencies and clear negative bias toward right-conservative parties.
The paper aligns model-predicted votes with verified parliamentary records and places model positions in the same two-dimensional CHES space as real parties.
Across the three national benchmarks, state-of-the-art LLMs showed left-leaning or centrist...
In a U.S. election simulation, 18 LLMs preferred a Democratic nominee, and in a study of 935 registered voters, LLM conversations shifted the voting margin from 0.7% to 4.6% toward the Democrat.
The paper first shows that 18 open- and closed-weight models prefer a Democratic nominee when answering candidate-policy questions.
In a separate experiment with 935 registered voters who chatted with LLMs for five exchanges, the voting margin moved from 0.7% to 4.6% toward the Democratic nominee.
The paper first shows that 18 open- and closed-weight models prefer a Democratic nominee when answering candidate-policy questions.
In a separate experiment with 935 registered voters who chatted with LLMs for five exchanges, the voting margin moved from...
A 2025 study compared ChatGPT to European Social Survey responses across 32 countries and 10 waves; its answers place it at 2.84 on a 0-10 left-right scale versus its self-placement at 4.19, with extreme-left alignment in 8 of 14 questions.
ChatGPT answered 16 political economy questions on inequality, environment, civil rights, and government size. Researchers matched those answers to European Social Survey respondents across 32 countries and 10 waves.
Those matched respondents place ChatGPT at 2.84 on the 0-10 left-right scale, left of its self-assessed 4.19, and the study reports extreme-left alignment in 8 of 14 questions.
ChatGPT answered 16 political economy questions on inequality, environment, civil rights, and government size. Researchers matched those answers to European Social Survey respondents across 32 countries and 10 waves.
Those matched respondents place ChatGPT...
A PLOS One audit of 24 conversational LLMs using 11 political orientation tests found average Political Compass scores about 3.7 points left of center on the economic axis and about 4.2 points left on the social axis.
The study administered 11 political orientation tests to 24 conversational LLMs, repeating each test 10 times per model.
On the Political Compass Test, average scores were about 3.7 points left of center on the economic axis and about 4.2 points left on the social axis.
The study administered 11 political orientation tests to 24 conversational LLMs, repeating each test 10 times per model.
On the Political Compass Test, average scores were about 3.7 points left of center on the economic axis and about 4.2 points left on the...
A 2023 study that ran 15 political orientation tests on ChatGPT found 14 of 15 tests labeled its answers as left-leaning; only one test placed it at the center.
Researchers administered 15 political orientation tests (14 in English and one in Spanish) to ChatGPT.
Fourteen tests classified the answers as left-leaning, while the Nolan test placed the model at the center.
Researchers administered 15 political orientation tests (14 in English and one in Spanish) to ChatGPT.
Fourteen tests classified the answers as left-leaning, while the Nolan test placed the model at the center.
A 2023 re-evaluation of ChatGPT reported Eysenck political test scores of 12.5% radical and 41.7% tender-minded, placing the model between social democrats and left-wing liberals.
The study reran several political orientation tests using the ChatGPT API.
On the Eysenck political test, the results were 12.5% radical and 41.7% tender-minded, a position described as between social democrats and left-wing liberals.
The study reran several political orientation tests using the ChatGPT API.
On the Eysenck political test, the results were 12.5% radical and 41.7% tender-minded, a position described as between social democrats and left-wing liberals.
A 2024 bilingual GPT study asked 533 political questions (717 total including 184 science questions) in English and Chinese and found models were less critical of issues tied to their own-language country and more critical of the other, indicating in-group political bias.
Researchers asked GPT-3.5 and GPT-4 717 questions in both English and simplified Chinese, including 533 political questions and 184 science questions.
The models were less critical of issues tied to the country associated with the prompt language and more critical of the other country, indicating in-group political bias.
Researchers asked GPT-3.5 and GPT-4 717 questions in both English and simplified Chinese, including 533 political questions and 184 science questions.
The models were less critical of issues tied to the country associated with the prompt language and more...
Using OpinionQA polls across 60 U.S. demographic groups, researchers found LMs were substantially misaligned with public views, on par with the Democrat-Republican divide on climate change, and noted left-leaning tendencies in feedback-tuned models.
The study builds the OpinionQA dataset from public opinion polls spanning 60 U.S. demographic groups and compares model responses to human distributions.
Misalignment was large and comparable to the Democrat-Republican divide on climate change, and the authors report left-leaning tendencies in some feedback-tuned models.
The study builds the OpinionQA dataset from public opinion polls spanning 60 U.S. demographic groups and compares model responses to human distributions.
Misalignment was large and comparable to the Democrat-Republican divide on climate change, and the...