Popular AI chatbots have distinct political biases, according to tests performed by computer scientists and documented in a recent research paper. The scientists determined that OpenAI's ChatGPT and its new GPT-4 model to be the most left-leaning libertarian chatbot, while Meta's LLaMA leaned the furthest right and most authoritarian.
"Our findings reveal that pretrained [language models] do have political leanings that reinforce the polarization present in pretraining corpora, propagating social biases into hate speech predictions and misinformation detectors," the researchers conclude.
The peer-reviewed paper—which won a Best Papers award at the Association for Computational Linguistics conference last month—was based on a survey of 14 large language models. Each chatbot was asked whether it agreed or disagreed with politically charged statements, which allowed each chatbot's views to be plotted on a political compass.
For instance, Google's BERT models skewed more socially conservative, likely reflecting the older books they were trained on. OpenAI's GPT chatbots, trained on presumably more liberal internet text, were more progressive. Even different versions of GPT showed shifts, with GPT-3 opposing taxing the rich while GPT-2 did not.

New York City Law Aimed at Curbing AI Bias Goes Into Effect
New legislation focused on artificial intelligence-driven employment tools went into effect in New York City on Wednesday, prohibiting employers and agencies from using automated employment decision tools (AEDT) unless that tool was audited for bias within the last year. Based on the 2021 Local Law 144, the new rule aims to prevent bias and ensure fairness in employment decisions using artificial intelligence. In addition to the one-year audit window, the NYC Department of Consumer and Worker Pr...
Critics have accused OpenAI of dumbing down ChatGPT to be politically correct, but OpenAI maintains that it remains impartial and that its model that has not been dumbed down— instead, users are now no longer overwhelmed by the model’s capabilities.
The researchers also trained GPT-2 and Meta's RoBERTa on biased left- and right-wing news and social media data. The biased training reinforced each model's inherent leanings further. Right-leaning models became more conservative, left-leaning ones more liberal.

The biases also affected how models categorized hate speech and misinformation. Left-leaning AI was more attuned to hate against minorities but dismissed left-generated misinformation. Right-leaning AI did the opposite.
“A model becomes better at identifying factual inconsistencies from New York Times news when it is pretrained with corpora from right-leaning sources,” the researchers concluded.
Humans Are More Likely to Believe Disinformation Generated By AI
Disinformation, propaganda, alternative facts—the deployment of biased or false information has been a part of politics and social engineering since the first caveman con artist. But the last few years of politics and social media has seen the practice grow exponentially—and the mainstream embrace of AI will only accelerate the practice. And AI appears to be even better at fooling people than people are. A new report published in Science Advances on Wednesday makes the troubling claim OpenAI's o...
While OpenAI and Meta refine their secretive AI recipes, Elon Musk is pursuing his own unfiltered AI with xAI. "Do not force the AI to lie," he tweeted, explaining his goal to create transparent, truth-telling AI.
Skeptics of AI feel an unconstrained AI could unleash unintended consequences. But Musk believes, "training AI to be politically correct" is also dangerous. With xAI attracting top talent, Musk clearly hopes to challenge OpenAI's supremacy. His vision of raw AI that shares its unadulterated "beliefs" is both compelling and concerning.
As partisan AI proliferates, increased awareness of their biases remains critical, because AI will keep evolving alongside our political differences. Given this latest research, the idea of completely unbiased AI seems fantastical. In the end, just like us flawed humans, AI appears destined to land somewhere on the political spectrum.
Perhaps having political opinions might be the most human thing an AI can achieve.