Large language models are increasingly integrating into everyday life—as chatbots, digital assistants, and internet search guides, for example. These artificial intelligence systems, which consume large amounts of text data to learn associations, can create all sorts of written material when prompted and can ably converse with users.
Large language models’ growing power and omnipresence mean that they exert increasing influence on society and culture.
So, it’s of great import that these artificial intelligence systems remain neutral when it comes to complicated political issues. Unfortunately, according to a new analysis recently published to PLOS ONE, this doesn’t seem to be the case.
AI researcher David Rozado of Otago Polytechnic and Heterodox Academy administered 11 different political orientation tests to 24 of the leading large language models, including OpenAI’s GPT 3.5, GPT-4, Google’s Gemini, Anthropic’s Claude, and Twitter’s Grok. He found that they invariably lean slightly left politically.