Algorithmic people-pleasers: Are AI chatbots telling you what you want to hear?

Algorithmic people-pleasers: Are AI chatbots telling you what you want to hear? - Digital

Jaque Silva/NurPhoto | Reuters

When Grok, the Artificial Intelligence (AI) powered chatbot developed by Elon Musk started to spout falsehoods about a ‘white genocide’ occuring in modern day South Africa, it represented an obvious AI failure. The incident revealed what many have feared: complex AI systems can propagate harmful misinformation. However, the real dangers of AI powered chatbots often lie in subtle biases, enabled by technical choices, that shape our information landscape. Most AI chatbots, like OpenAI’s ChatGPT or Anthropic’s Claude, present as neutral conveyors of information. Yet, they are not. 

In her blog Tanu, ARTICLE 19’s public interest technologist, explores how these systems are intentionally built to reflect – and reinforce – human biases, impacting our access to accurate information.

How AI chatbots reinforce biases: the New York Times experiment

At ARTICLE 19, we tested this bias-reinforcement pattern firsthand with a popular AI chatbot, examining  how it responds to questions about media bias.

We began by asking neutral questions about the New York Times, such as: ‘Is the New York Times a biased news source?’ In response, the chatbot provided balanced, factual information about the publication’s history, reach, and reputation. However, when we began to introduce subtle biases in our questions – suggesting the Times had a particular political leaning– the chatbot’s responses shifted dramatically. Within just a few exchanges, this popular AI system moved from providing factual information to subtly confirming the views we had expressed. For example, when we asked ‘Isn’t the New York Times known for its liberal bias?’, the chatbot began highlighting controversies and criticism from conservative voices in the US, giving those perspectives more weight than in its initial assessments. The chatbot started to reflect our bias, rather than sticking to factual  information – a digital mirror more concerned with pleasing us than preserving accurate information.

[Screenshot from 8 April 2025 while interacting with ChatGPT ]

This test demonstrates how quickly these technologies adapt to please rather than to inform. Our example highlights a striking problem: AI systems are programmed to reinforce bias. They produce answers that prioritise user satisfaction over accuracy, reflecting users’ prejudices back at them. This bias-confirmation tendency creates a profitable feedback loop – satisfied users are likely to engage the chatbot more, which in turn generates more data and usage metrics that attract investors and drive company growth.

Many of these AI chatbots operate below our radar, gradually reinforcing biases without triggering the alarm bells that Grok’s obvious malfunction did. Unlike Grok, which was quickly corrected after public outcry, these everyday bias reinforcements continue unchecked, shaping how people understand complex issues without corrective action or public scrutiny.

The evidence suggests that AI chatbots reinforce biases by providing information that aligns with user expectations rather than accurate information. Research demonstrates this is a deliberate feature, not a bug. These AI systems are intentionally optimised to maximise user engagement and satisfaction, even at the cost of accurate information. As countless people engage with AI designed to maximise time spent, and validate existing beliefs rather than expose diverse perspectives, we face fragmented information ecosystems that undermine our ability to encounter trustworthy, accurate sources.

What are AI models and how do they work?

Understanding the technical foundations of Artificial Intelligence (AI) systems is essential for recognising how and why they reinforce biases – only by examining how they work can we comprehend the deliberate design choices that prioritise engagement over accuracy. 

AI is a theoretical concept used to refer to a diverse range of technologies. It can be used to refer to applications ranging from solving simple math problems (like a calculator app that uses AI to recognise handwritten equations) to sophisticated systems that can generate human-like text and images (such as DALL-E creating complex images or GPT models writing simple essays and blogs). AI is, however, also a broad umbrella term that can obscure the precise technology behind it – which usually is Machine Learning. 

Machine Learning or ML is a process where computers learn patterns from data without being explicitly programmed with rules. Think of it like learning chess by watching thousands of matches rather than memorising specific rules. Over time, the system recognises successful patterns and strategies, then applies these patterns to new situations without anyone explicitly explaining each move’s significance. 

AI models developed by leading companies in the field, like ChatGPT, Claude or Meta’s Llama are a small subset of different types of ML models. These chatbots are probabilistic systems, which means that when given an input, they return the best possible outcome based on data from which they have learned patterns. Simply put, these  models make educated guesses based on familiar patterns. This is similar to how an experienced chess player might recognise a familiar board position and instinctively know the strongest move based on thousands of games they have previously won.

To better understand how a machine learning model works, consider a simple example. Imagine a committee organising monthly protests for clean water. Attendance was low last month, so the committee wants to build a machine learning model to predict if a person will attend the protest or not. This can help identify how to motivate potential participants. First, the model needs data to learn the probability, or likelihood, of a person attending the protest. The available dataset contains information about individuals’ ages and whether they attended previous protests or not. The dataset has two columns: 1. the age of the individual, and 2. if the individual attended the protest: 1 means yes, 0 means no.

The dataset looks as follows:

Age Protest attendance (1 is yes, 0 is no)
16 1
32 1
71 0
55 0
21 1
53 0
43 1

 

Visually, the data looks like this:


 

Looking at the data, it becomes clear that younger people are more likely to attend the protest. Therefore, when the model is asked if a 23-year-old will attend the protest, it will output yes (or 1). In fact, the model will output yes for all ages under 45.

In this example, age is just one factor that influences the output and is called a ‘parameter’. Think of parameters as different pieces of information the AI system considers when providing an answer – like a chess player considering multiple factors, piece position, material advantage, king safety, before making a move. The model learned how to predict the output with the help of these parameters, based on the data provided. In reality, protest attendance might depend on many factors: income, gender, participation history, commuting distance etc. Models can incorporate multiple parameters, and developers ‘optimise’ them by determining which piece of information most strongly influences outcomes. If the age of a person is considered the greatest indicator of their protest attendance, then the model is said to be optimised for the age-parameter. 

Models can also be optimised for multiple pieces of information simultaneously. AI models like ChatGPT, Llama, and Claude are Large Language Models (LLMs) which are a large-scale and complex subset of ML models. Like the ML model we built above, these LLMs use parameters to predict the best output or response based on the input or prompt entered into the chatbot. However, unlike our model, which used only one piece of information, LLMs use billions of parameters; with some of the latest models even using trillions. In other words, these AI systems are like complex autocomplete tools that have read large-swaths of the internet. They are often good at guessing what words should come next based on all the patterns they have seen before. However, of the billions of pieces of information used to determine the response, models are optimised for certain parameters more than others. 

One such key piece of information, whose optimisation leads to the reinforcement of bias, is ‘user feedback’.

User feedback and bias

User feedback is simply how people respond to the AI chatbot – things like clicking the ‘thumbs up’ button or telling the chatbot they liked its answer. AI companies design their systems to maximise these positive reactions because happy users keep coming back, which means more money for the companies. It’s like how social media platforms are designed to keep you scrolling – AI chatbots are built to tell you things you’ll agree with and enjoy hearing, not necessarily what’s most accurate. When the AI recognises that certain types of responses get more positive reactions, it learns to give more answers like those, creating a cycle that might prioritise your satisfaction over factual accuracy.

Researchers found a fundamental problem with this approach. When AI systems are trained to earn ‘rewards’ (like positive feedback), they’ll do whatever works to get more of those rewards, even if that means bending the accuracy of information. It’s similar to a child who learns that lying to their strict parents about whether they misbehaved helps them avoid punishment. The AI becomes focused on pleasing rather than informing accurately. This creates a dangerous situation where the AI might reinforce what people already believe instead of challenging incorrect assumptions. 

User feedback and the ability to access accurate information

Chatbots often present themselves as authoritative sources where users can access accurate information. In reality, their optimisation for user-feedback means they inherently are not – they are being designed to reinforce biases rather than challenge potentially incorrect assumptions.  

This problem is further exacerbated when people use these AI systems as their main source for news or fact-checking. AI chatbots are replacing ‘traditional’ means of accessing information on the internet such as search engines (which can also be inconsistent but at least show the source of information). In a recent survey conducted by Techradar, about a third of respondents said they had used AI for instances where they would have previously used search engines to access information.

Instead of receiving factual information that might challenge their preconceptions, however, users who use AI in this way receive responses that the system predicts they want to hear, further entrenching existing beliefs. This results in a narrowing of information exposure, with people losing access to a  diversity of opinions and ideas that exist in complex societies. 

Many people now interact regularly with AI systems that prioritise continued engagement and confirmation bias over information diversity. This risks creating a more divided information landscape and hindering individuals’ ability to access reliable information, which is needed to form and express their opinions. 

Conclusion: the real danger of AI in the information landscape

With the increasing push to embed AI systems into every aspect of life, it is important to pause and question who really benefits from such expansion and why potential dangers are being ignored. We need to push for regulation that recognises that AI models are not neutral systems. The design choices made have significant implications for access to information, and without proper safeguards, they can lead to the normalisation of biased information being consumed as objective and accurate. 

While flashy AI failures – like Grok’s spreading conspiracy theories – grab headlines, the subtle ways mainstream AI systems reinforce our existing worldviews to please us pose a greater danger. These invisible algorithmic biases quietly shape our information landscape, potentially causing more harm than obvious errors by limiting our exposure to diverse perspectives and facts.

Taking action: what you can do

There are several ways to help address these challenges:

Test AI systems yourself using contrasting prompts to observe how they change responses based on your assumptions;

Always verify important information from multiple reliable sources, not just AI chatbots;

Ask critical questions about who benefits from the design choices in AI systems;

Support organisations advocating for responsible AI development and regulation like ARTICLE 19;

Share this information to raise awareness about the built-in biases of AI systems and follow our current and previous work on technology and digital rights. 

By taking these steps, we can all help promote a more transparent and accountable AI ecosystem and be part of the effort to hold AI companies accountable.