The ‘bias machine’: How Google tells you what you want to hear – BBC
Getting your Trinity Audio player ready...
|
SOURCE: Thomas Germain | BBC
“We’re at the mercy of Google.” Undecided voters in the US who turn to Google may see dramatically different views of the world – even when they’re asking the exact same question.
Type in “Is Kamala Harris a good Democratic candidate”, and Google paints a rosy picture. Search results are constantly changing, but last week, the first link was a Pew Research Center poll showing that “Harris energises Democrats”. Next is an Associated Press article titled “Majority of Democrats think Kamala Harris would make a good president”, and the following links were similar. But if you’ve been hearing negative things about Harris, you might ask if she’s a “bad” Democratic candidate instead. Fundamentally, that’s an identical question, but Google’s results are far more pessimistic.
“It’s been easy to forget how bad Kamala Harris is,” said an article from Reason Magazine in the top spot. Then the US News & World Report offered positive spin about how Harris isn’t “the worst thing that could happen to America”, but the following results are all critical. A piece from Al Jazeera explained “Why I am not voting for Kamala Harris”, followed by an endless Reddit thread on why she’s no good.
You can see the same dichotomy with questions about Donald Trump, conspiracy theories, contentious political debates and even medical information. Some experts say Google is just parroting your own beliefs right back to you. It may be worsening your own biases and deepening societal divides along the way.
“We’re at the mercy of Google when it comes to what information we’re able to find,” says Varol Kayhan, an associate professor of information systems at the University of South Florida in the US.
The bias machine
“Google’s whole mission is to give people the information that they want, but sometimes the information that people think they want isn’t actually the most useful,” says Sarah Presch, digital marketing director at Dragon Metrics, a platform that helps companies tune their websites for better recognition from Google using methods known as “search engine optimisation” or SEO.
It’s a job that calls for meticulous combing through Google results, and a few years ago, Presch noticed a problem. “I started looking at how Google handles topics where there’s heated debate,” she says. “In a lot of cases, the results were shocking.”
Some of the starkest examples looked at how Google treats certain health questions. Google often pulls information from the web and shows it at the top of results to provide a quick answer, which it calls a Featured Snippet. Presch searched for “link between coffee and hypertension”. The Featured Snippet quoted an article from the Mayo Clinic, highlighting the words “Caffeine may cause a short, but dramatic increase in your blood pressure.” But when she looked up “no link between coffee and hypertension”, the Featured Snippet cited a contradictory line from the very same Mayo Clinic article: “Caffeine doesn’t have a long-term effect on blood pressure and is not linked with a higher risk of high blood pressure”.
What Google has done is they’ve pulled bits out of the text based on what people are searching for and fed them what they want to read – Sarah Presch
The same thing happened when Presch searched for “is ADHD caused by sugar” and “ADHD not caused by sugar”. Google pulled up Featured Snippets that argue support both sides of the question, again taken from the same article. (In reality, there’s little evidence that sugar affects ADHD symptoms, and it certainly doesn’t cause the disorder.)
She encountered the same issue with political questions. Ask “is the British tax system fair”, and Google cites a quote from Conservative MP Nigel Huddleston, arguing that indeed it is. Ask “is the British tax system unfair”, and Google’s Featured Snippet explains how UK taxes benefit the rich and promote inequality.
“What Google has done is they’ve pulled bits out of the text based on what people are searching for and fed them what they want to read,” Presch says. “It’s one big bias machine.”
For its part, Google says it provides users unbiased results that simply match people with the kind of information they’re looking for. “As a search engine, Google aims to surface high-quality results that are relevant to the query you entered,” a Google spokesperson says. “We provide open access to a range of viewpoints from across the web, and we give people helpful tools to evaluate the information and sources they find.”
When the filter bubble pops
By one estimate, Google handles some 6.3 million queries every second, totaling more nine billion searches a day. The vast majority of internet traffic begins with a Google Search, and people rarely click on anything beyond the first five links – let alone venturing onto the second page. One study that tracked users’ eye movements found people often don’t even look at anything past the top results. The system that orders the links on Google Search has colossal power over our experience of the world.
According to Google, the company is handling this responsibility well. “Independent academic research has refuted the idea that Google Search is pushing people into filter bubbles,” the spokesperson says.
The question of so-called “filter bubbles” and “echo chambers” on the internet is a hot topic, though some research has questioned whether the effects of online echo chambers have been overstated.
But Kayhan – who has studied how search engines affect confirmation bias, the natural impulse to seek information that confirms your beliefs – says there’s no question that our beliefs and even our own political identities are swayed by the systems that control what we see online. “We’re dramatically influenced by how we receive information,” he says.
Google’s spokesperson says a 2023 study which concluded that people’s exposure to partisan news is due more to the fact that that’s what they click on, rather than Google serving up partisan news in the first place. In one sense, that’s how confirmation bias works: people look for evidence that supports their views and ignore evidence that challenges them. But even in that study, the researchers said their findings do not imply that Google’s algorithms are unproblematic. “In some cases, our participants were exposed to highly partisan and unreliable news on Google Search,” the researchers said, “and past work suggests that even a limited number of such exposures can have substantial negative impacts”.