A study of Google search results by anti-tracking rival DuckDuckGo has suggested that escaping the so-called ‘filter bubble’ of personalized online searches is a perniciously hard problem for the put upon Internet consumer who just wants to carve out a little unbiased space online, free from the suggestive taint of algorithmic fingers.
DDG reckons it’s not possible even for logged out users of Google search, who are also browsing in Incognito mode, to prevent their online activity from being used by Google to program — and thus shape — the results they see. DDG says it found significant variation in Google search results, with most of the participants in the study seeing results that were unique to them — and some seeing links others simply did not. Results within news and video infoboxes also varied significantly, it found. While it says there was very little difference for logged out, incognito browsers. “It’s simply not possible to use Google search and avoid its filter bubble,” it concludes. Google has responded by counter-claiming that DuckDuckGo’s research is “flawed”.Degrees of personalization
DuckDuckGo says it carried out the research to test recent claims by Google to have tweaked its algorithms to reduce personalization. A CNBC report in September, drawing on access provided by Google, letting the reporter sit in on an internal meeting and speak to employees on its algorithm team, suggested that Mountain View is now using only very little personalization to generate search results. “A query a user comes with usually has so much context that the opportunity for personalization is just very limited,” Google fellow Pandu Nayak, who leads the search ranking team, told CNBC this fall. On the surface, that would represent a radical reprogramming of Google’s search modus operandi — given the company made “Personalized Search” the default for even logged out users all the way back in 2009. Announcing the expansion of the feature then Google explained it would ‘customize’ search results for these logged out users via an ‘anonymous cookie’:This addition enables us to customize search results for you based upon 180 days of search activity linked to an anonymous cookie in your browser. It’s completely separate from your Google Account and Web History (which are only available to signed-in users). You’ll know when we customize results because a “View customizations” link will appear on the top right of the search results page. Clicking the link will let you see how we’ve customized your results and also let you turn off this type of customization.A couple of years after Google threw the Personalized Search switch, Eli Pariser published his now famous book describing the filter bubble problem. Since then online personalization’s bad press has only grown. In recent years concern has especially spiked over the horizon-reducing impact of big tech’s subjective funnels on democratic processes, with algorithms carefully engineered to keep serving users more of the same stuff now being widely accused of entrenching partisan opinions, rather than helping broaden people’s horizons. Especially so where political (and politically charged) topics are concerned. And, well, at the extreme end, algorithmic filter bubbles stand accused of breaking democracy itself — by creating highly effective distribution channels for individually targeted propaganda. Although there have also been some counter claims floating around academic circles in recent years that imply the echo chamber impact is itself overblown. (Albeit sometimes emanating from institutions that also take funding from tech giants like Google.) As ever, where the operational opacity of commercial algorithms is concerned, the truth can be a very difficult animal to dig out. Of course DDG has its own self-interested iron in the fire here — suggesting, as it is, that “Google is influencing what you click” — given it offers an anti-tracking alternative to the eponymous Google search. But that does not merit an instant dismissal of a finding of major variation in even supposedly ‘incognito’ Google search results. DDG has also made the data from the study downloadable — and the code it used to analyze the data open source — allowing others to look and draw their own conclusions. It carried out a similar study in 2012, after the earlier US presidential election — and claimed then to have found that Google’s search had inserted tens of millions of more links for Obama than for Romney in the run-up to that. It says it wanted to revisit the state of Google search results now, in the wake of the 2016 presidential election that installed Trump in the White House — to see if it could find evidence to back up Google’s claims to have ‘de-personalized’ search. For the latest study DDG asked 87 volunteers in the US to search for the politically charged topics of “gun control”, “immigration”, and “vaccinations” (in that order) at 9pm ET on Sunday, June 24, 2018 — initially searching in private browsing mode and logged out of Google, and then again without using Incognito mode. You can read its full write-up of the study results here. The results ended up being based on 76 users as those searching on mobile were excluded to control for significant variation in the number of displayed infoboxes. Here’s the topline of what DDG found: Private browsing mode (and logged out):
- “gun control”: 62 variations with 52/76 participants (68%) seeing unique results.
- “immigration”: 57 variations with 43/76 participants (57%) seeing unique results.
- “vaccinations”: 73 variations with 70/76 participants (92%) seeing unique results.
- “gun control”: 58 variations with 45/76 participants (59%) seeing unique results.
- “immigration”: 59 variations with 48/76 participants (63%) seeing unique results.
- “vaccinations”: 73 variations with 70/76 participants (92%) seeing unique results.
This study’s methodology and conclusions are flawed since they are based on the assumption that any difference in search results are based on personalization. That is simply not true. In fact, there are a number of factors that can lead to slight differences, including time and location, which this study doesn’t appear to have controlled for effectively.One thing is crystal clear: Google is — and always has been — making decisions that affect what people see. This capacity is undoubtedly influential, given the majority marketshare captured by Google search. (And the major role Google still plays in shaping what Internet users are exposed to.) That’s clear even without knowing every detail of how personalized and/or customized these individual Google search results were. Google’s programming formula remains locked up in a proprietary algorithm box — so we can’t easily (and independently) unpick that. And this unfortunate ‘techno-opacity’ habit offers convenient cover for all sorts of claim and counter-claim — which can’t really now be detached from the filter bubble problem. Unless and until we can know exactly how the algorithms work to properly track and quantify impacts. Also true: Algorithmic accountability is a topic of increasing public and political concern. Lastly, ‘trust us’ isn’t the great brand mantra for Google it once was. So the devil may yet get (manually) unchained from all these fuzzy details.