by Greg Bruno, Rutgers University
Credit: CC0 Public Domain
Dr. Google is always in, but for Spanish speakers searching for health information online, the advice Google gives isn’t always helpful. Sometimes, it’s downright terrifying.
Research by Rutgers public health and information science experts found that during the height of the COVID-19 pandemic, Google search autocompletes—what the technology company calls “predictions”—returned different results in Spanish than in English. In many cases, the suggested Spanish search terms were more fear- and stress-inducing than the English equivalent.
For instance, the researchers found that typing “coronavirus is” into Google yielded search predictions including “it contagious,” “man made,” and “airborne,” while in Spanish, “coronavirus es” prompted “mortal,” “falso” (fake) and “peligroso” (dangerous).
“The health consequences of this data are profound,” said Pamela Valera, an assistant professor in the Department of Urban-Global Public Health at the Rutgers School of Public Health and lead author of the study published in the Journal of Community Psychology.
“This is one of the first studies to explore the differences in autocompletes and search quality across these two languages,” added Vivek K. Singh, an associate professor at the Rutgers School of Communication and Information and a study co-author.
Every day, billions of internet searches are powered by Google, and autocomplete predictions help deliver results faster by suggesting queries. Language, location, trending interest and past searches influence what is displayed, according to the company.
But predictive algorithms carry inherent bias, which can impact public health outcomes and behaviors, the researchers said. One example from the study: Researchers found English autocompletes for the term “hand sanitizer” returned mostly positive predictions, but in Spanish, the predictions were negative, potentially influencing the use of disinfectants among Spanish speakers.
To explore language differences in Google search autocompletes during the first wave of the pandemic, Valera, Singh and their colleagues conducted three focus groups with 29 participants.
The first meeting was in English with non‐Spanish speakers; the second with Spanish speakers only; and the third with bilingual and bicultural participants. Virtual meetings lasted between 90 and 120 minutes.
Several themes emerged from the conversations. First, autocompletes evoked fear and stress among Spanish speakers and skepticism and hesitation toward autocomplete searches themselves, Valera said.
The researchers also found that autocompletes promote search preselection, potentially leading people toward biased information. Finally, the study found that English autocompletes for COVID-19-related terms far outnumbered those in Spanish—three-to-one in some cases.
This study was part of a larger research project that focused on understanding algorithm bias on Google. Phase I of the study cataloged Google search autocompletes for COVID-19 topics between March 2020 and September 2020; that paper is awaiting publication. The current study reflects the qualitative results from Phase II.
With Google being such an important tool for obtaining health-related information, the researchers said technology companies should do more to ensure that search suggestions are consistent across languages.
“Arguably, more web content exists in English, but there are opportunities for Google to use the wealth of knowledge that exists in all languages to create products and autocomplete suggestions that are more comprehensive and more equitable across languages,” said Singh. “This would be good for the quality of their results and for anyone who Googles health-related information.”
Leave a Reply