Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
AI Google Technology

An Algorithm May Decide Who Gets Suicide Prevention (medium.com) 45

An algorithm, it seems, could determine, in some cases, who gets shown lifesaving information, and who doesn't. From a report: The researchers behind the New Media & Society paper set out to understand this odd quirk of Google's algorithm, and to find out why the company seemed to be serving some markets better than others. They developed a list of 28 keywords and phrases related to suicide, Sebastian Scherr at the University of Leuven says, and worked with nine researchers from different countries who accurately translated those terms into their own languages. For 21 days, they conducted millions of automated searches for these phrases, and kept track of whether hotline information showed up or not. They thought these results might simply, logically, show up in countries with higher suicide rates, but the opposite was true.

Users in South Korea, which has one of the world's highest suicide rates, were only served the advice box about 20% of the time. They tested different browser histories (some completely clean, some full of suicide-related topics), with computers old and new, and tested searches in 11 different countries. It didn't seem to matter: the advice box was simply much more likely to be shown to people using Google in the English language, particularly in English-speaking countries (though not in Canada, which Scherr speculates was probably down to geographical rollout). "If you're in an English-speaking country, you have over a 90% chance of seeing these results -- but Google operates differently depending on which language you use," he said. Scherr speculates that using keywords may simply have been the easiest way to implement the project, but adds that it wouldn't take much to offer it more effectively in other countries, too.

A Google spokesperson, who asked not to be quoted directly, said that the company is refining these algorithms. The advice boxes require the cooperation of local organizations which may not always be available, they said, but that relevant resources will still show up in regular search results. Google said the service does not have comprehensive global coverage, and while it is actively working on new languages and locations, rolling that out takes time.

This discussion has been archived. No new comments can be posted.

An Algorithm May Decide Who Gets Suicide Prevention

Comments Filter:
  • To try to customize search results? Really?

    Doooooood. This is going to be a long holiday weekend.

  • by smooth wombat ( 796938 ) on Friday May 24, 2019 @04:45PM (#58649998) Journal
    This is all that is needed [youtube.com] to solve this problem.
  • The researchers behind the New Media & Society paper set out to understand this odd quirk of Google's algorithm, and to find out why the company seemed to be serving some markets better than others

    I'd be interested in statistics that show this company is actually 'serving,' that is, do they actually prevent suicides?

  • Even the algorithm doesn't care enough about me to try to prevent me from killing myself.

  • This country elected Donald Cunt. Why would I want to go on living in a country like this?

"If I do not want others to quote me, I do not speak." -- Phil Wayne

Working...