女生小视频

Technology

A biased algorithm is delaying healthcare for black people in the US

By Jessica Hamzelou

24 October 2019

woman receiving medical care

A medical algorithm used in the US is prioritising the treatment of healthier white people over sicker black people

Joe Raedle/Getty Images

Black people in the US may be missing out on healthcare because a widely used algorithm is racially biased. The proportion of black people referred for extra care would more than double if the bias were removed, according to new research.

Algorithms are fast becoming a key part of healthcare. Such technologies are used to screen somewhere between 100 and 200 million people in the US, says at the University of California, Berkeley.

One example is an algorithm that is used to predict the future health of individuals based on their past health records. Once the algorithm is fed data about a person鈥檚 diagnoses, prescriptions and procedures, it spits out a number that predicts the cost of the person鈥檚 future healthcare.

Hospitals, healthcare systems and some health insurance providers use this score to identify people who are likely to need more care in the future. Those who have the highest predicted costs can be referred for extra medical care to help prevent them getting sicker, says Obermeyer.

Obermeyer and his colleagues wanted to find out more about how the algorithm worked. 鈥淎long the way, we noticed that there was this fairly stark difference in the risk scores that black and white patients had at the same level of health,鈥 he says.

Ahead in line

The team found that black people assigned the same score as white people went on to have worse health outcomes. 鈥淵ou can think of it as healthier white patients being put ahead in line of sicker black patients when it comes to allocating enrolment into this programme,鈥 says Obermeyer.

When the team ran a simulation that eliminated the bias, the proportion of people referred for extra treatment who were black increased from 17.7 per cent to 46.5 per cent. It is difficult to estimate how many black people are missing out on healthcare as a result, says Obermeyer, because he doesn鈥檛 know how many health organisations use the algorithm in this way.

This doesn鈥檛 mean that the algorithm isn鈥檛 working, says Obermeyer. 鈥淭he data are just a reflection of the society that produced the data,鈥 he says. 鈥Black patients will generate lower costs than white patients, and that鈥檚 because of a variety of socioeconomic factors related to access to healthcare, as well as direct effects of race on the doctor-patient relationship.鈥

鈥淭he design of different systems, whether we鈥檙e talking about legal systems or computer systems, can create and reinforce hierarchies precisely because the people who create them are not thinking about how social norms and structures shape their work,鈥 says at Princeton University, and author of Race After Technology. 鈥淚ndifference to social reality is, perhaps, more dangerous than outright bigotry.鈥

Significant and prevalent

Bias in algorithms is likely to be 鈥渟ignificant and prevalent鈥, says at Stanford University in California. 鈥淚t鈥檚 hard to put a number on it, but this could affect a large fraction of the [US] population,鈥 he says.

Obermeyer and his colleagues have tweaked the algorithm to predict other variables that are less racially biased, and have been working with the manufacturer of the algorithm to improve it.

The researchers have chosen not to name the manufacturer because they don鈥檛 want to single out one company. Many other companies and academics have developed similar algorithms, and none of them have realised they may be biased, says Obermeyer.

The hospitals, healthsystems and government agencies using the algorithms haven鈥檛 either. 鈥淚t鈥檚 tempting to think they should have known better, but nobody knew better,鈥 says Obermeyer. Regulations that require manufacturers to audit their algorithms might help in future, but users must also take responsibility for making sure the algorithms they use aren’t biased, he says.

In the meantime,聽there are some signs of progress, says聽, ethics fellow at the Alan Turing Institute in London. The organisation has helped create guidance on the responsible and ethical design and implementation of AI systems,聽.

Science

Topics:

Sign up to our weekly newsletter

Receive a weekly dose of discovery in your inbox. We'll also keep you up to date with New 女生小视频 events and special offers.

Sign up
Piano Exit Overlay Banner Mobile Piano Exit Overlay Banner Desktop