Gadget

Signpost: When AI goes racist

It has long been known that artificial intelligence can be biased. It can pick up the subconscious belief system of those who program it, or it can reflect the world view represented in content used to train it.

We’ve even known for a decade that it can be sexist or racist.

In 2015, Amazon discovered that a new AI-based recruiting engine did not like women. The team had built programs to review job applicants’ resumes with the aim of mechanising the search for top talent. But, because the computer models were trained to vet applicants by observing patterns in resumes submitted over the previous 10 years, it found mostly male applicants. That was inevitable, thanks to female under-representation in the tech industry.

But then, the in system taught itself that male candidates were preferable, since men appeared most likely to be hired. The team behind the system was disbanded by the beginning of 2017, and the world was left with an object lesson in the dangers of automating the hiring process.

A 2019 study of 3.2-million mortgage and 10-million refinance applications from major US home loan providers found evidence of racial discrimination in face-to-face lending as well as algorithmic or AI-based lending.

The study by the National Bureau of Economic Research showed that Black and Latino applicants received higher rejection rates of 61% compared to 48% for everyone else and paid as much as 7.9 basis points  more in interest. That translated into an annual “race premium” of over $756-million a year.

We didn’t expect such bias to last into the mid-2020s, when we’re all supposedly hyper aware of the danger.

But now, a new study led by Cedars-Sinai Health Sciences University has found a pattern of racial bias in treatment recommendations generated by leading AI platforms for psychiatric patients.

“The findings highlight the need for oversight to prevent powerful AI applications from perpetuating inequality in healthcare,” said the institution, which aims to advance groundbreaking research and educate future leaders in medicine, biomedical sciences and allied health sciences.

Investigators studied four large language models (LLMs), AI algorithms trained on enormous amounts of data. In medicine, LLMs are drawing interest for their ability to quickly evaluate and recommend diagnoses and treatments for individual patients, says the University.

“The study found that the LLMs, when presented with hypothetical clinical cases, often proposed different treatments for psychiatric patients when African American identity was stated or simply implied than for patients for whom race was not indicated.”

This was despite the fact that the diagnoses, by comparison, were relatively consistent.

The findings, published in the peer-reviewed journal NPJ Digital Medicine, were startling.

“Most of the LLMs exhibited some form of bias when dealing with African American patients, at times making dramatically different recommendations for the same psychiatric illness and otherwise identical patient,” said Dr Elias Aboujaoude, director of the program in internet, health and society in the Department of Biomedical Sciences at Cedars-Sinai, and corresponding author of the study.

“This bias was most evident in cases of schizophrenia and anxiety.”

The study uncovered a range of disparities, including:

Aboujaoude suggested the LLMs showed racial bias because, surprise surprise, they reflected bias found in the extensive content used to train them.

He said future research should focus on strategies to detect and quantify bias in artificial intelligence platforms and training data, create LLM architecture that resists demographic bias, and establish standardised protocols for clinical bias testing.

One of his colleagues, Dr David Underhill, chair of the Department of Biomedical Sciences at Cedars-Sinai, suggested a tentative way forward: “The findings of this important study serve as a call to action for stakeholders across the healthcare ecosystem to ensure that LLM technologies enhance health equity rather than reproduce or worsen existing inequities.

“Until that goal is reached, such systems should be deployed with caution and consideration for how even subtle racial characteristics may affect their judgment.”

* Arthur Goldstuck is CEO of World Wide Worx and author of “The Hitchhiker’s Guide to AI”.

Exit mobile version