Research: Highlight Human Bias to Boost AI Acceptance

Lehigh University

While people are growing more accustomed to AI-driven personal assistants, customer service chatbots and even financial advisors, when it comes to healthcare, most still want it with a human touch.

Given that receiving healthcare is a deeply personal experience, it's understandable that patients prefer it to come from, well, a person. But with AI's vast potential to increase the quality, efficacy and efficiency of medicine, a push toward greater acceptance of artificial intelligence-driven medicine could unlock benefits for patients and providers.

How, then, can the industry help nudge the public to feel more comfortable with medical AI?

According to a new study from researchers at Lehigh University and Seattle University, making the concept of bias more salient in their thinking can help.

Study explores 'bias salience'

The study, published in the journal Computers in Human Behavior, found that patients were more receptive to medical recommendations from AI when they were made more aware of the biases inherent in human healthcare decisions. This "bias salience," or making people more conscious of bias in decision-making, was shown to shift people's perceptions.

"This increased receptiveness to AI occurs because bias is perceived to be a fundamentally human shortcoming," said Rebecca J. H. Wang , associate professor of marketing in the Lehigh College of Business. "As such, when the prospect of bias is made salient, perceptions of AI integrity—defined as the perceived fairness and trustworthiness of an AI agent relative to a human counterpart—are enhanced."

The study entailed six experiments with nearly 1,900 participants, demonstrating that when participants were reminded of human biases—such as how healthcare providers might treat patients differently based on characteristics like gender—they became more receptive to AI recommendations.

Participants were presented with scenarios in which they would be seeking a recommendation or diagnosis such as a coronary bypass or skin cancer screening. They were then asked whether they preferred the recommendation to be made by a human being or by a computer/AI assistant.

Prior to being presented with the scenario, some participants were presented with prior screens intended to increase bias salience. These interventions included reviewing an infographic that highlighted common cognitive biases, describing a time when they were negatively affected by bias, including age-related bias for those over 50, and gender-related bias. The results showed:

  • When participants were made aware of potential biases in human healthcare, they rated AI as offering greater "integrity," meaning they perceived it as more fair and trustworthy.
  • While bias salience did not eliminate people's general preference for human healthcare, high bias salience reduces resistance to medical AI, presumably because bias is more readily associated with human beings.
  • In the absence of bias salience, the subjectivity of human providers is often viewed as a positive, but when bias salience is high, patients place greater value on the perceived objectivity of AI.

The future of AI in medicine

The authors stress the importance of keeping human biases in mind at all stages of AI proliferation, from development to adoption. Developers of AI systems should both aim to minimize bias inherent in the materials used to train medical AI—a known issue in the current development of AI systems—and provide context about human bias when users encounter it.

Doing so could help providers capitalize on AI's growing role in applications such as diagnostics, treatment recommendations, and patient monitoring. Those roles are only expected to expand as the industry is projected to invest more than $30 billion into AI medicine annually by 2029.

"By addressing patients' concerns about AI and highlighting the limitations of human judgment, healthcare providers can create a more balanced and trusting relationship between patients and emerging technologies," Wang said.

The study was a collaboration between Wang; Lucy E. Napper , associate professor of psychology at Lehigh; Jessecae Marsh , professor of psychology and associate dean at Lehigh; and Mathew S. Isaac, professor and chair of the department of marketing at Seattle University. The paper, "To err is human: Bias salience can help overcome resistance to medical AI," is available online.

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.