University of Toronto researcher Parham Aarabi has created an artificial intelligence model that mimics how people human users e-commerce websites - and it may be able to help retailers optimize their sites for people experiencing colour blindness and other conditions.
Called PRE, the AI-generated tool sees virtual users browse, pause on a page, add items to cart and click on discounted items.
While the tool shows that users tend to be drawn to colourful images, Aarabi also wanted to see how those experiencing full and partial colour blindness might respond.
"Around eight to 10 per cent of the population has a type of colour-blindness," says Aarabi, an associate professor in the Edward S. Rogers Sr. department of electrical and computer engineering in the Faculty of Applied Science & Engineering. "There are a number of ways the eye can be confused by colour, commonly between red and green or blue and yellow.
"I wanted to see how this might impact web navigation."
Aarabi set up an experiment. He altered a retail clothing website to simulate how it would appear to someone with protanomaly, or a reduced ability to perceive red light. One might think of it as applying a filter, or lens, which Aarabi then modified to approximate eight other variations of colour deficiency.
For each variation, Aarabi initiated one million navigation sessions with AI virtual users and tracked the image click rates. He found that, in general, someone with colour-blindness is 30 per cent more likely than a colour-abled user to click on a monochrome image. The results will be presented in a paper at the 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society this summer.
The boost factor that website designers count on with colour doesn't translate to everyone, notes Aarabi.
"When people are designing sites or presenting products, they need to stay cognizant that eight per cent of the population is not going to be impacted. You need to add better descriptions and more textual information to guide users through the shopping process."
Aarabi sees this study as one of many that can benefit from PRE, whose neural net took two years to train with data from 110,000 real-life user sessions.
"To measure its accuracy, we set up a sample site and predicted what actions the AI virtual users will take - what percentage would add to cart, what percentage would buy a particular product, and so on - and also ran a test of the site with people," says Aarabi. "PRE correctly mimics a human user's actions 90 per cent of the time."
There are benefits with using AI virtual users for a study. One can run experiments more quickly, on a larger scale, and can recreate as many sessions as desired. The AI model eliminates the need, for example, to locate and coordinate many thousands of willing colour-blind participants.
Aarabi has plans to use PRE to test other barriers to accessibility, such as dyslexia or motor impairments. His long-term goal is provide an auditing service for companies that allows them to test a web design's impact on users with various conditions before or after launch.
Such goals are part of Aarabi's research effort to mitigate negativity about AI.
"There's a lot of worry, even within the tech community, about AI taking over or replacing us in some capacity," he says. "If we can make AI more humanlike in some way, build in some empathy and have it mirror the reactions that humans have, we could dispel some of those concerns."
"Professor Aarabi has been a pioneer in the application of AI, from past research cautioning against bias in training data sets to this current project, which uses the AI advantage to address accessibility issues," says Professor Deepa Kundur, chair of the department of electrical and computer engineering. "Parham brings a valuable, forward-thinking approach to leveraging AI for positive outcomes."