Researchers aim to bridge the gap between AI technology and human understanding — ScienceDaily

University of Waterloo researchers have developed a new explainable artificial intelligence (AI) model to reduce bias and enhance trust and accuracy in machine learning-generated decision-making and knowledge organization.

Traditional machine learning models often yield biased results, favouring groups with large populations or being influenced by unknown factors, and take extensive effort to identify from instances containing patterns and sub-patterns coming from different classes or primary sources.

The medical field is one area where there are severe implications for biased machine learning results. Hospital staff and medical professionals rely on datasets containing thousands of medical records and complex computer algorithms to make critical decisions about patient care. Machine learning is used to sort the data, which saves time. However, specific patient groups with rare symptomatic patterns may go undetected, and mislabeled patients and anomalies could impact diagnostic outcomes. This inherent bias and pattern entanglement leads to misdiagnoses and inequitable healthcare outcomes for specific patient groups.

Thanks to new research led by Dr. Andrew Wong, a distinguished professor emeritus of systems design engineering at Waterloo, an innovative model aims to eliminate these barriers by untangling complex patterns from data to relate them to specific underlying causes unaffected by anomalies and mislabeled instances. It can enhance trust and reliability in Explainable Artificial Intelligence (XAI.)

“This research represents a significant contribution to the field of XAI,” Wong said. “While analyzing a vast amount of protein binding data from X-ray crystallography, my team revealed the statistics of the physicochemical amino acid interacting patterns which were masked and mixed at the data level due to the entanglement of multiple factors present in the binding environment. That was the first time we showed entangled statistics can be disentangled to give a correct picture of the deep knowledge missed at the data level with scientific evidence.”

This revelation led Wong and his team to develop the new XAI model called Pattern Discovery and Disentanglement (PDD).

“With PDD, we aim to bridge the gap between AI technology and human understanding to help enable trustworthy decision-making and unlock deeper knowledge from complex data sources,” said Dr. Peiyuan Zhou, the lead researcher on Wong’s team.

Professor Annie Lee, a co-author and collaborator from the University of Toronto, specializing in Natural Language Processing, foresees the immense value of PDD contribution to clinical decision-making.

The PDD model has revolutionized pattern discovery. Various case studies have showcased PDD, demonstrating an ability to predict patients’ medical results based on their clinical records. The PDD system can also discover new and rare patterns in datasets. This allows researchers and practitioners alike to detect mislabels or anomalies in machine learning.

The result shows that healthcare professionals can make more reliable diagnoses supported by rigorous statistics and explainable patterns for better treatment recommendations for various diseases at different stages.

The study, Theory and rationale of interpretable all-in-one pattern discovery and disentanglement system, appears in the journal npj Digital Medicine.

The recent award of an NSER Idea-to-Innovation Grant of $125 K on PDD indicates its industrial recognition. PDD is commercialized via Waterloo Commercialization Office.

Comments (0)
Add Comment