acm-header
Sign In

Communications of the ACM

ACM Careers

AI's Health Advice Equivalent to Human Doctors, Startup Says


View as: Print Mobile App Share:
robot doctor, illustration

Startup Babylon Health announced that the company's AI, in a series of robust tests (including the relevant sections of the MRCGP exam), has demonstrated its ability to provide health advice which is on-par with practicing clinicians1.

The MRCGP exam is the final test for trainee General Practitioners, set by the Royal College of General Practitioners (RCGP) in the U.K. Trainee General Practitioners who pass this assessment have demonstrated their competence and clinical skills to a level which is sufficiently high enough for them to undertake independent practice.

A key part of this exam tests a doctor's ability to diagnose.

Babylon took a representative sample-set of questions testing diagnostic skills from publicly available RCGP sources2, as well as independently published examination preparation materials, and mapped these to the current RCGP curriculum in order to ensure the questions resembled actual MRCGP questions as closely as possible.

The average pass mark over the past five years for real-life doctors was 72% 3. In sitting the exam for the first time, Babylon's AI scored 81%. As the AI continues to learn and accumulate knowledge, Babylon expects that subsequent testing will produce significant improvements in terms of results.

Important though exams are, doctors are presented with a much wider range of illnesses and conditions in their daily practice. Therefore, to further test the AI's capabilities, Babylon's team of scientists, clinicians, and engineers next collaborated with the Royal College of Physicians, Dr. Megan Mahoney (Chief of General Primary Care, Division of Primary Care and Population Health, Stanford University), and Dr. Arnold DoRosario (Chief Population Health Officer, Yale New Haven Health), to test Babylon's AI alongside seven highly-experienced primary care doctors using 100 independently-devised symptom sets, or "vignettes."

Babylon's AI scored 80% for accuracy, while the seven doctors achieved an accuracy range of 64-to-94%.

The accuracy of the AI was 98% when assessed against conditions seen most frequently in primary care medicine. In comparison, when Babylon's research team assessed experienced clinicians using the same measure, their accuracy ranged from 52-to-99%.

Crucially, the safety of the AI was 97%. This compares favorably to the doctors, whose average was 93.1%.

The AI system and performance is described in "A Comparative Study of Artificial Intelligence and Human Doctors for the Purpose of Triage and Diagnosis."

Dr. Ali Parsa, Babylon's founder and CEO, says, "The World Health Organization estimates that there is a shortage of over five million doctors globally, leaving more than half the world's population without access to even the most basic healthcare services. Even in the richest nations, primary care is becoming increasingly unaffordable and inconvenient, often with waiting times that make it not readily accessible. Babylon's latest artificial intelligence capabilities show that it is possible for anyone, irrespective of their geography, wealth, or circumstances, to have free access to health advice that is on-par with top-rated practicing clinicians.

"[The] results clearly illustrate how AI-augmented health services can reduce the burden on healthcare systems around the world," Dr. Parsa says. "Our mission is to put accessible and affordable health services into the hands of every person on Earth. These landmark results take humanity a significant step closer to achieving a world where no-one is denied safe and accurate health advice."

References:

((1. Babylon's technology provides health information, rather than a medical diagnosis, for regulatory reasons. The tests carried out relate to the diagnostic exams taken by doctors as a benchmark for accuracy, however, Babylon's AI service remains an information service, rather than a medical diagnosis.

2. As the RCGP does not publish past papers, Babylon used example questions - some published directly by the College, some which were sourced from publicly available resources (which all are referenced) – during its AI exam preparation and testing.))

3. Average pass mark was calculated using publicly available RCGP pass mark data from the period 2013 – 2018.


 

No entries found

Sign In for Full Access
» Forgot Password? » Create an ACM Web Account