As an industry reliant on patient records and beset by outdated technology, health care is widely thought to be a prime target for an artificial intelligence revolution.
Many believe the technology will provide a host of benefits to clinical practitioners, speeding up the overall experience and diagnosing illnesses early on to identify potential treatment.
Just two days ago, DeepMind, an AI (artificial intelligence) firm owned by Google, said it had lent its technology to London’s Moorfields Eye Hospital for groundbreaking research into detecting eye diseases. It was used to scan and identify more than 50 ophthalmological conditions. DeepMind’s machine-learning technology made correct diagnoses 94 percent of the time, Moorfields said.
The development indicated that AI can analyze health problems with as much accuracy as a doctor. But some doctors worry that those in the tech world think AI can not only help clinicians, but even do a better job.
Take Babylon Health for instance, which in June said its AI chatbot was able to diagnose medical conditions as accurately as a doctor. The firm’s chatbot scored a higher-than-average test score on a practice exam compiled for physicians.
Babylon’s chatbot passed 82 percent of the test’s questions, versus the average mark for human doctors of 72 percent.
But the Royal College of General Practitioners (RCGP), an industry body representing doctors that treat a wide range of common illnesses, quickly disputed the claim that AI could diagnose illnesses with the same effectiveness of a human medical practitioner.
“No app or algorithm can do what a GP does,” Helen Stokes-Lampard, a professor and chair of the RCGP, told CNBC earlier this week. “Every day we deliver care to more than a million people across the U.K., taking into account the physical, psychological and social factors that may be impacting on each person’s health.”
Stokes-Lampard continued: “We consider the different health conditions a patient is living with, their family history, any medications they might be taking, and a myriad of other considerations when formulating a treatment plan.”
Babylon at the time denied it had claimed an AI could do the job of a GP, saying that it supported a model where AI is complementary to medical practice.
‘The role of the doctor will have to adapt’
Nevertheless, the spat highlighted a serious question that may one day need to be addressed by those in the health industry: How should health professionals respond to the rapid growth in new, data-driven technologies like AI?
“Over the next decade or two, AI will certainly play a big role in supporting doctors make decisions,” Dan Vahdat, chief executive of health tech start-up Medopad, told CNBC via email.
“The role of the doctor will have to adapt in learning how to use AI to complement their clinical judgments. This will take time, but it’s inevitable.”
Medopad specializes in connecting health-care providers, doctors and patients to monitor a patient’s health data and see how their care can be improved.
In the U.K., the National Health Service, the country’s universal health care system, has come under strain both in terms of funding and resources. The promise of AI to reduce the financial burden on medical services by cutting out some disposable roles and functions could be music to the ears of government and health authorities.
Vahdat said that one area in which AI could be hugely beneficial in improving efficiency and cutting costs was cardiological care.
“If we collect data in real time and combine it with their historical data as well as data from millions of similar patients, an intelligent system could predict a heart attack with a high rate of precision,” he said.
“This will allow us to save countless lives and save health care systems around the world huge spend.”
But some experts fear the fast-paced nature of the still nascent AI industry could come at the risk of patient safety.
“We are concerned that in the rush to roll out AI and push the boundaries of technology, there is a risk that important checks and balances that have been established to keep patients safe might be seen as an afterthought, or be bypassed entirely,” Stokes-Lampard said.